2006-01-02 09:14:23 +00:00
|
|
|
/* i915_irq.c -- IRQ support for the I915 -*- linux-c -*-
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2006-01-02 09:14:23 +00:00
|
|
|
/*
|
2005-04-16 22:20:36 +00:00
|
|
|
* Copyright 2003 Tungsten Graphics, Inc., Cedar Park, Texas.
|
|
|
|
* All Rights Reserved.
|
2005-06-23 12:46:46 +00:00
|
|
|
*
|
|
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
|
|
* copy of this software and associated documentation files (the
|
|
|
|
* "Software"), to deal in the Software without restriction, including
|
|
|
|
* without limitation the rights to use, copy, modify, merge, publish,
|
|
|
|
* distribute, sub license, and/or sell copies of the Software, and to
|
|
|
|
* permit persons to whom the Software is furnished to do so, subject to
|
|
|
|
* the following conditions:
|
|
|
|
*
|
|
|
|
* The above copyright notice and this permission notice (including the
|
|
|
|
* next paragraph) shall be included in all copies or substantial portions
|
|
|
|
* of the Software.
|
|
|
|
*
|
|
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
|
|
|
|
* OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
|
|
|
|
* IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR
|
|
|
|
* ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
|
|
|
|
* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
|
|
|
|
* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
|
|
|
*
|
2006-01-02 09:14:23 +00:00
|
|
|
*/
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2012-03-18 20:00:11 +00:00
|
|
|
#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
|
|
|
|
|
2009-06-18 23:56:52 +00:00
|
|
|
#include <linux/sysrq.h>
|
include cleanup: Update gfp.h and slab.h includes to prepare for breaking implicit slab.h inclusion from percpu.h
percpu.h is included by sched.h and module.h and thus ends up being
included when building most .c files. percpu.h includes slab.h which
in turn includes gfp.h making everything defined by the two files
universally available and complicating inclusion dependencies.
percpu.h -> slab.h dependency is about to be removed. Prepare for
this change by updating users of gfp and slab facilities include those
headers directly instead of assuming availability. As this conversion
needs to touch large number of source files, the following script is
used as the basis of conversion.
http://userweb.kernel.org/~tj/misc/slabh-sweep.py
The script does the followings.
* Scan files for gfp and slab usages and update includes such that
only the necessary includes are there. ie. if only gfp is used,
gfp.h, if slab is used, slab.h.
* When the script inserts a new include, it looks at the include
blocks and try to put the new include such that its order conforms
to its surrounding. It's put in the include block which contains
core kernel includes, in the same order that the rest are ordered -
alphabetical, Christmas tree, rev-Xmas-tree or at the end if there
doesn't seem to be any matching order.
* If the script can't find a place to put a new include (mostly
because the file doesn't have fitting include block), it prints out
an error message indicating which .h file needs to be added to the
file.
The conversion was done in the following steps.
1. The initial automatic conversion of all .c files updated slightly
over 4000 files, deleting around 700 includes and adding ~480 gfp.h
and ~3000 slab.h inclusions. The script emitted errors for ~400
files.
2. Each error was manually checked. Some didn't need the inclusion,
some needed manual addition while adding it to implementation .h or
embedding .c file was more appropriate for others. This step added
inclusions to around 150 files.
3. The script was run again and the output was compared to the edits
from #2 to make sure no file was left behind.
4. Several build tests were done and a couple of problems were fixed.
e.g. lib/decompress_*.c used malloc/free() wrappers around slab
APIs requiring slab.h to be added manually.
5. The script was run on all .h files but without automatically
editing them as sprinkling gfp.h and slab.h inclusions around .h
files could easily lead to inclusion dependency hell. Most gfp.h
inclusion directives were ignored as stuff from gfp.h was usually
wildly available and often used in preprocessor macros. Each
slab.h inclusion directive was examined and added manually as
necessary.
6. percpu.h was updated not to include slab.h.
7. Build test were done on the following configurations and failures
were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my
distributed build env didn't work with gcov compiles) and a few
more options had to be turned off depending on archs to make things
build (like ipr on powerpc/64 which failed due to missing writeq).
* x86 and x86_64 UP and SMP allmodconfig and a custom test config.
* powerpc and powerpc64 SMP allmodconfig
* sparc and sparc64 SMP allmodconfig
* ia64 SMP allmodconfig
* s390 SMP allmodconfig
* alpha SMP allmodconfig
* um on x86_64 SMP allmodconfig
8. percpu.h modifications were reverted so that it could be applied as
a separate patch and serve as bisection point.
Given the fact that I had only a couple of failures from tests on step
6, I'm fairly confident about the coverage of this conversion patch.
If there is a breakage, it's likely to be something in one of the arch
headers which should be easily discoverable easily on most builds of
the specific arch.
Signed-off-by: Tejun Heo <tj@kernel.org>
Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
2010-03-24 08:04:11 +00:00
|
|
|
#include <linux/slab.h>
|
2013-10-15 17:55:29 +00:00
|
|
|
#include <linux/circ_buf.h>
|
2012-10-02 17:01:07 +00:00
|
|
|
#include <drm/drmP.h>
|
|
|
|
#include <drm/i915_drm.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
#include "i915_drv.h"
|
2009-08-25 10:15:50 +00:00
|
|
|
#include "i915_trace.h"
|
DRM: i915: add mode setting support
This commit adds i915 driver support for the DRM mode setting APIs.
Currently, VGA, LVDS, SDVO DVI & VGA, TV and DVO LVDS outputs are
supported. HDMI, DisplayPort and additional SDVO output support will
follow.
Support for the mode setting code is controlled by the new 'modeset'
module option. A new config option, CONFIG_DRM_I915_KMS controls the
default behavior, and whether a PCI ID list is built into the module for
use by user level module utilities.
Note that if mode setting is enabled, user level drivers that access
display registers directly or that don't use the kernel graphics memory
manager will likely corrupt kernel graphics memory, disrupt output
configuration (possibly leading to hangs and/or blank displays), and
prevent panic/oops messages from appearing. So use caution when
enabling this code; be sure your user level code supports the new
interfaces.
A new SysRq key, 'g', provides emergency support for switching back to
the kernel's framebuffer console; which is useful for testing.
Co-authors: Dave Airlie <airlied@linux.ie>, Hong Liu <hong.liu@intel.com>
Signed-off-by: Jesse Barnes <jbarnes@virtuousgeek.org>
Signed-off-by: Eric Anholt <eric@anholt.net>
Signed-off-by: Dave Airlie <airlied@redhat.com>
2008-11-07 22:24:08 +00:00
|
|
|
#include "intel_drv.h"
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2014-09-30 08:56:45 +00:00
|
|
|
/**
|
|
|
|
* DOC: interrupt handling
|
|
|
|
*
|
|
|
|
* These functions provide the basic support for enabling and disabling the
|
|
|
|
* interrupt handling support. There's a lot more functionality in i915_irq.c
|
|
|
|
* and related files, but that will be described in separate chapters.
|
|
|
|
*/
|
|
|
|
|
2015-08-27 20:56:03 +00:00
|
|
|
static const u32 hpd_ilk[HPD_NUM_PINS] = {
|
|
|
|
[HPD_PORT_A] = DE_DP_A_HOTPLUG,
|
|
|
|
};
|
|
|
|
|
2015-08-27 20:56:04 +00:00
|
|
|
static const u32 hpd_ivb[HPD_NUM_PINS] = {
|
|
|
|
[HPD_PORT_A] = DE_DP_A_HOTPLUG_IVB,
|
|
|
|
};
|
|
|
|
|
2015-08-27 20:56:06 +00:00
|
|
|
static const u32 hpd_bdw[HPD_NUM_PINS] = {
|
|
|
|
[HPD_PORT_A] = GEN8_PORT_DP_A_HOTPLUG,
|
|
|
|
};
|
|
|
|
|
2015-01-09 12:21:12 +00:00
|
|
|
static const u32 hpd_ibx[HPD_NUM_PINS] = {
|
2013-02-28 09:17:12 +00:00
|
|
|
[HPD_CRT] = SDE_CRT_HOTPLUG,
|
|
|
|
[HPD_SDVO_B] = SDE_SDVOB_HOTPLUG,
|
|
|
|
[HPD_PORT_B] = SDE_PORTB_HOTPLUG,
|
|
|
|
[HPD_PORT_C] = SDE_PORTC_HOTPLUG,
|
|
|
|
[HPD_PORT_D] = SDE_PORTD_HOTPLUG
|
|
|
|
};
|
|
|
|
|
2015-01-09 12:21:12 +00:00
|
|
|
static const u32 hpd_cpt[HPD_NUM_PINS] = {
|
2013-02-28 09:17:12 +00:00
|
|
|
[HPD_CRT] = SDE_CRT_HOTPLUG_CPT,
|
2013-03-26 21:38:43 +00:00
|
|
|
[HPD_SDVO_B] = SDE_SDVOB_HOTPLUG_CPT,
|
2013-02-28 09:17:12 +00:00
|
|
|
[HPD_PORT_B] = SDE_PORTB_HOTPLUG_CPT,
|
|
|
|
[HPD_PORT_C] = SDE_PORTC_HOTPLUG_CPT,
|
|
|
|
[HPD_PORT_D] = SDE_PORTD_HOTPLUG_CPT
|
|
|
|
};
|
|
|
|
|
2015-08-17 07:55:50 +00:00
|
|
|
static const u32 hpd_spt[HPD_NUM_PINS] = {
|
2015-08-27 20:56:07 +00:00
|
|
|
[HPD_PORT_A] = SDE_PORTA_HOTPLUG_SPT,
|
2015-08-17 07:55:50 +00:00
|
|
|
[HPD_PORT_B] = SDE_PORTB_HOTPLUG_CPT,
|
|
|
|
[HPD_PORT_C] = SDE_PORTC_HOTPLUG_CPT,
|
|
|
|
[HPD_PORT_D] = SDE_PORTD_HOTPLUG_CPT,
|
|
|
|
[HPD_PORT_E] = SDE_PORTE_HOTPLUG_SPT
|
|
|
|
};
|
|
|
|
|
2015-01-09 12:21:12 +00:00
|
|
|
static const u32 hpd_mask_i915[HPD_NUM_PINS] = {
|
2013-02-28 09:17:12 +00:00
|
|
|
[HPD_CRT] = CRT_HOTPLUG_INT_EN,
|
|
|
|
[HPD_SDVO_B] = SDVOB_HOTPLUG_INT_EN,
|
|
|
|
[HPD_SDVO_C] = SDVOC_HOTPLUG_INT_EN,
|
|
|
|
[HPD_PORT_B] = PORTB_HOTPLUG_INT_EN,
|
|
|
|
[HPD_PORT_C] = PORTC_HOTPLUG_INT_EN,
|
|
|
|
[HPD_PORT_D] = PORTD_HOTPLUG_INT_EN
|
|
|
|
};
|
|
|
|
|
2015-01-09 12:21:12 +00:00
|
|
|
static const u32 hpd_status_g4x[HPD_NUM_PINS] = {
|
2013-02-28 09:17:12 +00:00
|
|
|
[HPD_CRT] = CRT_HOTPLUG_INT_STATUS,
|
|
|
|
[HPD_SDVO_B] = SDVOB_HOTPLUG_INT_STATUS_G4X,
|
|
|
|
[HPD_SDVO_C] = SDVOC_HOTPLUG_INT_STATUS_G4X,
|
|
|
|
[HPD_PORT_B] = PORTB_HOTPLUG_INT_STATUS,
|
|
|
|
[HPD_PORT_C] = PORTC_HOTPLUG_INT_STATUS,
|
|
|
|
[HPD_PORT_D] = PORTD_HOTPLUG_INT_STATUS
|
|
|
|
};
|
|
|
|
|
drm/i915: Use HOTPLUG_INT_STATUS_G4X on VLV/CHV
Use HOTPLUG_INT_STATUS_G4X instead of HOTPLUG_INT_STATUS_I915 on VLV/CHV
so that we don't confuse the AUX status bits with SDVO status bits.
Avoid pointless log spam as below while handling AUX interrupts:
[drm:intel_hpd_irq_handler] hotplug event received, stat 0x00000040, dig 0x00000000
[drm:intel_hpd_irq_handler] hotplug event received, stat 0x00000040, dig 0x00000000
[drm:intel_hpd_irq_handler] hotplug event received, stat 0x00000040, dig 0x00000000
[drm:intel_hpd_irq_handler] hotplug event received, stat 0x00000040, dig 0x00000000
[drm:intel_hpd_irq_handler] hotplug event received, stat 0x00000040, dig 0x00000000
[drm:intel_dp_aux_ch] dp_aux_ch timeout status 0x71450064
Note that there's no functional issue, it's just that the sdvo bits
overlap with the dp aux bits. Hence every time we receive an aux
interrupt we also think there's an sdvo hpd interrupt, but due to lack
of any sdvo encoders nothing ever happens because of that.
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Jani Nikula <jani.nikula@intel.com>
[danvet: Add Ville's explanation why nothing functional really
changes.]
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2015-05-11 17:49:10 +00:00
|
|
|
static const u32 hpd_status_i915[HPD_NUM_PINS] = {
|
2013-02-28 09:17:12 +00:00
|
|
|
[HPD_CRT] = CRT_HOTPLUG_INT_STATUS,
|
|
|
|
[HPD_SDVO_B] = SDVOB_HOTPLUG_INT_STATUS_I915,
|
|
|
|
[HPD_SDVO_C] = SDVOC_HOTPLUG_INT_STATUS_I915,
|
|
|
|
[HPD_PORT_B] = PORTB_HOTPLUG_INT_STATUS,
|
|
|
|
[HPD_PORT_C] = PORTC_HOTPLUG_INT_STATUS,
|
|
|
|
[HPD_PORT_D] = PORTD_HOTPLUG_INT_STATUS
|
|
|
|
};
|
|
|
|
|
2015-03-27 12:54:14 +00:00
|
|
|
/* BXT hpd list */
|
|
|
|
static const u32 hpd_bxt[HPD_NUM_PINS] = {
|
2015-08-10 05:05:35 +00:00
|
|
|
[HPD_PORT_A] = BXT_DE_PORT_HP_DDIA,
|
2015-03-27 12:54:14 +00:00
|
|
|
[HPD_PORT_B] = BXT_DE_PORT_HP_DDIB,
|
|
|
|
[HPD_PORT_C] = BXT_DE_PORT_HP_DDIC
|
|
|
|
};
|
|
|
|
|
2014-04-01 18:37:11 +00:00
|
|
|
/* IIR can theoretically queue up two events. Be paranoid. */
|
2014-04-01 18:37:14 +00:00
|
|
|
#define GEN8_IRQ_RESET_NDX(type, which) do { \
|
2014-04-01 18:37:11 +00:00
|
|
|
I915_WRITE(GEN8_##type##_IMR(which), 0xffffffff); \
|
|
|
|
POSTING_READ(GEN8_##type##_IMR(which)); \
|
|
|
|
I915_WRITE(GEN8_##type##_IER(which), 0); \
|
|
|
|
I915_WRITE(GEN8_##type##_IIR(which), 0xffffffff); \
|
|
|
|
POSTING_READ(GEN8_##type##_IIR(which)); \
|
|
|
|
I915_WRITE(GEN8_##type##_IIR(which), 0xffffffff); \
|
|
|
|
POSTING_READ(GEN8_##type##_IIR(which)); \
|
|
|
|
} while (0)
|
|
|
|
|
2017-08-18 18:36:52 +00:00
|
|
|
#define GEN3_IRQ_RESET(type) do { \
|
2014-04-01 18:37:09 +00:00
|
|
|
I915_WRITE(type##IMR, 0xffffffff); \
|
2014-04-01 18:37:11 +00:00
|
|
|
POSTING_READ(type##IMR); \
|
2014-04-01 18:37:09 +00:00
|
|
|
I915_WRITE(type##IER, 0); \
|
2014-04-01 18:37:11 +00:00
|
|
|
I915_WRITE(type##IIR, 0xffffffff); \
|
|
|
|
POSTING_READ(type##IIR); \
|
|
|
|
I915_WRITE(type##IIR, 0xffffffff); \
|
|
|
|
POSTING_READ(type##IIR); \
|
2014-04-01 18:37:09 +00:00
|
|
|
} while (0)
|
|
|
|
|
2017-08-18 18:36:54 +00:00
|
|
|
#define GEN2_IRQ_RESET(type) do { \
|
|
|
|
I915_WRITE16(type##IMR, 0xffff); \
|
|
|
|
POSTING_READ16(type##IMR); \
|
|
|
|
I915_WRITE16(type##IER, 0); \
|
|
|
|
I915_WRITE16(type##IIR, 0xffff); \
|
|
|
|
POSTING_READ16(type##IIR); \
|
|
|
|
I915_WRITE16(type##IIR, 0xffff); \
|
|
|
|
POSTING_READ16(type##IIR); \
|
|
|
|
} while (0)
|
|
|
|
|
2014-04-01 18:37:16 +00:00
|
|
|
/*
|
|
|
|
* We should clear IMR at preinstall/uninstall, and just check at postinstall.
|
|
|
|
*/
|
2017-08-18 18:36:52 +00:00
|
|
|
static void gen3_assert_iir_is_zero(struct drm_i915_private *dev_priv,
|
drm/i915: Type safe register read/write
Make I915_READ and I915_WRITE more type safe by wrapping the register
offset in a struct. This should eliminate most of the fumbles we've had
with misplaced parens.
This only takes care of normal mmio registers. We could extend the idea
to other register types and define each with its own struct. That way
you wouldn't be able to accidentally pass the wrong thing to a specific
register access function.
The gpio_reg setup is probably the ugliest thing left. But I figure I'd
just leave it for now, and wait for some divine inspiration to strike
before making it nice.
As for the generated code, it's actually a bit better sometimes. Eg.
looking at i915_irq_handler(), we can see the following change:
lea 0x70024(%rdx,%rax,1),%r9d
mov $0x1,%edx
- movslq %r9d,%r9
- mov %r9,%rsi
- mov %r9,-0x58(%rbp)
- callq *0xd8(%rbx)
+ mov %r9d,%esi
+ mov %r9d,-0x48(%rbp)
callq *0xd8(%rbx)
So previously gcc thought the register offset might be signed and
decided to sign extend it, just in case. The rest appears to be
mostly just minor shuffling of instructions.
v2: i915_mmio_reg_{offset,equal,valid}() helpers added
s/_REG/_MMIO/ in the register defines
mo more switch statements left to worry about
ring_emit stuff got sorted in a prep patch
cmd parser, lrc context and w/a batch buildup also in prep patch
vgpu stuff cleaned up and moved to a prep patch
all other unrelated changes split out
v3: Rebased due to BXT DSI/BLC, MOCS, etc.
v4: Rebased due to churn, s/i915_mmio_reg_t/i915_reg_t/
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/1447853606-2751-1-git-send-email-ville.syrjala@linux.intel.com
2015-11-18 13:33:26 +00:00
|
|
|
i915_reg_t reg)
|
2015-09-18 17:03:41 +00:00
|
|
|
{
|
|
|
|
u32 val = I915_READ(reg);
|
|
|
|
|
|
|
|
if (val == 0)
|
|
|
|
return;
|
|
|
|
|
|
|
|
WARN(1, "Interrupt register 0x%x is not zero: 0x%08x\n",
|
drm/i915: Type safe register read/write
Make I915_READ and I915_WRITE more type safe by wrapping the register
offset in a struct. This should eliminate most of the fumbles we've had
with misplaced parens.
This only takes care of normal mmio registers. We could extend the idea
to other register types and define each with its own struct. That way
you wouldn't be able to accidentally pass the wrong thing to a specific
register access function.
The gpio_reg setup is probably the ugliest thing left. But I figure I'd
just leave it for now, and wait for some divine inspiration to strike
before making it nice.
As for the generated code, it's actually a bit better sometimes. Eg.
looking at i915_irq_handler(), we can see the following change:
lea 0x70024(%rdx,%rax,1),%r9d
mov $0x1,%edx
- movslq %r9d,%r9
- mov %r9,%rsi
- mov %r9,-0x58(%rbp)
- callq *0xd8(%rbx)
+ mov %r9d,%esi
+ mov %r9d,-0x48(%rbp)
callq *0xd8(%rbx)
So previously gcc thought the register offset might be signed and
decided to sign extend it, just in case. The rest appears to be
mostly just minor shuffling of instructions.
v2: i915_mmio_reg_{offset,equal,valid}() helpers added
s/_REG/_MMIO/ in the register defines
mo more switch statements left to worry about
ring_emit stuff got sorted in a prep patch
cmd parser, lrc context and w/a batch buildup also in prep patch
vgpu stuff cleaned up and moved to a prep patch
all other unrelated changes split out
v3: Rebased due to BXT DSI/BLC, MOCS, etc.
v4: Rebased due to churn, s/i915_mmio_reg_t/i915_reg_t/
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/1447853606-2751-1-git-send-email-ville.syrjala@linux.intel.com
2015-11-18 13:33:26 +00:00
|
|
|
i915_mmio_reg_offset(reg), val);
|
2015-09-18 17:03:41 +00:00
|
|
|
I915_WRITE(reg, 0xffffffff);
|
|
|
|
POSTING_READ(reg);
|
|
|
|
I915_WRITE(reg, 0xffffffff);
|
|
|
|
POSTING_READ(reg);
|
|
|
|
}
|
2014-04-01 18:37:16 +00:00
|
|
|
|
2017-08-18 18:36:54 +00:00
|
|
|
static void gen2_assert_iir_is_zero(struct drm_i915_private *dev_priv,
|
|
|
|
i915_reg_t reg)
|
|
|
|
{
|
|
|
|
u16 val = I915_READ16(reg);
|
|
|
|
|
|
|
|
if (val == 0)
|
|
|
|
return;
|
|
|
|
|
|
|
|
WARN(1, "Interrupt register 0x%x is not zero: 0x%08x\n",
|
|
|
|
i915_mmio_reg_offset(reg), val);
|
|
|
|
I915_WRITE16(reg, 0xffff);
|
|
|
|
POSTING_READ16(reg);
|
|
|
|
I915_WRITE16(reg, 0xffff);
|
|
|
|
POSTING_READ16(reg);
|
|
|
|
}
|
|
|
|
|
2014-04-01 18:37:15 +00:00
|
|
|
#define GEN8_IRQ_INIT_NDX(type, which, imr_val, ier_val) do { \
|
2017-08-18 18:36:52 +00:00
|
|
|
gen3_assert_iir_is_zero(dev_priv, GEN8_##type##_IIR(which)); \
|
2014-04-01 18:37:15 +00:00
|
|
|
I915_WRITE(GEN8_##type##_IER(which), (ier_val)); \
|
drm/i915: Apply some ocd for IMR vs. IER order during irq enable
When disabling interrupts we do the writes in this order:
IMR,IER,IIR,IIR. But when enabling interrupts we don't do use the
mirrored order, and instead do IIR,IIR,IMR,IER.
I like consistency unless there's a good reason against it, which I
can't think of here, so change the enable order to IIR,IIR,IER,IMR.
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2014-10-30 17:42:50 +00:00
|
|
|
I915_WRITE(GEN8_##type##_IMR(which), (imr_val)); \
|
|
|
|
POSTING_READ(GEN8_##type##_IMR(which)); \
|
2014-04-01 18:37:15 +00:00
|
|
|
} while (0)
|
|
|
|
|
2017-08-18 18:36:52 +00:00
|
|
|
#define GEN3_IRQ_INIT(type, imr_val, ier_val) do { \
|
|
|
|
gen3_assert_iir_is_zero(dev_priv, type##IIR); \
|
2014-04-01 18:37:15 +00:00
|
|
|
I915_WRITE(type##IER, (ier_val)); \
|
drm/i915: Apply some ocd for IMR vs. IER order during irq enable
When disabling interrupts we do the writes in this order:
IMR,IER,IIR,IIR. But when enabling interrupts we don't do use the
mirrored order, and instead do IIR,IIR,IMR,IER.
I like consistency unless there's a good reason against it, which I
can't think of here, so change the enable order to IIR,IIR,IER,IMR.
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2014-10-30 17:42:50 +00:00
|
|
|
I915_WRITE(type##IMR, (imr_val)); \
|
|
|
|
POSTING_READ(type##IMR); \
|
2014-04-01 18:37:15 +00:00
|
|
|
} while (0)
|
|
|
|
|
2017-08-18 18:36:54 +00:00
|
|
|
#define GEN2_IRQ_INIT(type, imr_val, ier_val) do { \
|
|
|
|
gen2_assert_iir_is_zero(dev_priv, type##IIR); \
|
|
|
|
I915_WRITE16(type##IER, (ier_val)); \
|
|
|
|
I915_WRITE16(type##IMR, (imr_val)); \
|
|
|
|
POSTING_READ16(type##IMR); \
|
|
|
|
} while (0)
|
|
|
|
|
2014-11-05 18:48:37 +00:00
|
|
|
static void gen6_rps_irq_handler(struct drm_i915_private *dev_priv, u32 pm_iir);
|
2016-10-12 16:24:31 +00:00
|
|
|
static void gen9_guc_irq_handler(struct drm_i915_private *dev_priv, u32 pm_iir);
|
2014-11-05 18:48:37 +00:00
|
|
|
|
2015-09-23 14:15:27 +00:00
|
|
|
/* For display hotplug interrupt */
|
|
|
|
static inline void
|
|
|
|
i915_hotplug_interrupt_update_locked(struct drm_i915_private *dev_priv,
|
|
|
|
uint32_t mask,
|
|
|
|
uint32_t bits)
|
|
|
|
{
|
|
|
|
uint32_t val;
|
|
|
|
|
2017-03-02 13:28:01 +00:00
|
|
|
lockdep_assert_held(&dev_priv->irq_lock);
|
2015-09-23 14:15:27 +00:00
|
|
|
WARN_ON(bits & ~mask);
|
|
|
|
|
|
|
|
val = I915_READ(PORT_HOTPLUG_EN);
|
|
|
|
val &= ~mask;
|
|
|
|
val |= bits;
|
|
|
|
I915_WRITE(PORT_HOTPLUG_EN, val);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* i915_hotplug_interrupt_update - update hotplug interrupt enable
|
|
|
|
* @dev_priv: driver private
|
|
|
|
* @mask: bits to update
|
|
|
|
* @bits: bits to enable
|
|
|
|
* NOTE: the HPD enable bits are modified both inside and outside
|
|
|
|
* of an interrupt context. To avoid that read-modify-write cycles
|
|
|
|
* interfer, these bits are protected by a spinlock. Since this
|
|
|
|
* function is usually not called from a context where the lock is
|
|
|
|
* held already, this function acquires the lock itself. A non-locking
|
|
|
|
* version is also available.
|
|
|
|
*/
|
|
|
|
void i915_hotplug_interrupt_update(struct drm_i915_private *dev_priv,
|
|
|
|
uint32_t mask,
|
|
|
|
uint32_t bits)
|
|
|
|
{
|
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
|
|
|
i915_hotplug_interrupt_update_locked(dev_priv, mask, bits);
|
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
|
|
|
}
|
|
|
|
|
2015-08-27 20:55:58 +00:00
|
|
|
/**
|
|
|
|
* ilk_update_display_irq - update DEIMR
|
|
|
|
* @dev_priv: driver private
|
|
|
|
* @interrupt_mask: mask of interrupt bits to update
|
|
|
|
* @enabled_irq_mask: mask of interrupt bits to enable
|
|
|
|
*/
|
2015-11-23 16:06:16 +00:00
|
|
|
void ilk_update_display_irq(struct drm_i915_private *dev_priv,
|
|
|
|
uint32_t interrupt_mask,
|
|
|
|
uint32_t enabled_irq_mask)
|
2009-06-08 06:40:19 +00:00
|
|
|
{
|
2015-08-27 20:55:58 +00:00
|
|
|
uint32_t new_val;
|
|
|
|
|
2017-03-02 13:28:01 +00:00
|
|
|
lockdep_assert_held(&dev_priv->irq_lock);
|
2013-06-27 11:44:58 +00:00
|
|
|
|
2015-08-27 20:55:58 +00:00
|
|
|
WARN_ON(enabled_irq_mask & ~interrupt_mask);
|
|
|
|
|
2014-06-20 16:29:20 +00:00
|
|
|
if (WARN_ON(!intel_irqs_enabled(dev_priv)))
|
2013-08-19 16:18:09 +00:00
|
|
|
return;
|
|
|
|
|
2015-08-27 20:55:58 +00:00
|
|
|
new_val = dev_priv->irq_mask;
|
|
|
|
new_val &= ~interrupt_mask;
|
|
|
|
new_val |= (~enabled_irq_mask & interrupt_mask);
|
|
|
|
|
|
|
|
if (new_val != dev_priv->irq_mask) {
|
|
|
|
dev_priv->irq_mask = new_val;
|
2010-12-04 11:30:53 +00:00
|
|
|
I915_WRITE(DEIMR, dev_priv->irq_mask);
|
2010-11-16 15:55:10 +00:00
|
|
|
POSTING_READ(DEIMR);
|
2009-06-08 06:40:19 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-08-06 21:57:12 +00:00
|
|
|
/**
|
|
|
|
* ilk_update_gt_irq - update GTIMR
|
|
|
|
* @dev_priv: driver private
|
|
|
|
* @interrupt_mask: mask of interrupt bits to update
|
|
|
|
* @enabled_irq_mask: mask of interrupt bits to enable
|
|
|
|
*/
|
|
|
|
static void ilk_update_gt_irq(struct drm_i915_private *dev_priv,
|
|
|
|
uint32_t interrupt_mask,
|
|
|
|
uint32_t enabled_irq_mask)
|
|
|
|
{
|
2017-03-02 13:28:01 +00:00
|
|
|
lockdep_assert_held(&dev_priv->irq_lock);
|
2013-08-06 21:57:12 +00:00
|
|
|
|
2014-12-08 15:30:00 +00:00
|
|
|
WARN_ON(enabled_irq_mask & ~interrupt_mask);
|
|
|
|
|
2014-06-20 16:29:20 +00:00
|
|
|
if (WARN_ON(!intel_irqs_enabled(dev_priv)))
|
2013-08-19 16:18:09 +00:00
|
|
|
return;
|
|
|
|
|
2013-08-06 21:57:12 +00:00
|
|
|
dev_priv->gt_irq_mask &= ~interrupt_mask;
|
|
|
|
dev_priv->gt_irq_mask |= (~enabled_irq_mask & interrupt_mask);
|
|
|
|
I915_WRITE(GTIMR, dev_priv->gt_irq_mask);
|
|
|
|
}
|
|
|
|
|
2014-07-16 07:49:40 +00:00
|
|
|
void gen5_enable_gt_irq(struct drm_i915_private *dev_priv, uint32_t mask)
|
2013-08-06 21:57:12 +00:00
|
|
|
{
|
|
|
|
ilk_update_gt_irq(dev_priv, mask, mask);
|
2016-07-01 16:23:27 +00:00
|
|
|
POSTING_READ_FW(GTIMR);
|
2013-08-06 21:57:12 +00:00
|
|
|
}
|
|
|
|
|
2014-07-16 07:49:40 +00:00
|
|
|
void gen5_disable_gt_irq(struct drm_i915_private *dev_priv, uint32_t mask)
|
2013-08-06 21:57:12 +00:00
|
|
|
{
|
|
|
|
ilk_update_gt_irq(dev_priv, mask, 0);
|
|
|
|
}
|
|
|
|
|
drm/i915: Type safe register read/write
Make I915_READ and I915_WRITE more type safe by wrapping the register
offset in a struct. This should eliminate most of the fumbles we've had
with misplaced parens.
This only takes care of normal mmio registers. We could extend the idea
to other register types and define each with its own struct. That way
you wouldn't be able to accidentally pass the wrong thing to a specific
register access function.
The gpio_reg setup is probably the ugliest thing left. But I figure I'd
just leave it for now, and wait for some divine inspiration to strike
before making it nice.
As for the generated code, it's actually a bit better sometimes. Eg.
looking at i915_irq_handler(), we can see the following change:
lea 0x70024(%rdx,%rax,1),%r9d
mov $0x1,%edx
- movslq %r9d,%r9
- mov %r9,%rsi
- mov %r9,-0x58(%rbp)
- callq *0xd8(%rbx)
+ mov %r9d,%esi
+ mov %r9d,-0x48(%rbp)
callq *0xd8(%rbx)
So previously gcc thought the register offset might be signed and
decided to sign extend it, just in case. The rest appears to be
mostly just minor shuffling of instructions.
v2: i915_mmio_reg_{offset,equal,valid}() helpers added
s/_REG/_MMIO/ in the register defines
mo more switch statements left to worry about
ring_emit stuff got sorted in a prep patch
cmd parser, lrc context and w/a batch buildup also in prep patch
vgpu stuff cleaned up and moved to a prep patch
all other unrelated changes split out
v3: Rebased due to BXT DSI/BLC, MOCS, etc.
v4: Rebased due to churn, s/i915_mmio_reg_t/i915_reg_t/
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/1447853606-2751-1-git-send-email-ville.syrjala@linux.intel.com
2015-11-18 13:33:26 +00:00
|
|
|
static i915_reg_t gen6_pm_iir(struct drm_i915_private *dev_priv)
|
2014-11-05 18:48:48 +00:00
|
|
|
{
|
2017-07-18 18:28:00 +00:00
|
|
|
return INTEL_GEN(dev_priv) >= 8 ? GEN8_GT_IIR(2) : GEN6_PMIIR;
|
2014-11-05 18:48:48 +00:00
|
|
|
}
|
|
|
|
|
drm/i915: Type safe register read/write
Make I915_READ and I915_WRITE more type safe by wrapping the register
offset in a struct. This should eliminate most of the fumbles we've had
with misplaced parens.
This only takes care of normal mmio registers. We could extend the idea
to other register types and define each with its own struct. That way
you wouldn't be able to accidentally pass the wrong thing to a specific
register access function.
The gpio_reg setup is probably the ugliest thing left. But I figure I'd
just leave it for now, and wait for some divine inspiration to strike
before making it nice.
As for the generated code, it's actually a bit better sometimes. Eg.
looking at i915_irq_handler(), we can see the following change:
lea 0x70024(%rdx,%rax,1),%r9d
mov $0x1,%edx
- movslq %r9d,%r9
- mov %r9,%rsi
- mov %r9,-0x58(%rbp)
- callq *0xd8(%rbx)
+ mov %r9d,%esi
+ mov %r9d,-0x48(%rbp)
callq *0xd8(%rbx)
So previously gcc thought the register offset might be signed and
decided to sign extend it, just in case. The rest appears to be
mostly just minor shuffling of instructions.
v2: i915_mmio_reg_{offset,equal,valid}() helpers added
s/_REG/_MMIO/ in the register defines
mo more switch statements left to worry about
ring_emit stuff got sorted in a prep patch
cmd parser, lrc context and w/a batch buildup also in prep patch
vgpu stuff cleaned up and moved to a prep patch
all other unrelated changes split out
v3: Rebased due to BXT DSI/BLC, MOCS, etc.
v4: Rebased due to churn, s/i915_mmio_reg_t/i915_reg_t/
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/1447853606-2751-1-git-send-email-ville.syrjala@linux.intel.com
2015-11-18 13:33:26 +00:00
|
|
|
static i915_reg_t gen6_pm_imr(struct drm_i915_private *dev_priv)
|
2014-11-05 18:48:31 +00:00
|
|
|
{
|
2017-07-18 18:28:00 +00:00
|
|
|
return INTEL_GEN(dev_priv) >= 8 ? GEN8_GT_IMR(2) : GEN6_PMIMR;
|
2014-11-05 18:48:31 +00:00
|
|
|
}
|
|
|
|
|
drm/i915: Type safe register read/write
Make I915_READ and I915_WRITE more type safe by wrapping the register
offset in a struct. This should eliminate most of the fumbles we've had
with misplaced parens.
This only takes care of normal mmio registers. We could extend the idea
to other register types and define each with its own struct. That way
you wouldn't be able to accidentally pass the wrong thing to a specific
register access function.
The gpio_reg setup is probably the ugliest thing left. But I figure I'd
just leave it for now, and wait for some divine inspiration to strike
before making it nice.
As for the generated code, it's actually a bit better sometimes. Eg.
looking at i915_irq_handler(), we can see the following change:
lea 0x70024(%rdx,%rax,1),%r9d
mov $0x1,%edx
- movslq %r9d,%r9
- mov %r9,%rsi
- mov %r9,-0x58(%rbp)
- callq *0xd8(%rbx)
+ mov %r9d,%esi
+ mov %r9d,-0x48(%rbp)
callq *0xd8(%rbx)
So previously gcc thought the register offset might be signed and
decided to sign extend it, just in case. The rest appears to be
mostly just minor shuffling of instructions.
v2: i915_mmio_reg_{offset,equal,valid}() helpers added
s/_REG/_MMIO/ in the register defines
mo more switch statements left to worry about
ring_emit stuff got sorted in a prep patch
cmd parser, lrc context and w/a batch buildup also in prep patch
vgpu stuff cleaned up and moved to a prep patch
all other unrelated changes split out
v3: Rebased due to BXT DSI/BLC, MOCS, etc.
v4: Rebased due to churn, s/i915_mmio_reg_t/i915_reg_t/
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/1447853606-2751-1-git-send-email-ville.syrjala@linux.intel.com
2015-11-18 13:33:26 +00:00
|
|
|
static i915_reg_t gen6_pm_ier(struct drm_i915_private *dev_priv)
|
2014-11-05 18:48:48 +00:00
|
|
|
{
|
2017-07-18 18:28:00 +00:00
|
|
|
return INTEL_GEN(dev_priv) >= 8 ? GEN8_GT_IER(2) : GEN6_PMIER;
|
2014-11-05 18:48:48 +00:00
|
|
|
}
|
|
|
|
|
2013-08-06 21:57:13 +00:00
|
|
|
/**
|
2015-11-25 14:21:30 +00:00
|
|
|
* snb_update_pm_irq - update GEN6_PMIMR
|
|
|
|
* @dev_priv: driver private
|
|
|
|
* @interrupt_mask: mask of interrupt bits to update
|
|
|
|
* @enabled_irq_mask: mask of interrupt bits to enable
|
|
|
|
*/
|
2013-08-06 21:57:13 +00:00
|
|
|
static void snb_update_pm_irq(struct drm_i915_private *dev_priv,
|
|
|
|
uint32_t interrupt_mask,
|
|
|
|
uint32_t enabled_irq_mask)
|
|
|
|
{
|
2013-08-06 21:57:15 +00:00
|
|
|
uint32_t new_val;
|
2013-08-06 21:57:13 +00:00
|
|
|
|
2014-12-08 15:30:00 +00:00
|
|
|
WARN_ON(enabled_irq_mask & ~interrupt_mask);
|
|
|
|
|
2017-03-02 13:28:01 +00:00
|
|
|
lockdep_assert_held(&dev_priv->irq_lock);
|
2013-08-06 21:57:13 +00:00
|
|
|
|
2016-10-12 16:24:30 +00:00
|
|
|
new_val = dev_priv->pm_imr;
|
2013-08-06 21:57:14 +00:00
|
|
|
new_val &= ~interrupt_mask;
|
|
|
|
new_val |= (~enabled_irq_mask & interrupt_mask);
|
|
|
|
|
2016-10-12 16:24:30 +00:00
|
|
|
if (new_val != dev_priv->pm_imr) {
|
|
|
|
dev_priv->pm_imr = new_val;
|
|
|
|
I915_WRITE(gen6_pm_imr(dev_priv), dev_priv->pm_imr);
|
2014-11-05 18:48:31 +00:00
|
|
|
POSTING_READ(gen6_pm_imr(dev_priv));
|
2013-08-06 21:57:14 +00:00
|
|
|
}
|
2013-08-06 21:57:13 +00:00
|
|
|
}
|
|
|
|
|
2016-10-12 16:24:30 +00:00
|
|
|
void gen6_unmask_pm_irq(struct drm_i915_private *dev_priv, u32 mask)
|
2013-08-06 21:57:13 +00:00
|
|
|
{
|
2014-11-20 21:01:47 +00:00
|
|
|
if (WARN_ON(!intel_irqs_enabled(dev_priv)))
|
|
|
|
return;
|
|
|
|
|
2013-08-06 21:57:13 +00:00
|
|
|
snb_update_pm_irq(dev_priv, mask, mask);
|
|
|
|
}
|
|
|
|
|
2016-10-12 16:24:30 +00:00
|
|
|
static void __gen6_mask_pm_irq(struct drm_i915_private *dev_priv, u32 mask)
|
2013-08-06 21:57:13 +00:00
|
|
|
{
|
|
|
|
snb_update_pm_irq(dev_priv, mask, 0);
|
|
|
|
}
|
|
|
|
|
2016-10-12 16:24:30 +00:00
|
|
|
void gen6_mask_pm_irq(struct drm_i915_private *dev_priv, u32 mask)
|
2014-11-20 21:01:47 +00:00
|
|
|
{
|
|
|
|
if (WARN_ON(!intel_irqs_enabled(dev_priv)))
|
|
|
|
return;
|
|
|
|
|
2016-10-12 16:24:30 +00:00
|
|
|
__gen6_mask_pm_irq(dev_priv, mask);
|
2014-11-20 21:01:47 +00:00
|
|
|
}
|
|
|
|
|
2017-08-23 23:58:24 +00:00
|
|
|
static void gen6_reset_pm_iir(struct drm_i915_private *dev_priv, u32 reset_mask)
|
2014-11-19 13:30:03 +00:00
|
|
|
{
|
drm/i915: Type safe register read/write
Make I915_READ and I915_WRITE more type safe by wrapping the register
offset in a struct. This should eliminate most of the fumbles we've had
with misplaced parens.
This only takes care of normal mmio registers. We could extend the idea
to other register types and define each with its own struct. That way
you wouldn't be able to accidentally pass the wrong thing to a specific
register access function.
The gpio_reg setup is probably the ugliest thing left. But I figure I'd
just leave it for now, and wait for some divine inspiration to strike
before making it nice.
As for the generated code, it's actually a bit better sometimes. Eg.
looking at i915_irq_handler(), we can see the following change:
lea 0x70024(%rdx,%rax,1),%r9d
mov $0x1,%edx
- movslq %r9d,%r9
- mov %r9,%rsi
- mov %r9,-0x58(%rbp)
- callq *0xd8(%rbx)
+ mov %r9d,%esi
+ mov %r9d,-0x48(%rbp)
callq *0xd8(%rbx)
So previously gcc thought the register offset might be signed and
decided to sign extend it, just in case. The rest appears to be
mostly just minor shuffling of instructions.
v2: i915_mmio_reg_{offset,equal,valid}() helpers added
s/_REG/_MMIO/ in the register defines
mo more switch statements left to worry about
ring_emit stuff got sorted in a prep patch
cmd parser, lrc context and w/a batch buildup also in prep patch
vgpu stuff cleaned up and moved to a prep patch
all other unrelated changes split out
v3: Rebased due to BXT DSI/BLC, MOCS, etc.
v4: Rebased due to churn, s/i915_mmio_reg_t/i915_reg_t/
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/1447853606-2751-1-git-send-email-ville.syrjala@linux.intel.com
2015-11-18 13:33:26 +00:00
|
|
|
i915_reg_t reg = gen6_pm_iir(dev_priv);
|
2014-11-19 13:30:03 +00:00
|
|
|
|
2017-03-02 13:28:01 +00:00
|
|
|
lockdep_assert_held(&dev_priv->irq_lock);
|
2016-10-12 16:24:30 +00:00
|
|
|
|
|
|
|
I915_WRITE(reg, reset_mask);
|
|
|
|
I915_WRITE(reg, reset_mask);
|
2014-11-19 13:30:03 +00:00
|
|
|
POSTING_READ(reg);
|
2016-10-12 16:24:30 +00:00
|
|
|
}
|
|
|
|
|
2017-08-23 23:58:24 +00:00
|
|
|
static void gen6_enable_pm_irq(struct drm_i915_private *dev_priv, u32 enable_mask)
|
2016-10-12 16:24:30 +00:00
|
|
|
{
|
2017-03-02 13:28:01 +00:00
|
|
|
lockdep_assert_held(&dev_priv->irq_lock);
|
2016-10-12 16:24:30 +00:00
|
|
|
|
|
|
|
dev_priv->pm_ier |= enable_mask;
|
|
|
|
I915_WRITE(gen6_pm_ier(dev_priv), dev_priv->pm_ier);
|
|
|
|
gen6_unmask_pm_irq(dev_priv, enable_mask);
|
|
|
|
/* unmask_pm_irq provides an implicit barrier (POSTING_READ) */
|
|
|
|
}
|
|
|
|
|
2017-08-23 23:58:24 +00:00
|
|
|
static void gen6_disable_pm_irq(struct drm_i915_private *dev_priv, u32 disable_mask)
|
2016-10-12 16:24:30 +00:00
|
|
|
{
|
2017-03-02 13:28:01 +00:00
|
|
|
lockdep_assert_held(&dev_priv->irq_lock);
|
2016-10-12 16:24:30 +00:00
|
|
|
|
|
|
|
dev_priv->pm_ier &= ~disable_mask;
|
|
|
|
__gen6_mask_pm_irq(dev_priv, disable_mask);
|
|
|
|
I915_WRITE(gen6_pm_ier(dev_priv), dev_priv->pm_ier);
|
|
|
|
/* though a barrier is missing here, but don't really need a one */
|
|
|
|
}
|
|
|
|
|
|
|
|
void gen6_reset_rps_interrupts(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
|
|
|
gen6_reset_pm_iir(dev_priv, dev_priv->pm_rps_events);
|
2015-03-23 17:11:35 +00:00
|
|
|
dev_priv->rps.pm_iir = 0;
|
2014-11-19 13:30:03 +00:00
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
void gen6_enable_rps_interrupts(struct drm_i915_private *dev_priv)
|
2014-11-05 18:48:48 +00:00
|
|
|
{
|
2016-09-21 13:51:06 +00:00
|
|
|
if (READ_ONCE(dev_priv->rps.interrupts_enabled))
|
|
|
|
return;
|
|
|
|
|
2014-11-05 18:48:48 +00:00
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
2016-07-04 07:08:36 +00:00
|
|
|
WARN_ON_ONCE(dev_priv->rps.pm_iir);
|
|
|
|
WARN_ON_ONCE(I915_READ(gen6_pm_iir(dev_priv)) & dev_priv->pm_rps_events);
|
drm/i915: sanitize rps irq disabling
When disabling the RPS interrupts there is a tricky dependency between
the thread disabling the interrupts, the RPS interrupt handler and the
corresponding RPS work. The RPS work can reenable the interrupts, so
there is no straightforward order in the disabling thread to (1) make
sure that any RPS work is flushed and to (2) disable all RPS
interrupts. Currently this is solved by masking the interrupts using two
separate mask registers (first level display IMR and PM IMR) and doing
the disabling when all first level interrupts are disabled.
This works, but the requirement to run with all first level interrupts
disabled is unnecessary making the suspend / unload time ordering of RPS
disabling wrt. other unitialization steps difficult and error prone.
Removing this restriction allows us to disable RPS early during suspend
/ unload and forget about it for the rest of the sequence. By adding a
more explicit method for avoiding the above race, it also becomes easier
to prove its correctness. Finally currently we can hit the WARN in
snb_update_pm_irq(), when a final RPS work runs with the first level
interrupts already disabled. This won't lead to any problem (due to the
separate interrupt masks), but with the change in this and the next
patch we can get rid of the WARN, while leaving it in place for other
scenarios.
To address the above points, add a new RPS interrupts_enabled flag and
use this during RPS disabling to avoid requeuing the RPS work and
reenabling of the RPS interrupts. Since the interrupt disabling happens
now in intel_suspend_gt_powersave(), we will disable RPS interrupts
explicitly during suspend (and not just through the first level mask),
but there is no problem doing so, it's also more consistent and allows
us to unify more of the RPS disabling during suspend and unload time in
the next patch.
v2/v3:
- rebase on patch "drm/i915: move rps irq disable one level up" in the
patchset
Signed-off-by: Imre Deak <imre.deak@intel.com>
Reviewed-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2014-11-19 13:30:04 +00:00
|
|
|
dev_priv->rps.interrupts_enabled = true;
|
2014-11-05 18:48:48 +00:00
|
|
|
gen6_enable_pm_irq(dev_priv, dev_priv->pm_rps_events);
|
2014-12-15 16:59:27 +00:00
|
|
|
|
2014-11-05 18:48:48 +00:00
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
void gen6_disable_rps_interrupts(struct drm_i915_private *dev_priv)
|
2014-11-05 18:48:48 +00:00
|
|
|
{
|
2016-09-21 13:51:06 +00:00
|
|
|
if (!READ_ONCE(dev_priv->rps.interrupts_enabled))
|
|
|
|
return;
|
|
|
|
|
drm/i915: sanitize rps irq disabling
When disabling the RPS interrupts there is a tricky dependency between
the thread disabling the interrupts, the RPS interrupt handler and the
corresponding RPS work. The RPS work can reenable the interrupts, so
there is no straightforward order in the disabling thread to (1) make
sure that any RPS work is flushed and to (2) disable all RPS
interrupts. Currently this is solved by masking the interrupts using two
separate mask registers (first level display IMR and PM IMR) and doing
the disabling when all first level interrupts are disabled.
This works, but the requirement to run with all first level interrupts
disabled is unnecessary making the suspend / unload time ordering of RPS
disabling wrt. other unitialization steps difficult and error prone.
Removing this restriction allows us to disable RPS early during suspend
/ unload and forget about it for the rest of the sequence. By adding a
more explicit method for avoiding the above race, it also becomes easier
to prove its correctness. Finally currently we can hit the WARN in
snb_update_pm_irq(), when a final RPS work runs with the first level
interrupts already disabled. This won't lead to any problem (due to the
separate interrupt masks), but with the change in this and the next
patch we can get rid of the WARN, while leaving it in place for other
scenarios.
To address the above points, add a new RPS interrupts_enabled flag and
use this during RPS disabling to avoid requeuing the RPS work and
reenabling of the RPS interrupts. Since the interrupt disabling happens
now in intel_suspend_gt_powersave(), we will disable RPS interrupts
explicitly during suspend (and not just through the first level mask),
but there is no problem doing so, it's also more consistent and allows
us to unify more of the RPS disabling during suspend and unload time in
the next patch.
v2/v3:
- rebase on patch "drm/i915: move rps irq disable one level up" in the
patchset
Signed-off-by: Imre Deak <imre.deak@intel.com>
Reviewed-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2014-11-19 13:30:04 +00:00
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
|
|
|
dev_priv->rps.interrupts_enabled = false;
|
2014-11-20 21:01:47 +00:00
|
|
|
|
2016-09-12 20:19:35 +00:00
|
|
|
I915_WRITE(GEN6_PMINTRMSK, gen6_sanitize_rps_pm_mask(dev_priv, ~0u));
|
2014-11-20 21:01:47 +00:00
|
|
|
|
2016-10-12 16:24:30 +00:00
|
|
|
gen6_disable_pm_irq(dev_priv, dev_priv->pm_rps_events);
|
2015-03-23 17:11:34 +00:00
|
|
|
|
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
2016-07-05 09:40:23 +00:00
|
|
|
synchronize_irq(dev_priv->drm.irq);
|
2016-07-04 07:08:36 +00:00
|
|
|
|
|
|
|
/* Now that we will not be generating any more work, flush any
|
2017-08-23 23:58:24 +00:00
|
|
|
* outstanding tasks. As we are called on the RPS idle path,
|
2016-07-04 07:08:36 +00:00
|
|
|
* we will reset the GPU to minimum frequencies, so the current
|
|
|
|
* state of the worker can be discarded.
|
|
|
|
*/
|
|
|
|
cancel_work_sync(&dev_priv->rps.work);
|
|
|
|
gen6_reset_rps_interrupts(dev_priv);
|
2014-11-05 18:48:48 +00:00
|
|
|
}
|
|
|
|
|
2016-10-12 16:24:31 +00:00
|
|
|
void gen9_reset_guc_interrupts(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
|
|
|
gen6_reset_pm_iir(dev_priv, dev_priv->pm_guc_events);
|
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
void gen9_enable_guc_interrupts(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
|
|
|
if (!dev_priv->guc.interrupts_enabled) {
|
|
|
|
WARN_ON_ONCE(I915_READ(gen6_pm_iir(dev_priv)) &
|
|
|
|
dev_priv->pm_guc_events);
|
|
|
|
dev_priv->guc.interrupts_enabled = true;
|
|
|
|
gen6_enable_pm_irq(dev_priv, dev_priv->pm_guc_events);
|
|
|
|
}
|
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
void gen9_disable_guc_interrupts(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
|
|
|
dev_priv->guc.interrupts_enabled = false;
|
|
|
|
|
|
|
|
gen6_disable_pm_irq(dev_priv, dev_priv->pm_guc_events);
|
|
|
|
|
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
|
|
|
synchronize_irq(dev_priv->drm.irq);
|
|
|
|
|
|
|
|
gen9_reset_guc_interrupts(dev_priv);
|
|
|
|
}
|
|
|
|
|
2015-08-27 20:56:06 +00:00
|
|
|
/**
|
2015-11-25 14:21:30 +00:00
|
|
|
* bdw_update_port_irq - update DE port interrupt
|
|
|
|
* @dev_priv: driver private
|
|
|
|
* @interrupt_mask: mask of interrupt bits to update
|
|
|
|
* @enabled_irq_mask: mask of interrupt bits to enable
|
|
|
|
*/
|
2015-08-27 20:56:06 +00:00
|
|
|
static void bdw_update_port_irq(struct drm_i915_private *dev_priv,
|
|
|
|
uint32_t interrupt_mask,
|
|
|
|
uint32_t enabled_irq_mask)
|
|
|
|
{
|
|
|
|
uint32_t new_val;
|
|
|
|
uint32_t old_val;
|
|
|
|
|
2017-03-02 13:28:01 +00:00
|
|
|
lockdep_assert_held(&dev_priv->irq_lock);
|
2015-08-27 20:56:06 +00:00
|
|
|
|
|
|
|
WARN_ON(enabled_irq_mask & ~interrupt_mask);
|
|
|
|
|
|
|
|
if (WARN_ON(!intel_irqs_enabled(dev_priv)))
|
|
|
|
return;
|
|
|
|
|
|
|
|
old_val = I915_READ(GEN8_DE_PORT_IMR);
|
|
|
|
|
|
|
|
new_val = old_val;
|
|
|
|
new_val &= ~interrupt_mask;
|
|
|
|
new_val |= (~enabled_irq_mask & interrupt_mask);
|
|
|
|
|
|
|
|
if (new_val != old_val) {
|
|
|
|
I915_WRITE(GEN8_DE_PORT_IMR, new_val);
|
|
|
|
POSTING_READ(GEN8_DE_PORT_IMR);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-11-23 16:06:17 +00:00
|
|
|
/**
|
|
|
|
* bdw_update_pipe_irq - update DE pipe interrupt
|
|
|
|
* @dev_priv: driver private
|
|
|
|
* @pipe: pipe whose interrupt to update
|
|
|
|
* @interrupt_mask: mask of interrupt bits to update
|
|
|
|
* @enabled_irq_mask: mask of interrupt bits to enable
|
|
|
|
*/
|
|
|
|
void bdw_update_pipe_irq(struct drm_i915_private *dev_priv,
|
|
|
|
enum pipe pipe,
|
|
|
|
uint32_t interrupt_mask,
|
|
|
|
uint32_t enabled_irq_mask)
|
|
|
|
{
|
|
|
|
uint32_t new_val;
|
|
|
|
|
2017-03-02 13:28:01 +00:00
|
|
|
lockdep_assert_held(&dev_priv->irq_lock);
|
2015-11-23 16:06:17 +00:00
|
|
|
|
|
|
|
WARN_ON(enabled_irq_mask & ~interrupt_mask);
|
|
|
|
|
|
|
|
if (WARN_ON(!intel_irqs_enabled(dev_priv)))
|
|
|
|
return;
|
|
|
|
|
|
|
|
new_val = dev_priv->de_irq_mask[pipe];
|
|
|
|
new_val &= ~interrupt_mask;
|
|
|
|
new_val |= (~enabled_irq_mask & interrupt_mask);
|
|
|
|
|
|
|
|
if (new_val != dev_priv->de_irq_mask[pipe]) {
|
|
|
|
dev_priv->de_irq_mask[pipe] = new_val;
|
|
|
|
I915_WRITE(GEN8_DE_PIPE_IMR(pipe), dev_priv->de_irq_mask[pipe]);
|
|
|
|
POSTING_READ(GEN8_DE_PIPE_IMR(pipe));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-07-04 21:35:21 +00:00
|
|
|
/**
|
|
|
|
* ibx_display_interrupt_update - update SDEIMR
|
|
|
|
* @dev_priv: driver private
|
|
|
|
* @interrupt_mask: mask of interrupt bits to update
|
|
|
|
* @enabled_irq_mask: mask of interrupt bits to enable
|
|
|
|
*/
|
2014-09-30 08:56:46 +00:00
|
|
|
void ibx_display_interrupt_update(struct drm_i915_private *dev_priv,
|
|
|
|
uint32_t interrupt_mask,
|
|
|
|
uint32_t enabled_irq_mask)
|
2013-07-04 21:35:21 +00:00
|
|
|
{
|
|
|
|
uint32_t sdeimr = I915_READ(SDEIMR);
|
|
|
|
sdeimr &= ~interrupt_mask;
|
|
|
|
sdeimr |= (~enabled_irq_mask & interrupt_mask);
|
|
|
|
|
2014-12-08 15:30:00 +00:00
|
|
|
WARN_ON(enabled_irq_mask & ~interrupt_mask);
|
|
|
|
|
2017-03-02 13:28:01 +00:00
|
|
|
lockdep_assert_held(&dev_priv->irq_lock);
|
2013-07-04 21:35:21 +00:00
|
|
|
|
2014-06-20 16:29:20 +00:00
|
|
|
if (WARN_ON(!intel_irqs_enabled(dev_priv)))
|
2013-08-19 16:18:09 +00:00
|
|
|
return;
|
|
|
|
|
2013-07-04 21:35:21 +00:00
|
|
|
I915_WRITE(SDEIMR, sdeimr);
|
|
|
|
POSTING_READ(SDEIMR);
|
|
|
|
}
|
drm/i915: report Gen5+ CPU and PCH FIFO underruns
In this commit we enable both CPU and PCH FIFO underrun reporting and
start reporting them. We follow a few rules:
- after we receive one of these errors, we mask the interrupt, so
we won't get an "interrupt storm" and we also won't flood dmesg;
- at each mode set we enable the interrupts again, so we'll see each
message at most once per mode set;
- in the specific places where we need to ignore the errors, we
completely mask the interrupts.
The downside of this patch is that since we're completely disabling
(masking) the interrupts instead of just not printing error messages,
we will mask more than just what we want on IVB/HSW CPU interrupts
(due to GEN7_ERR_INT) and on CPT/PPT/LPT PCHs (due to SERR_INT). So
when we decide to mask PCH FIFO underruns for pipe A on CPT, we'll
also be masking PCH FIFO underruns for pipe B, because both are
reported by SERR_INT, which has to be either completely enabled or
completely disabled (in othe words, there's no way to disable/enable
specific bits of GEN7_ERR_INT and SERR_INT).
V2: Rename some functions and variables, downgrade messages to
DRM_DEBUG_DRIVER and rebase.
Signed-off-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Reviewed-by: Imre Deak <imre.deak@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-04-12 20:57:57 +00:00
|
|
|
|
2017-09-14 15:17:31 +00:00
|
|
|
u32 i915_pipestat_enable_mask(struct drm_i915_private *dev_priv,
|
|
|
|
enum pipe pipe)
|
2008-11-04 10:03:27 +00:00
|
|
|
{
|
2017-09-14 15:17:31 +00:00
|
|
|
u32 status_mask = dev_priv->pipestat_irq_mask[pipe];
|
|
|
|
u32 enable_mask = status_mask << 16;
|
2008-11-04 10:03:27 +00:00
|
|
|
|
2017-03-02 13:28:01 +00:00
|
|
|
lockdep_assert_held(&dev_priv->irq_lock);
|
2008-11-04 10:03:27 +00:00
|
|
|
|
2017-09-14 15:17:31 +00:00
|
|
|
if (INTEL_GEN(dev_priv) < 5)
|
|
|
|
goto out;
|
2014-02-10 16:42:48 +00:00
|
|
|
|
|
|
|
/*
|
2014-04-09 10:28:48 +00:00
|
|
|
* On pipe A we don't support the PSR interrupt yet,
|
|
|
|
* on pipe B and C the same bit MBZ.
|
2014-02-10 16:42:48 +00:00
|
|
|
*/
|
|
|
|
if (WARN_ON_ONCE(status_mask & PIPE_A_PSR_STATUS_VLV))
|
|
|
|
return 0;
|
2014-04-09 10:28:48 +00:00
|
|
|
/*
|
|
|
|
* On pipe B and C we don't support the PSR interrupt yet, on pipe
|
|
|
|
* A the same bit is for perf counters which we don't use either.
|
|
|
|
*/
|
|
|
|
if (WARN_ON_ONCE(status_mask & PIPE_B_PSR_STATUS_VLV))
|
|
|
|
return 0;
|
2014-02-10 16:42:48 +00:00
|
|
|
|
|
|
|
enable_mask &= ~(PIPE_FIFO_UNDERRUN_STATUS |
|
|
|
|
SPRITE0_FLIP_DONE_INT_EN_VLV |
|
|
|
|
SPRITE1_FLIP_DONE_INT_EN_VLV);
|
|
|
|
if (status_mask & SPRITE0_FLIP_DONE_INT_STATUS_VLV)
|
|
|
|
enable_mask |= SPRITE0_FLIP_DONE_INT_EN_VLV;
|
|
|
|
if (status_mask & SPRITE1_FLIP_DONE_INT_STATUS_VLV)
|
|
|
|
enable_mask |= SPRITE1_FLIP_DONE_INT_EN_VLV;
|
|
|
|
|
2017-09-14 15:17:31 +00:00
|
|
|
out:
|
|
|
|
WARN_ONCE(enable_mask & ~PIPESTAT_INT_ENABLE_MASK ||
|
|
|
|
status_mask & ~PIPESTAT_INT_STATUS_MASK,
|
|
|
|
"pipe %c: enable_mask=0x%x, status_mask=0x%x\n",
|
|
|
|
pipe_name(pipe), enable_mask, status_mask);
|
|
|
|
|
2014-02-10 16:42:48 +00:00
|
|
|
return enable_mask;
|
|
|
|
}
|
|
|
|
|
2017-09-14 15:17:31 +00:00
|
|
|
void i915_enable_pipestat(struct drm_i915_private *dev_priv,
|
|
|
|
enum pipe pipe, u32 status_mask)
|
2014-02-10 16:42:47 +00:00
|
|
|
{
|
2017-09-14 15:17:31 +00:00
|
|
|
i915_reg_t reg = PIPESTAT(pipe);
|
2014-02-10 16:42:47 +00:00
|
|
|
u32 enable_mask;
|
|
|
|
|
2017-09-14 15:17:31 +00:00
|
|
|
WARN_ONCE(status_mask & ~PIPESTAT_INT_STATUS_MASK,
|
|
|
|
"pipe %c: status_mask=0x%x\n",
|
|
|
|
pipe_name(pipe), status_mask);
|
|
|
|
|
|
|
|
lockdep_assert_held(&dev_priv->irq_lock);
|
|
|
|
WARN_ON(!intel_irqs_enabled(dev_priv));
|
|
|
|
|
|
|
|
if ((dev_priv->pipestat_irq_mask[pipe] & status_mask) == status_mask)
|
|
|
|
return;
|
|
|
|
|
|
|
|
dev_priv->pipestat_irq_mask[pipe] |= status_mask;
|
|
|
|
enable_mask = i915_pipestat_enable_mask(dev_priv, pipe);
|
|
|
|
|
|
|
|
I915_WRITE(reg, enable_mask | status_mask);
|
|
|
|
POSTING_READ(reg);
|
2014-02-10 16:42:47 +00:00
|
|
|
}
|
|
|
|
|
2017-09-14 15:17:31 +00:00
|
|
|
void i915_disable_pipestat(struct drm_i915_private *dev_priv,
|
|
|
|
enum pipe pipe, u32 status_mask)
|
2014-02-10 16:42:47 +00:00
|
|
|
{
|
2017-09-14 15:17:31 +00:00
|
|
|
i915_reg_t reg = PIPESTAT(pipe);
|
2014-02-10 16:42:47 +00:00
|
|
|
u32 enable_mask;
|
|
|
|
|
2017-09-14 15:17:31 +00:00
|
|
|
WARN_ONCE(status_mask & ~PIPESTAT_INT_STATUS_MASK,
|
|
|
|
"pipe %c: status_mask=0x%x\n",
|
|
|
|
pipe_name(pipe), status_mask);
|
|
|
|
|
|
|
|
lockdep_assert_held(&dev_priv->irq_lock);
|
|
|
|
WARN_ON(!intel_irqs_enabled(dev_priv));
|
|
|
|
|
|
|
|
if ((dev_priv->pipestat_irq_mask[pipe] & status_mask) == 0)
|
|
|
|
return;
|
|
|
|
|
|
|
|
dev_priv->pipestat_irq_mask[pipe] &= ~status_mask;
|
|
|
|
enable_mask = i915_pipestat_enable_mask(dev_priv, pipe);
|
|
|
|
|
|
|
|
I915_WRITE(reg, enable_mask | status_mask);
|
|
|
|
POSTING_READ(reg);
|
2014-02-10 16:42:47 +00:00
|
|
|
}
|
|
|
|
|
2009-10-28 05:10:00 +00:00
|
|
|
/**
|
2013-04-29 10:02:54 +00:00
|
|
|
* i915_enable_asle_pipestat - enable ASLE pipestat for OpRegion
|
2016-06-03 13:02:17 +00:00
|
|
|
* @dev_priv: i915 device private
|
2009-10-28 05:10:00 +00:00
|
|
|
*/
|
2016-05-06 13:48:28 +00:00
|
|
|
static void i915_enable_asle_pipestat(struct drm_i915_private *dev_priv)
|
2009-10-28 05:10:00 +00:00
|
|
|
{
|
2016-05-06 13:48:28 +00:00
|
|
|
if (!dev_priv->opregion.asle || !IS_MOBILE(dev_priv))
|
2013-04-29 10:02:54 +00:00
|
|
|
return;
|
|
|
|
|
2014-09-15 12:55:29 +00:00
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
2009-10-28 05:10:00 +00:00
|
|
|
|
2014-02-10 16:42:47 +00:00
|
|
|
i915_enable_pipestat(dev_priv, PIPE_B, PIPE_LEGACY_BLC_EVENT_STATUS);
|
2016-05-06 13:48:28 +00:00
|
|
|
if (INTEL_GEN(dev_priv) >= 4)
|
2013-10-21 16:04:35 +00:00
|
|
|
i915_enable_pipestat(dev_priv, PIPE_A,
|
2014-02-10 16:42:47 +00:00
|
|
|
PIPE_LEGACY_BLC_EVENT_STATUS);
|
2010-12-04 11:30:53 +00:00
|
|
|
|
2014-09-15 12:55:29 +00:00
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
2009-10-28 05:10:00 +00:00
|
|
|
}
|
|
|
|
|
2014-05-15 17:20:36 +00:00
|
|
|
/*
|
|
|
|
* This timing diagram depicts the video signal in and
|
|
|
|
* around the vertical blanking period.
|
|
|
|
*
|
|
|
|
* Assumptions about the fictitious mode used in this example:
|
|
|
|
* vblank_start >= 3
|
|
|
|
* vsync_start = vblank_start + 1
|
|
|
|
* vsync_end = vblank_start + 2
|
|
|
|
* vtotal = vblank_start + 3
|
|
|
|
*
|
|
|
|
* start of vblank:
|
|
|
|
* latch double buffered registers
|
|
|
|
* increment frame counter (ctg+)
|
|
|
|
* generate start of vblank interrupt (gen4+)
|
|
|
|
* |
|
|
|
|
* | frame start:
|
|
|
|
* | generate frame start interrupt (aka. vblank interrupt) (gmch)
|
|
|
|
* | may be shifted forward 1-3 extra lines via PIPECONF
|
|
|
|
* | |
|
|
|
|
* | | start of vsync:
|
|
|
|
* | | generate vsync interrupt
|
|
|
|
* | | |
|
|
|
|
* ___xxxx___ ___xxxx___ ___xxxx___ ___xxxx___ ___xxxx___ ___xxxx
|
|
|
|
* . \hs/ . \hs/ \hs/ \hs/ . \hs/
|
|
|
|
* ----va---> <-----------------vb--------------------> <--------va-------------
|
|
|
|
* | | <----vs-----> |
|
|
|
|
* -vbs-----> <---vbs+1---> <---vbs+2---> <-----0-----> <-----1-----> <-----2--- (scanline counter gen2)
|
|
|
|
* -vbs-2---> <---vbs-1---> <---vbs-----> <---vbs+1---> <---vbs+2---> <-----0--- (scanline counter gen3+)
|
|
|
|
* -vbs-2---> <---vbs-2---> <---vbs-1---> <---vbs-----> <---vbs+1---> <---vbs+2- (scanline counter hsw+ hdmi)
|
|
|
|
* | | |
|
|
|
|
* last visible pixel first visible pixel
|
|
|
|
* | increment frame counter (gen3/4)
|
|
|
|
* pixel counter = vblank_start * htotal pixel counter = 0 (gen3/4)
|
|
|
|
*
|
|
|
|
* x = horizontal active
|
|
|
|
* _ = horizontal blanking
|
|
|
|
* hs = horizontal sync
|
|
|
|
* va = vertical active
|
|
|
|
* vb = vertical blanking
|
|
|
|
* vs = vertical sync
|
|
|
|
* vbs = vblank_start (number)
|
|
|
|
*
|
|
|
|
* Summary:
|
|
|
|
* - most events happen at the start of horizontal sync
|
|
|
|
* - frame start happens at the start of horizontal blank, 1-4 lines
|
|
|
|
* (depending on PIPECONF settings) after the start of vblank
|
|
|
|
* - gen3/4 pixel and frame counter are synchronized with the start
|
|
|
|
* of horizontal active on the first line of vertical active
|
|
|
|
*/
|
|
|
|
|
2008-10-19 02:39:29 +00:00
|
|
|
/* Called from drm generic code, passed a 'crtc', which
|
|
|
|
* we use as a pipe index
|
|
|
|
*/
|
2015-09-24 16:35:31 +00:00
|
|
|
static u32 i915_get_vblank_counter(struct drm_device *dev, unsigned int pipe)
|
2008-09-30 19:14:26 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
drm/i915: Type safe register read/write
Make I915_READ and I915_WRITE more type safe by wrapping the register
offset in a struct. This should eliminate most of the fumbles we've had
with misplaced parens.
This only takes care of normal mmio registers. We could extend the idea
to other register types and define each with its own struct. That way
you wouldn't be able to accidentally pass the wrong thing to a specific
register access function.
The gpio_reg setup is probably the ugliest thing left. But I figure I'd
just leave it for now, and wait for some divine inspiration to strike
before making it nice.
As for the generated code, it's actually a bit better sometimes. Eg.
looking at i915_irq_handler(), we can see the following change:
lea 0x70024(%rdx,%rax,1),%r9d
mov $0x1,%edx
- movslq %r9d,%r9
- mov %r9,%rsi
- mov %r9,-0x58(%rbp)
- callq *0xd8(%rbx)
+ mov %r9d,%esi
+ mov %r9d,-0x48(%rbp)
callq *0xd8(%rbx)
So previously gcc thought the register offset might be signed and
decided to sign extend it, just in case. The rest appears to be
mostly just minor shuffling of instructions.
v2: i915_mmio_reg_{offset,equal,valid}() helpers added
s/_REG/_MMIO/ in the register defines
mo more switch statements left to worry about
ring_emit stuff got sorted in a prep patch
cmd parser, lrc context and w/a batch buildup also in prep patch
vgpu stuff cleaned up and moved to a prep patch
all other unrelated changes split out
v3: Rebased due to BXT DSI/BLC, MOCS, etc.
v4: Rebased due to churn, s/i915_mmio_reg_t/i915_reg_t/
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/1447853606-2751-1-git-send-email-ville.syrjala@linux.intel.com
2015-11-18 13:33:26 +00:00
|
|
|
i915_reg_t high_frame, low_frame;
|
2014-04-29 10:35:50 +00:00
|
|
|
u32 high1, high2, low, pixel, vbl_start, hsync_start, htotal;
|
2017-05-09 14:03:29 +00:00
|
|
|
const struct drm_display_mode *mode = &dev->vblank[pipe].hwmode;
|
2017-03-09 15:44:30 +00:00
|
|
|
unsigned long irqflags;
|
2008-09-30 19:14:26 +00:00
|
|
|
|
2015-02-13 20:03:44 +00:00
|
|
|
htotal = mode->crtc_htotal;
|
|
|
|
hsync_start = mode->crtc_hsync_start;
|
|
|
|
vbl_start = mode->crtc_vblank_start;
|
|
|
|
if (mode->flags & DRM_MODE_FLAG_INTERLACE)
|
|
|
|
vbl_start = DIV_ROUND_UP(vbl_start, 2);
|
2013-09-25 16:55:26 +00:00
|
|
|
|
2014-04-29 10:35:50 +00:00
|
|
|
/* Convert to pixel count */
|
|
|
|
vbl_start *= htotal;
|
|
|
|
|
|
|
|
/* Start of vblank event occurs at start of hsync */
|
|
|
|
vbl_start -= htotal - hsync_start;
|
|
|
|
|
2011-02-07 20:26:52 +00:00
|
|
|
high_frame = PIPEFRAME(pipe);
|
|
|
|
low_frame = PIPEFRAMEPIXEL(pipe);
|
2010-09-11 12:48:45 +00:00
|
|
|
|
2017-03-09 15:44:30 +00:00
|
|
|
spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
|
|
|
|
|
2008-09-30 19:14:26 +00:00
|
|
|
/*
|
|
|
|
* High & low register fields aren't synchronized, so make sure
|
|
|
|
* we get a low value that's stable across two reads of the high
|
|
|
|
* register.
|
|
|
|
*/
|
|
|
|
do {
|
2017-03-09 15:44:30 +00:00
|
|
|
high1 = I915_READ_FW(high_frame) & PIPE_FRAME_HIGH_MASK;
|
|
|
|
low = I915_READ_FW(low_frame);
|
|
|
|
high2 = I915_READ_FW(high_frame) & PIPE_FRAME_HIGH_MASK;
|
2008-09-30 19:14:26 +00:00
|
|
|
} while (high1 != high2);
|
|
|
|
|
2017-03-09 15:44:30 +00:00
|
|
|
spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
|
|
|
|
|
2010-09-11 12:48:45 +00:00
|
|
|
high1 >>= PIPE_FRAME_HIGH_SHIFT;
|
2013-09-25 16:55:26 +00:00
|
|
|
pixel = low & PIPE_PIXEL_MASK;
|
2010-09-11 12:48:45 +00:00
|
|
|
low >>= PIPE_FRAME_LOW_SHIFT;
|
2013-09-25 16:55:26 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* The frame counter increments at beginning of active.
|
|
|
|
* Cook up a vblank counter by also checking the pixel
|
|
|
|
* counter against vblank start.
|
|
|
|
*/
|
2013-11-06 15:56:27 +00:00
|
|
|
return (((high1 << 8) | low) + (pixel >= vbl_start)) & 0xffffff;
|
2008-09-30 19:14:26 +00:00
|
|
|
}
|
|
|
|
|
2015-10-29 23:45:33 +00:00
|
|
|
static u32 g4x_get_vblank_counter(struct drm_device *dev, unsigned int pipe)
|
2009-02-06 18:22:41 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2009-02-06 18:22:41 +00:00
|
|
|
|
2015-09-22 16:50:01 +00:00
|
|
|
return I915_READ(PIPE_FRMCOUNT_G4X(pipe));
|
2009-02-06 18:22:41 +00:00
|
|
|
}
|
|
|
|
|
2017-09-25 13:56:01 +00:00
|
|
|
/*
|
|
|
|
* On certain encoders on certain platforms, pipe
|
|
|
|
* scanline register will not work to get the scanline,
|
|
|
|
* since the timings are driven from the PORT or issues
|
|
|
|
* with scanline register updates.
|
|
|
|
* This function will use Framestamp and current
|
|
|
|
* timestamp registers to calculate the scanline.
|
|
|
|
*/
|
|
|
|
static u32 __intel_get_crtc_scanline_from_timestamp(struct intel_crtc *crtc)
|
|
|
|
{
|
|
|
|
struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
|
|
|
|
struct drm_vblank_crtc *vblank =
|
|
|
|
&crtc->base.dev->vblank[drm_crtc_index(&crtc->base)];
|
|
|
|
const struct drm_display_mode *mode = &vblank->hwmode;
|
|
|
|
u32 vblank_start = mode->crtc_vblank_start;
|
|
|
|
u32 vtotal = mode->crtc_vtotal;
|
|
|
|
u32 htotal = mode->crtc_htotal;
|
|
|
|
u32 clock = mode->crtc_clock;
|
|
|
|
u32 scanline, scan_prev_time, scan_curr_time, scan_post_time;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* To avoid the race condition where we might cross into the
|
|
|
|
* next vblank just between the PIPE_FRMTMSTMP and TIMESTAMP_CTR
|
|
|
|
* reads. We make sure we read PIPE_FRMTMSTMP and TIMESTAMP_CTR
|
|
|
|
* during the same frame.
|
|
|
|
*/
|
|
|
|
do {
|
|
|
|
/*
|
|
|
|
* This field provides read back of the display
|
|
|
|
* pipe frame time stamp. The time stamp value
|
|
|
|
* is sampled at every start of vertical blank.
|
|
|
|
*/
|
|
|
|
scan_prev_time = I915_READ_FW(PIPE_FRMTMSTMP(crtc->pipe));
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The TIMESTAMP_CTR register has the current
|
|
|
|
* time stamp value.
|
|
|
|
*/
|
|
|
|
scan_curr_time = I915_READ_FW(IVB_TIMESTAMP_CTR);
|
|
|
|
|
|
|
|
scan_post_time = I915_READ_FW(PIPE_FRMTMSTMP(crtc->pipe));
|
|
|
|
} while (scan_post_time != scan_prev_time);
|
|
|
|
|
|
|
|
scanline = div_u64(mul_u32_u32(scan_curr_time - scan_prev_time,
|
|
|
|
clock), 1000 * htotal);
|
|
|
|
scanline = min(scanline, vtotal - 1);
|
|
|
|
scanline = (scanline + vblank_start) % vtotal;
|
|
|
|
|
|
|
|
return scanline;
|
|
|
|
}
|
|
|
|
|
2015-10-22 12:34:56 +00:00
|
|
|
/* I915_READ_FW, only for fast reads of display block, no need for forcewake etc. */
|
2014-04-29 10:35:45 +00:00
|
|
|
static int __intel_get_crtc_scanline(struct intel_crtc *crtc)
|
|
|
|
{
|
|
|
|
struct drm_device *dev = crtc->base.dev;
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2017-05-09 14:03:29 +00:00
|
|
|
const struct drm_display_mode *mode;
|
|
|
|
struct drm_vblank_crtc *vblank;
|
2014-04-29 10:35:45 +00:00
|
|
|
enum pipe pipe = crtc->pipe;
|
2014-05-15 17:23:23 +00:00
|
|
|
int position, vtotal;
|
2014-04-29 10:35:45 +00:00
|
|
|
|
2017-03-02 17:15:05 +00:00
|
|
|
if (!crtc->active)
|
|
|
|
return -1;
|
|
|
|
|
2017-05-09 14:03:29 +00:00
|
|
|
vblank = &crtc->base.dev->vblank[drm_crtc_index(&crtc->base)];
|
|
|
|
mode = &vblank->hwmode;
|
|
|
|
|
2017-09-25 13:56:01 +00:00
|
|
|
if (mode->private_flags & I915_MODE_FLAG_GET_SCANLINE_FROM_TIMESTAMP)
|
|
|
|
return __intel_get_crtc_scanline_from_timestamp(crtc);
|
|
|
|
|
2014-05-15 17:23:23 +00:00
|
|
|
vtotal = mode->crtc_vtotal;
|
2014-04-29 10:35:45 +00:00
|
|
|
if (mode->flags & DRM_MODE_FLAG_INTERLACE)
|
|
|
|
vtotal /= 2;
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
if (IS_GEN2(dev_priv))
|
2015-10-22 12:34:56 +00:00
|
|
|
position = I915_READ_FW(PIPEDSL(pipe)) & DSL_LINEMASK_GEN2;
|
2014-04-29 10:35:45 +00:00
|
|
|
else
|
2015-10-22 12:34:56 +00:00
|
|
|
position = I915_READ_FW(PIPEDSL(pipe)) & DSL_LINEMASK_GEN3;
|
2014-04-29 10:35:45 +00:00
|
|
|
|
2015-09-22 19:15:54 +00:00
|
|
|
/*
|
|
|
|
* On HSW, the DSL reg (0x70000) appears to return 0 if we
|
|
|
|
* read it just before the start of vblank. So try it again
|
|
|
|
* so we don't accidentally end up spanning a vblank frame
|
|
|
|
* increment, causing the pipe_update_end() code to squak at us.
|
|
|
|
*
|
|
|
|
* The nature of this problem means we can't simply check the ISR
|
|
|
|
* bit and return the vblank start value; nor can we use the scanline
|
|
|
|
* debug register in the transcoder as it appears to have the same
|
|
|
|
* problem. We may need to extend this to include other platforms,
|
|
|
|
* but so far testing only shows the problem on HSW.
|
|
|
|
*/
|
2016-05-06 13:48:28 +00:00
|
|
|
if (HAS_DDI(dev_priv) && !position) {
|
2015-09-22 19:15:54 +00:00
|
|
|
int i, temp;
|
|
|
|
|
|
|
|
for (i = 0; i < 100; i++) {
|
|
|
|
udelay(1);
|
2017-03-09 15:44:31 +00:00
|
|
|
temp = I915_READ_FW(PIPEDSL(pipe)) & DSL_LINEMASK_GEN3;
|
2015-09-22 19:15:54 +00:00
|
|
|
if (temp != position) {
|
|
|
|
position = temp;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-04-29 10:35:45 +00:00
|
|
|
/*
|
2014-05-15 17:23:23 +00:00
|
|
|
* See update_scanline_offset() for the details on the
|
|
|
|
* scanline_offset adjustment.
|
2014-04-29 10:35:45 +00:00
|
|
|
*/
|
2014-05-15 17:23:23 +00:00
|
|
|
return (position + crtc->scanline_offset) % vtotal;
|
2014-04-29 10:35:45 +00:00
|
|
|
}
|
|
|
|
|
drm/vblank: drop the mode argument from drm_calc_vbltimestamp_from_scanoutpos
If we restrict this helper to only kms drivers (which is the case) we
can look up the correct mode easily ourselves. But it's a bit tricky:
- All legacy drivers look at crtc->hwmode. But that is updated already
at the beginning of the modeset helper, which means when we disable
a pipe. Hence the final timestamps might be a bit off. But since
this is an existing bug I'm not going to change it, but just try to
be bug-for-bug compatible with the current code. This only applies
to radeon&amdgpu.
- i915 tries to get it perfect by updating crtc->hwmode when the pipe
is off (i.e. vblank->enabled = false).
- All other atomic drivers look at crtc->state->adjusted_mode. Those
that look at state->requested_mode simply don't adjust their mode,
so it's the same. That has two problems: Accessing crtc->state from
interrupt handling code is unsafe, and it's updated before we shut
down the pipe. For nonblocking modesets it's even worse.
For atomic drivers try to implement what i915 does. To do that we add
a new hwmode field to the vblank structure, and update it from
drm_calc_timestamping_constants(). For atomic drivers that's called
from the right spot by the helper library already, so all fine. But
for safety let's enforce that.
For legacy driver this function is only called at the end (oh the
fun), which is broken, so again let's not bother and just stay
bug-for-bug compatible.
The benefit is that we can use drm_calc_vbltimestamp_from_scanoutpos
directly to implement ->get_vblank_timestamp in every driver, deleting
a lot of code.
v2: Completely new approach, trying to mimick the i915 solution.
v3: Fixup kerneldoc.
v4: Drop the WARN_ON to check that the vblank is off, atomic helpers
currently unconditionally call this. Recomputing the same stuff should
be harmless.
v5: Fix typos and move misplaced hunks to the right patches (Neil).
v6: Undo hunk movement (kbuild).
Cc: Mario Kleiner <mario.kleiner@tuebingen.mpg.de>
Cc: Eric Anholt <eric@anholt.net>
Cc: Rob Clark <robdclark@gmail.com>
Cc: linux-arm-msm@vger.kernel.org
Cc: freedreno@lists.freedesktop.org
Cc: Alex Deucher <alexander.deucher@amd.com>
Cc: Christian König <christian.koenig@amd.com>
Cc: Ben Skeggs <bskeggs@redhat.com>
Reviewed-by: Neil Armstrong <narmstrong@baylibre.com>
Acked-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/20170509140329.24114-4-daniel.vetter@ffwll.ch
2017-05-09 14:03:28 +00:00
|
|
|
static bool i915_get_crtc_scanoutpos(struct drm_device *dev, unsigned int pipe,
|
|
|
|
bool in_vblank_irq, int *vpos, int *hpos,
|
|
|
|
ktime_t *stime, ktime_t *etime,
|
|
|
|
const struct drm_display_mode *mode)
|
2010-12-08 03:07:19 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2016-10-31 20:37:10 +00:00
|
|
|
struct intel_crtc *intel_crtc = intel_get_crtc_for_pipe(dev_priv,
|
|
|
|
pipe);
|
2013-10-11 16:10:32 +00:00
|
|
|
int position;
|
2014-04-29 10:35:44 +00:00
|
|
|
int vbl_start, vbl_end, hsync_start, htotal, vtotal;
|
2013-10-30 04:13:08 +00:00
|
|
|
unsigned long irqflags;
|
2010-12-08 03:07:19 +00:00
|
|
|
|
2015-06-01 10:50:07 +00:00
|
|
|
if (WARN_ON(!mode->crtc_clock)) {
|
2010-12-08 03:07:19 +00:00
|
|
|
DRM_DEBUG_DRIVER("trying to get scanoutpos for disabled "
|
2011-02-07 20:26:52 +00:00
|
|
|
"pipe %c\n", pipe_name(pipe));
|
drm/vblank: drop the mode argument from drm_calc_vbltimestamp_from_scanoutpos
If we restrict this helper to only kms drivers (which is the case) we
can look up the correct mode easily ourselves. But it's a bit tricky:
- All legacy drivers look at crtc->hwmode. But that is updated already
at the beginning of the modeset helper, which means when we disable
a pipe. Hence the final timestamps might be a bit off. But since
this is an existing bug I'm not going to change it, but just try to
be bug-for-bug compatible with the current code. This only applies
to radeon&amdgpu.
- i915 tries to get it perfect by updating crtc->hwmode when the pipe
is off (i.e. vblank->enabled = false).
- All other atomic drivers look at crtc->state->adjusted_mode. Those
that look at state->requested_mode simply don't adjust their mode,
so it's the same. That has two problems: Accessing crtc->state from
interrupt handling code is unsafe, and it's updated before we shut
down the pipe. For nonblocking modesets it's even worse.
For atomic drivers try to implement what i915 does. To do that we add
a new hwmode field to the vblank structure, and update it from
drm_calc_timestamping_constants(). For atomic drivers that's called
from the right spot by the helper library already, so all fine. But
for safety let's enforce that.
For legacy driver this function is only called at the end (oh the
fun), which is broken, so again let's not bother and just stay
bug-for-bug compatible.
The benefit is that we can use drm_calc_vbltimestamp_from_scanoutpos
directly to implement ->get_vblank_timestamp in every driver, deleting
a lot of code.
v2: Completely new approach, trying to mimick the i915 solution.
v3: Fixup kerneldoc.
v4: Drop the WARN_ON to check that the vblank is off, atomic helpers
currently unconditionally call this. Recomputing the same stuff should
be harmless.
v5: Fix typos and move misplaced hunks to the right patches (Neil).
v6: Undo hunk movement (kbuild).
Cc: Mario Kleiner <mario.kleiner@tuebingen.mpg.de>
Cc: Eric Anholt <eric@anholt.net>
Cc: Rob Clark <robdclark@gmail.com>
Cc: linux-arm-msm@vger.kernel.org
Cc: freedreno@lists.freedesktop.org
Cc: Alex Deucher <alexander.deucher@amd.com>
Cc: Christian König <christian.koenig@amd.com>
Cc: Ben Skeggs <bskeggs@redhat.com>
Reviewed-by: Neil Armstrong <narmstrong@baylibre.com>
Acked-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/20170509140329.24114-4-daniel.vetter@ffwll.ch
2017-05-09 14:03:28 +00:00
|
|
|
return false;
|
2010-12-08 03:07:19 +00:00
|
|
|
}
|
|
|
|
|
2013-09-23 11:48:50 +00:00
|
|
|
htotal = mode->crtc_htotal;
|
2014-04-29 10:35:44 +00:00
|
|
|
hsync_start = mode->crtc_hsync_start;
|
2013-09-23 11:48:50 +00:00
|
|
|
vtotal = mode->crtc_vtotal;
|
|
|
|
vbl_start = mode->crtc_vblank_start;
|
|
|
|
vbl_end = mode->crtc_vblank_end;
|
2010-12-08 03:07:19 +00:00
|
|
|
|
2013-10-28 14:31:41 +00:00
|
|
|
if (mode->flags & DRM_MODE_FLAG_INTERLACE) {
|
|
|
|
vbl_start = DIV_ROUND_UP(vbl_start, 2);
|
|
|
|
vbl_end /= 2;
|
|
|
|
vtotal /= 2;
|
|
|
|
}
|
|
|
|
|
2013-10-30 04:13:08 +00:00
|
|
|
/*
|
|
|
|
* Lock uncore.lock, as we will do multiple timing critical raw
|
|
|
|
* register reads, potentially with preemption disabled, so the
|
|
|
|
* following code must not block on uncore.lock.
|
|
|
|
*/
|
|
|
|
spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
|
2014-04-29 10:35:44 +00:00
|
|
|
|
2013-10-30 04:13:08 +00:00
|
|
|
/* preempt_disable_rt() should go right here in PREEMPT_RT patchset. */
|
|
|
|
|
|
|
|
/* Get optional system timestamp before query. */
|
|
|
|
if (stime)
|
|
|
|
*stime = ktime_get();
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
if (IS_GEN2(dev_priv) || IS_G4X(dev_priv) || INTEL_GEN(dev_priv) >= 5) {
|
2010-12-08 03:07:19 +00:00
|
|
|
/* No obvious pixelcount register. Only query vertical
|
|
|
|
* scanout position from Display scan line register.
|
|
|
|
*/
|
2014-04-29 10:35:45 +00:00
|
|
|
position = __intel_get_crtc_scanline(intel_crtc);
|
2010-12-08 03:07:19 +00:00
|
|
|
} else {
|
|
|
|
/* Have access to pixelcount since start of frame.
|
|
|
|
* We can split this into vertical and horizontal
|
|
|
|
* scanout position.
|
|
|
|
*/
|
2015-10-22 12:34:56 +00:00
|
|
|
position = (I915_READ_FW(PIPEFRAMEPIXEL(pipe)) & PIPE_PIXEL_MASK) >> PIPE_PIXEL_SHIFT;
|
2010-12-08 03:07:19 +00:00
|
|
|
|
2013-10-11 16:10:32 +00:00
|
|
|
/* convert to pixel counts */
|
|
|
|
vbl_start *= htotal;
|
|
|
|
vbl_end *= htotal;
|
|
|
|
vtotal *= htotal;
|
2014-04-29 10:35:44 +00:00
|
|
|
|
2014-04-29 10:35:49 +00:00
|
|
|
/*
|
|
|
|
* In interlaced modes, the pixel counter counts all pixels,
|
|
|
|
* so one field will have htotal more pixels. In order to avoid
|
|
|
|
* the reported position from jumping backwards when the pixel
|
|
|
|
* counter is beyond the length of the shorter field, just
|
|
|
|
* clamp the position the length of the shorter field. This
|
|
|
|
* matches how the scanline counter based position works since
|
|
|
|
* the scanline counter doesn't count the two half lines.
|
|
|
|
*/
|
|
|
|
if (position >= vtotal)
|
|
|
|
position = vtotal - 1;
|
|
|
|
|
2014-04-29 10:35:44 +00:00
|
|
|
/*
|
|
|
|
* Start of vblank interrupt is triggered at start of hsync,
|
|
|
|
* just prior to the first active line of vblank. However we
|
|
|
|
* consider lines to start at the leading edge of horizontal
|
|
|
|
* active. So, should we get here before we've crossed into
|
|
|
|
* the horizontal active of the first line in vblank, we would
|
|
|
|
* not set the DRM_SCANOUTPOS_INVBL flag. In order to fix that,
|
|
|
|
* always add htotal-hsync_start to the current pixel position.
|
|
|
|
*/
|
|
|
|
position = (position + htotal - hsync_start) % vtotal;
|
2010-12-08 03:07:19 +00:00
|
|
|
}
|
|
|
|
|
2013-10-30 04:13:08 +00:00
|
|
|
/* Get optional system timestamp after query. */
|
|
|
|
if (etime)
|
|
|
|
*etime = ktime_get();
|
|
|
|
|
|
|
|
/* preempt_enable_rt() should go right here in PREEMPT_RT patchset. */
|
|
|
|
|
|
|
|
spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
|
|
|
|
|
2013-10-11 16:10:32 +00:00
|
|
|
/*
|
|
|
|
* While in vblank, position will be negative
|
|
|
|
* counting up towards 0 at vbl_end. And outside
|
|
|
|
* vblank, position will be positive counting
|
|
|
|
* up since vbl_end.
|
|
|
|
*/
|
|
|
|
if (position >= vbl_start)
|
|
|
|
position -= vbl_end;
|
|
|
|
else
|
|
|
|
position += vtotal - vbl_end;
|
2010-12-08 03:07:19 +00:00
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
if (IS_GEN2(dev_priv) || IS_G4X(dev_priv) || INTEL_GEN(dev_priv) >= 5) {
|
2013-10-11 16:10:32 +00:00
|
|
|
*vpos = position;
|
|
|
|
*hpos = 0;
|
|
|
|
} else {
|
|
|
|
*vpos = position / htotal;
|
|
|
|
*hpos = position - (*vpos * htotal);
|
|
|
|
}
|
2010-12-08 03:07:19 +00:00
|
|
|
|
drm/vblank: drop the mode argument from drm_calc_vbltimestamp_from_scanoutpos
If we restrict this helper to only kms drivers (which is the case) we
can look up the correct mode easily ourselves. But it's a bit tricky:
- All legacy drivers look at crtc->hwmode. But that is updated already
at the beginning of the modeset helper, which means when we disable
a pipe. Hence the final timestamps might be a bit off. But since
this is an existing bug I'm not going to change it, but just try to
be bug-for-bug compatible with the current code. This only applies
to radeon&amdgpu.
- i915 tries to get it perfect by updating crtc->hwmode when the pipe
is off (i.e. vblank->enabled = false).
- All other atomic drivers look at crtc->state->adjusted_mode. Those
that look at state->requested_mode simply don't adjust their mode,
so it's the same. That has two problems: Accessing crtc->state from
interrupt handling code is unsafe, and it's updated before we shut
down the pipe. For nonblocking modesets it's even worse.
For atomic drivers try to implement what i915 does. To do that we add
a new hwmode field to the vblank structure, and update it from
drm_calc_timestamping_constants(). For atomic drivers that's called
from the right spot by the helper library already, so all fine. But
for safety let's enforce that.
For legacy driver this function is only called at the end (oh the
fun), which is broken, so again let's not bother and just stay
bug-for-bug compatible.
The benefit is that we can use drm_calc_vbltimestamp_from_scanoutpos
directly to implement ->get_vblank_timestamp in every driver, deleting
a lot of code.
v2: Completely new approach, trying to mimick the i915 solution.
v3: Fixup kerneldoc.
v4: Drop the WARN_ON to check that the vblank is off, atomic helpers
currently unconditionally call this. Recomputing the same stuff should
be harmless.
v5: Fix typos and move misplaced hunks to the right patches (Neil).
v6: Undo hunk movement (kbuild).
Cc: Mario Kleiner <mario.kleiner@tuebingen.mpg.de>
Cc: Eric Anholt <eric@anholt.net>
Cc: Rob Clark <robdclark@gmail.com>
Cc: linux-arm-msm@vger.kernel.org
Cc: freedreno@lists.freedesktop.org
Cc: Alex Deucher <alexander.deucher@amd.com>
Cc: Christian König <christian.koenig@amd.com>
Cc: Ben Skeggs <bskeggs@redhat.com>
Reviewed-by: Neil Armstrong <narmstrong@baylibre.com>
Acked-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/20170509140329.24114-4-daniel.vetter@ffwll.ch
2017-05-09 14:03:28 +00:00
|
|
|
return true;
|
2010-12-08 03:07:19 +00:00
|
|
|
}
|
|
|
|
|
2014-04-29 10:35:45 +00:00
|
|
|
int intel_get_crtc_scanline(struct intel_crtc *crtc)
|
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
|
2014-04-29 10:35:45 +00:00
|
|
|
unsigned long irqflags;
|
|
|
|
int position;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
|
|
|
|
position = __intel_get_crtc_scanline(crtc);
|
|
|
|
spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
|
|
|
|
|
|
|
|
return position;
|
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void ironlake_rps_change_irq_handler(struct drm_i915_private *dev_priv)
|
2010-01-29 19:27:07 +00:00
|
|
|
{
|
2010-02-02 18:30:47 +00:00
|
|
|
u32 busy_up, busy_down, max_avg, min_avg;
|
2012-08-09 14:46:01 +00:00
|
|
|
u8 new_delay;
|
|
|
|
|
2013-07-04 21:35:25 +00:00
|
|
|
spin_lock(&mchdev_lock);
|
2010-01-29 19:27:07 +00:00
|
|
|
|
2012-08-08 21:35:37 +00:00
|
|
|
I915_WRITE16(MEMINTRSTS, I915_READ(MEMINTRSTS));
|
|
|
|
|
2012-08-08 21:35:39 +00:00
|
|
|
new_delay = dev_priv->ips.cur_delay;
|
2012-08-09 14:46:01 +00:00
|
|
|
|
2010-05-20 21:28:11 +00:00
|
|
|
I915_WRITE16(MEMINTRSTS, MEMINT_EVAL_CHG);
|
2010-02-02 18:30:47 +00:00
|
|
|
busy_up = I915_READ(RCPREVBSYTUPAVG);
|
|
|
|
busy_down = I915_READ(RCPREVBSYTDNAVG);
|
2010-01-29 19:27:07 +00:00
|
|
|
max_avg = I915_READ(RCBMAXAVG);
|
|
|
|
min_avg = I915_READ(RCBMINAVG);
|
|
|
|
|
|
|
|
/* Handle RCS change request from hw */
|
2010-02-02 18:30:47 +00:00
|
|
|
if (busy_up > max_avg) {
|
2012-08-08 21:35:39 +00:00
|
|
|
if (dev_priv->ips.cur_delay != dev_priv->ips.max_delay)
|
|
|
|
new_delay = dev_priv->ips.cur_delay - 1;
|
|
|
|
if (new_delay < dev_priv->ips.max_delay)
|
|
|
|
new_delay = dev_priv->ips.max_delay;
|
2010-02-02 18:30:47 +00:00
|
|
|
} else if (busy_down < min_avg) {
|
2012-08-08 21:35:39 +00:00
|
|
|
if (dev_priv->ips.cur_delay != dev_priv->ips.min_delay)
|
|
|
|
new_delay = dev_priv->ips.cur_delay + 1;
|
|
|
|
if (new_delay > dev_priv->ips.min_delay)
|
|
|
|
new_delay = dev_priv->ips.min_delay;
|
2010-01-29 19:27:07 +00:00
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
if (ironlake_set_drps(dev_priv, new_delay))
|
2012-08-08 21:35:39 +00:00
|
|
|
dev_priv->ips.cur_delay = new_delay;
|
2010-01-29 19:27:07 +00:00
|
|
|
|
2013-07-04 21:35:25 +00:00
|
|
|
spin_unlock(&mchdev_lock);
|
2012-08-09 14:46:01 +00:00
|
|
|
|
2010-01-29 19:27:07 +00:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2016-03-16 11:00:37 +00:00
|
|
|
static void notify_ring(struct intel_engine_cs *engine)
|
2010-10-19 10:19:32 +00:00
|
|
|
{
|
2017-02-27 20:58:48 +00:00
|
|
|
struct drm_i915_gem_request *rq = NULL;
|
|
|
|
struct intel_wait *wait;
|
2017-02-21 09:13:48 +00:00
|
|
|
|
2017-02-17 15:13:00 +00:00
|
|
|
atomic_inc(&engine->irq_count);
|
2017-01-24 15:18:05 +00:00
|
|
|
set_bit(ENGINE_IRQ_BREADCRUMB, &engine->irq_posted);
|
2017-02-27 20:58:48 +00:00
|
|
|
|
2017-03-03 19:08:24 +00:00
|
|
|
spin_lock(&engine->breadcrumbs.irq_lock);
|
|
|
|
wait = engine->breadcrumbs.irq_wait;
|
2017-02-27 20:58:48 +00:00
|
|
|
if (wait) {
|
2017-09-18 16:27:33 +00:00
|
|
|
bool wakeup = engine->irq_seqno_barrier;
|
|
|
|
|
2017-02-27 20:58:48 +00:00
|
|
|
/* We use a callback from the dma-fence to submit
|
|
|
|
* requests after waiting on our own requests. To
|
|
|
|
* ensure minimum delay in queuing the next request to
|
|
|
|
* hardware, signal the fence now rather than wait for
|
|
|
|
* the signaler to be woken up. We still wake up the
|
|
|
|
* waiter in order to handle the irq-seqno coherency
|
|
|
|
* issues (we may receive the interrupt before the
|
|
|
|
* seqno is written, see __i915_request_irq_complete())
|
|
|
|
* and to handle coalescing of multiple seqno updates
|
|
|
|
* and many waiters.
|
|
|
|
*/
|
|
|
|
if (i915_seqno_passed(intel_engine_get_seqno(engine),
|
2017-09-18 16:27:33 +00:00
|
|
|
wait->seqno)) {
|
2017-09-18 16:27:34 +00:00
|
|
|
struct drm_i915_gem_request *waiter = wait->request;
|
|
|
|
|
2017-09-18 16:27:33 +00:00
|
|
|
wakeup = true;
|
|
|
|
if (!test_bit(DMA_FENCE_FLAG_SIGNALED_BIT,
|
2017-09-18 16:27:34 +00:00
|
|
|
&waiter->fence.flags) &&
|
|
|
|
intel_wait_check_request(wait, waiter))
|
|
|
|
rq = i915_gem_request_get(waiter);
|
2017-09-18 16:27:33 +00:00
|
|
|
}
|
2017-02-27 20:58:48 +00:00
|
|
|
|
2017-09-18 16:27:33 +00:00
|
|
|
if (wakeup)
|
|
|
|
wake_up_process(wait->tsk);
|
2017-02-27 20:58:50 +00:00
|
|
|
} else {
|
|
|
|
__intel_engine_disarm_breadcrumbs(engine);
|
2017-02-27 20:58:48 +00:00
|
|
|
}
|
2017-03-03 19:08:24 +00:00
|
|
|
spin_unlock(&engine->breadcrumbs.irq_lock);
|
2017-02-27 20:58:48 +00:00
|
|
|
|
2017-03-03 14:45:57 +00:00
|
|
|
if (rq) {
|
2017-02-27 20:58:48 +00:00
|
|
|
dma_fence_signal(&rq->fence);
|
2017-03-03 14:45:57 +00:00
|
|
|
i915_gem_request_put(rq);
|
|
|
|
}
|
2017-02-27 20:58:48 +00:00
|
|
|
|
|
|
|
trace_intel_engine_notify(engine, wait);
|
2010-10-19 10:19:32 +00:00
|
|
|
}
|
|
|
|
|
2015-03-18 09:48:22 +00:00
|
|
|
static void vlv_c0_read(struct drm_i915_private *dev_priv,
|
|
|
|
struct intel_rps_ei *ei)
|
2014-07-03 21:33:01 +00:00
|
|
|
{
|
2017-03-15 15:43:03 +00:00
|
|
|
ei->ktime = ktime_get_raw();
|
2015-03-18 09:48:22 +00:00
|
|
|
ei->render_c0 = I915_READ(VLV_RENDER_C0_COUNT);
|
|
|
|
ei->media_c0 = I915_READ(VLV_MEDIA_C0_COUNT);
|
|
|
|
}
|
2014-07-03 21:33:01 +00:00
|
|
|
|
2015-03-18 09:48:22 +00:00
|
|
|
void gen6_rps_reset_ei(struct drm_i915_private *dev_priv)
|
2014-07-03 21:33:01 +00:00
|
|
|
{
|
2017-03-09 21:12:30 +00:00
|
|
|
memset(&dev_priv->rps.ei, 0, sizeof(dev_priv->rps.ei));
|
2015-03-18 09:48:22 +00:00
|
|
|
}
|
2014-07-03 21:33:01 +00:00
|
|
|
|
2015-03-18 09:48:22 +00:00
|
|
|
static u32 vlv_wa_c0_ei(struct drm_i915_private *dev_priv, u32 pm_iir)
|
|
|
|
{
|
2017-03-09 21:12:30 +00:00
|
|
|
const struct intel_rps_ei *prev = &dev_priv->rps.ei;
|
2015-03-18 09:48:22 +00:00
|
|
|
struct intel_rps_ei now;
|
|
|
|
u32 events = 0;
|
2014-07-03 21:33:01 +00:00
|
|
|
|
2017-03-09 21:12:30 +00:00
|
|
|
if ((pm_iir & GEN6_PM_RP_UP_EI_EXPIRED) == 0)
|
2015-03-18 09:48:22 +00:00
|
|
|
return 0;
|
2014-07-03 21:33:01 +00:00
|
|
|
|
2015-03-18 09:48:22 +00:00
|
|
|
vlv_c0_read(dev_priv, &now);
|
2014-07-03 21:33:01 +00:00
|
|
|
|
2017-03-15 15:43:03 +00:00
|
|
|
if (prev->ktime) {
|
2017-03-09 21:12:30 +00:00
|
|
|
u64 time, c0;
|
2017-03-09 21:12:31 +00:00
|
|
|
u32 render, media;
|
2017-03-09 21:12:30 +00:00
|
|
|
|
2017-03-15 15:43:03 +00:00
|
|
|
time = ktime_us_delta(now.ktime, prev->ktime);
|
2017-03-13 17:06:17 +00:00
|
|
|
|
2017-03-09 21:12:30 +00:00
|
|
|
time *= dev_priv->czclk_freq;
|
|
|
|
|
|
|
|
/* Workload can be split between render + media,
|
|
|
|
* e.g. SwapBuffers being blitted in X after being rendered in
|
|
|
|
* mesa. To account for this we need to combine both engines
|
|
|
|
* into our activity counter.
|
|
|
|
*/
|
2017-03-09 21:12:31 +00:00
|
|
|
render = now.render_c0 - prev->render_c0;
|
|
|
|
media = now.media_c0 - prev->media_c0;
|
|
|
|
c0 = max(render, media);
|
2017-03-15 16:12:59 +00:00
|
|
|
c0 *= 1000 * 100 << 8; /* to usecs and scale to threshold% */
|
2017-03-09 21:12:30 +00:00
|
|
|
|
|
|
|
if (c0 > time * dev_priv->rps.up_threshold)
|
|
|
|
events = GEN6_PM_RP_UP_THRESHOLD;
|
|
|
|
else if (c0 < time * dev_priv->rps.down_threshold)
|
|
|
|
events = GEN6_PM_RP_DOWN_THRESHOLD;
|
2014-07-03 21:33:01 +00:00
|
|
|
}
|
|
|
|
|
2017-03-09 21:12:30 +00:00
|
|
|
dev_priv->rps.ei = now;
|
2015-03-18 09:48:22 +00:00
|
|
|
return events;
|
2014-07-03 21:33:01 +00:00
|
|
|
}
|
|
|
|
|
2011-04-25 18:25:20 +00:00
|
|
|
static void gen6_pm_rps_work(struct work_struct *work)
|
2010-12-17 22:19:02 +00:00
|
|
|
{
|
2014-03-31 11:27:17 +00:00
|
|
|
struct drm_i915_private *dev_priv =
|
|
|
|
container_of(work, struct drm_i915_private, rps.work);
|
2017-03-09 21:12:32 +00:00
|
|
|
bool client_boost = false;
|
2015-05-21 20:01:47 +00:00
|
|
|
int new_delay, adj, min, max;
|
2017-03-09 21:12:32 +00:00
|
|
|
u32 pm_iir = 0;
|
2011-04-25 18:25:20 +00:00
|
|
|
|
2013-07-04 21:35:28 +00:00
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
2017-03-09 21:12:32 +00:00
|
|
|
if (dev_priv->rps.interrupts_enabled) {
|
|
|
|
pm_iir = fetch_and_zero(&dev_priv->rps.pm_iir);
|
2017-06-28 12:35:48 +00:00
|
|
|
client_boost = atomic_read(&dev_priv->rps.num_waiters);
|
drm/i915: sanitize rps irq disabling
When disabling the RPS interrupts there is a tricky dependency between
the thread disabling the interrupts, the RPS interrupt handler and the
corresponding RPS work. The RPS work can reenable the interrupts, so
there is no straightforward order in the disabling thread to (1) make
sure that any RPS work is flushed and to (2) disable all RPS
interrupts. Currently this is solved by masking the interrupts using two
separate mask registers (first level display IMR and PM IMR) and doing
the disabling when all first level interrupts are disabled.
This works, but the requirement to run with all first level interrupts
disabled is unnecessary making the suspend / unload time ordering of RPS
disabling wrt. other unitialization steps difficult and error prone.
Removing this restriction allows us to disable RPS early during suspend
/ unload and forget about it for the rest of the sequence. By adding a
more explicit method for avoiding the above race, it also becomes easier
to prove its correctness. Finally currently we can hit the WARN in
snb_update_pm_irq(), when a final RPS work runs with the first level
interrupts already disabled. This won't lead to any problem (due to the
separate interrupt masks), but with the change in this and the next
patch we can get rid of the WARN, while leaving it in place for other
scenarios.
To address the above points, add a new RPS interrupts_enabled flag and
use this during RPS disabling to avoid requeuing the RPS work and
reenabling of the RPS interrupts. Since the interrupt disabling happens
now in intel_suspend_gt_powersave(), we will disable RPS interrupts
explicitly during suspend (and not just through the first level mask),
but there is no problem doing so, it's also more consistent and allows
us to unify more of the RPS disabling during suspend and unload time in
the next patch.
v2/v3:
- rebase on patch "drm/i915: move rps irq disable one level up" in the
patchset
Signed-off-by: Imre Deak <imre.deak@intel.com>
Reviewed-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2014-11-19 13:30:04 +00:00
|
|
|
}
|
2013-07-04 21:35:28 +00:00
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
2010-12-17 22:19:02 +00:00
|
|
|
|
2013-08-15 14:50:01 +00:00
|
|
|
/* Make sure we didn't queue anything we're not going to process. */
|
2014-03-15 14:53:22 +00:00
|
|
|
WARN_ON(pm_iir & ~dev_priv->pm_rps_events);
|
2015-05-21 20:01:47 +00:00
|
|
|
if ((pm_iir & dev_priv->pm_rps_events) == 0 && !client_boost)
|
2017-03-09 21:12:32 +00:00
|
|
|
goto out;
|
2010-12-17 22:19:02 +00:00
|
|
|
|
2012-11-02 18:14:01 +00:00
|
|
|
mutex_lock(&dev_priv->rps.hw_lock);
|
2012-04-28 07:56:39 +00:00
|
|
|
|
2015-03-18 09:48:22 +00:00
|
|
|
pm_iir |= vlv_wa_c0_ei(dev_priv, pm_iir);
|
|
|
|
|
drm/i915: Tweak RPS thresholds to more aggressively downclock
After applying wait-boost we often find ourselves stuck at higher clocks
than required. The current threshold value requires the GPU to be
continuously and completely idle for 313ms before it is dropped by one
bin. Conversely, we require the GPU to be busy for an average of 90% over
a 84ms period before we upclock. So the current thresholds almost never
downclock the GPU, and respond very slowly to sudden demands for more
power. It is easy to observe that we currently lock into the wrong bin
and both underperform in benchmarks and consume more power than optimal
(just by repeating the task and measuring the different results).
An alternative approach, as discussed in the bspec, is to use a
continuous threshold for upclocking, and an average value for downclocking.
This is good for quickly detecting and reacting to state changes within a
frame, however it fails with the common throttling method of waiting
upon the outstanding frame - at least it is difficult to choose a
threshold that works well at 15,000fps and at 60fps. So continue to use
average busy/idle loads to determine frequency change.
v2: Use 3 power zones to keep frequencies low in steady-state mostly
idle (e.g. scrolling, interactive 2D drawing), and frequencies high
for demanding games. In between those end-states, we use a
fast-reclocking algorithm to converge more quickly on the desired bin.
v3: Bug fixes - make sure we reset adj after switching power zones.
v4: Tune - drop the continuous busy thresholds as it prevents us from
choosing the right frequency for glxgears style swap benchmarks. Instead
the goal is to be able to find the right clocks irrespective of the
wait-boost.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Kenneth Graunke <kenneth@whitecape.org>
Cc: Stéphane Marchesin <stephane.marchesin@gmail.com>
Cc: Owen Taylor <otaylor@redhat.com>
Cc: "Meng, Mengmeng" <mengmeng.meng@intel.com>
Cc: "Zhuang, Lena" <lena.zhuang@intel.com>
Reviewed-by: Jesse Barnes <jbarnes@virtuousgeek.org>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-09-25 16:34:57 +00:00
|
|
|
adj = dev_priv->rps.last_adj;
|
2015-04-07 15:20:29 +00:00
|
|
|
new_delay = dev_priv->rps.cur_freq;
|
2015-05-21 20:01:47 +00:00
|
|
|
min = dev_priv->rps.min_freq_softlimit;
|
|
|
|
max = dev_priv->rps.max_freq_softlimit;
|
2017-06-28 12:35:48 +00:00
|
|
|
if (client_boost)
|
2016-07-13 08:10:35 +00:00
|
|
|
max = dev_priv->rps.max_freq;
|
|
|
|
if (client_boost && new_delay < dev_priv->rps.boost_freq) {
|
|
|
|
new_delay = dev_priv->rps.boost_freq;
|
2015-05-21 20:01:47 +00:00
|
|
|
adj = 0;
|
|
|
|
} else if (pm_iir & GEN6_PM_RP_UP_THRESHOLD) {
|
drm/i915: Tweak RPS thresholds to more aggressively downclock
After applying wait-boost we often find ourselves stuck at higher clocks
than required. The current threshold value requires the GPU to be
continuously and completely idle for 313ms before it is dropped by one
bin. Conversely, we require the GPU to be busy for an average of 90% over
a 84ms period before we upclock. So the current thresholds almost never
downclock the GPU, and respond very slowly to sudden demands for more
power. It is easy to observe that we currently lock into the wrong bin
and both underperform in benchmarks and consume more power than optimal
(just by repeating the task and measuring the different results).
An alternative approach, as discussed in the bspec, is to use a
continuous threshold for upclocking, and an average value for downclocking.
This is good for quickly detecting and reacting to state changes within a
frame, however it fails with the common throttling method of waiting
upon the outstanding frame - at least it is difficult to choose a
threshold that works well at 15,000fps and at 60fps. So continue to use
average busy/idle loads to determine frequency change.
v2: Use 3 power zones to keep frequencies low in steady-state mostly
idle (e.g. scrolling, interactive 2D drawing), and frequencies high
for demanding games. In between those end-states, we use a
fast-reclocking algorithm to converge more quickly on the desired bin.
v3: Bug fixes - make sure we reset adj after switching power zones.
v4: Tune - drop the continuous busy thresholds as it prevents us from
choosing the right frequency for glxgears style swap benchmarks. Instead
the goal is to be able to find the right clocks irrespective of the
wait-boost.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Kenneth Graunke <kenneth@whitecape.org>
Cc: Stéphane Marchesin <stephane.marchesin@gmail.com>
Cc: Owen Taylor <otaylor@redhat.com>
Cc: "Meng, Mengmeng" <mengmeng.meng@intel.com>
Cc: "Zhuang, Lena" <lena.zhuang@intel.com>
Reviewed-by: Jesse Barnes <jbarnes@virtuousgeek.org>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-09-25 16:34:57 +00:00
|
|
|
if (adj > 0)
|
|
|
|
adj *= 2;
|
2015-04-07 15:20:29 +00:00
|
|
|
else /* CHV needs even encode values */
|
|
|
|
adj = IS_CHERRYVIEW(dev_priv) ? 2 : 1;
|
2017-01-20 03:48:24 +00:00
|
|
|
|
|
|
|
if (new_delay >= dev_priv->rps.max_freq_softlimit)
|
|
|
|
adj = 0;
|
2017-06-28 12:35:48 +00:00
|
|
|
} else if (client_boost) {
|
2015-04-27 12:41:23 +00:00
|
|
|
adj = 0;
|
drm/i915: Tweak RPS thresholds to more aggressively downclock
After applying wait-boost we often find ourselves stuck at higher clocks
than required. The current threshold value requires the GPU to be
continuously and completely idle for 313ms before it is dropped by one
bin. Conversely, we require the GPU to be busy for an average of 90% over
a 84ms period before we upclock. So the current thresholds almost never
downclock the GPU, and respond very slowly to sudden demands for more
power. It is easy to observe that we currently lock into the wrong bin
and both underperform in benchmarks and consume more power than optimal
(just by repeating the task and measuring the different results).
An alternative approach, as discussed in the bspec, is to use a
continuous threshold for upclocking, and an average value for downclocking.
This is good for quickly detecting and reacting to state changes within a
frame, however it fails with the common throttling method of waiting
upon the outstanding frame - at least it is difficult to choose a
threshold that works well at 15,000fps and at 60fps. So continue to use
average busy/idle loads to determine frequency change.
v2: Use 3 power zones to keep frequencies low in steady-state mostly
idle (e.g. scrolling, interactive 2D drawing), and frequencies high
for demanding games. In between those end-states, we use a
fast-reclocking algorithm to converge more quickly on the desired bin.
v3: Bug fixes - make sure we reset adj after switching power zones.
v4: Tune - drop the continuous busy thresholds as it prevents us from
choosing the right frequency for glxgears style swap benchmarks. Instead
the goal is to be able to find the right clocks irrespective of the
wait-boost.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Kenneth Graunke <kenneth@whitecape.org>
Cc: Stéphane Marchesin <stephane.marchesin@gmail.com>
Cc: Owen Taylor <otaylor@redhat.com>
Cc: "Meng, Mengmeng" <mengmeng.meng@intel.com>
Cc: "Zhuang, Lena" <lena.zhuang@intel.com>
Reviewed-by: Jesse Barnes <jbarnes@virtuousgeek.org>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-09-25 16:34:57 +00:00
|
|
|
} else if (pm_iir & GEN6_PM_RP_DOWN_TIMEOUT) {
|
2014-03-20 01:31:11 +00:00
|
|
|
if (dev_priv->rps.cur_freq > dev_priv->rps.efficient_freq)
|
|
|
|
new_delay = dev_priv->rps.efficient_freq;
|
2017-02-10 15:03:47 +00:00
|
|
|
else if (dev_priv->rps.cur_freq > dev_priv->rps.min_freq_softlimit)
|
2014-03-20 01:31:11 +00:00
|
|
|
new_delay = dev_priv->rps.min_freq_softlimit;
|
drm/i915: Tweak RPS thresholds to more aggressively downclock
After applying wait-boost we often find ourselves stuck at higher clocks
than required. The current threshold value requires the GPU to be
continuously and completely idle for 313ms before it is dropped by one
bin. Conversely, we require the GPU to be busy for an average of 90% over
a 84ms period before we upclock. So the current thresholds almost never
downclock the GPU, and respond very slowly to sudden demands for more
power. It is easy to observe that we currently lock into the wrong bin
and both underperform in benchmarks and consume more power than optimal
(just by repeating the task and measuring the different results).
An alternative approach, as discussed in the bspec, is to use a
continuous threshold for upclocking, and an average value for downclocking.
This is good for quickly detecting and reacting to state changes within a
frame, however it fails with the common throttling method of waiting
upon the outstanding frame - at least it is difficult to choose a
threshold that works well at 15,000fps and at 60fps. So continue to use
average busy/idle loads to determine frequency change.
v2: Use 3 power zones to keep frequencies low in steady-state mostly
idle (e.g. scrolling, interactive 2D drawing), and frequencies high
for demanding games. In between those end-states, we use a
fast-reclocking algorithm to converge more quickly on the desired bin.
v3: Bug fixes - make sure we reset adj after switching power zones.
v4: Tune - drop the continuous busy thresholds as it prevents us from
choosing the right frequency for glxgears style swap benchmarks. Instead
the goal is to be able to find the right clocks irrespective of the
wait-boost.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Kenneth Graunke <kenneth@whitecape.org>
Cc: Stéphane Marchesin <stephane.marchesin@gmail.com>
Cc: Owen Taylor <otaylor@redhat.com>
Cc: "Meng, Mengmeng" <mengmeng.meng@intel.com>
Cc: "Zhuang, Lena" <lena.zhuang@intel.com>
Reviewed-by: Jesse Barnes <jbarnes@virtuousgeek.org>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-09-25 16:34:57 +00:00
|
|
|
adj = 0;
|
|
|
|
} else if (pm_iir & GEN6_PM_RP_DOWN_THRESHOLD) {
|
|
|
|
if (adj < 0)
|
|
|
|
adj *= 2;
|
2015-04-07 15:20:29 +00:00
|
|
|
else /* CHV needs even encode values */
|
|
|
|
adj = IS_CHERRYVIEW(dev_priv) ? -2 : -1;
|
2017-01-20 03:48:24 +00:00
|
|
|
|
|
|
|
if (new_delay <= dev_priv->rps.min_freq_softlimit)
|
|
|
|
adj = 0;
|
drm/i915: Tweak RPS thresholds to more aggressively downclock
After applying wait-boost we often find ourselves stuck at higher clocks
than required. The current threshold value requires the GPU to be
continuously and completely idle for 313ms before it is dropped by one
bin. Conversely, we require the GPU to be busy for an average of 90% over
a 84ms period before we upclock. So the current thresholds almost never
downclock the GPU, and respond very slowly to sudden demands for more
power. It is easy to observe that we currently lock into the wrong bin
and both underperform in benchmarks and consume more power than optimal
(just by repeating the task and measuring the different results).
An alternative approach, as discussed in the bspec, is to use a
continuous threshold for upclocking, and an average value for downclocking.
This is good for quickly detecting and reacting to state changes within a
frame, however it fails with the common throttling method of waiting
upon the outstanding frame - at least it is difficult to choose a
threshold that works well at 15,000fps and at 60fps. So continue to use
average busy/idle loads to determine frequency change.
v2: Use 3 power zones to keep frequencies low in steady-state mostly
idle (e.g. scrolling, interactive 2D drawing), and frequencies high
for demanding games. In between those end-states, we use a
fast-reclocking algorithm to converge more quickly on the desired bin.
v3: Bug fixes - make sure we reset adj after switching power zones.
v4: Tune - drop the continuous busy thresholds as it prevents us from
choosing the right frequency for glxgears style swap benchmarks. Instead
the goal is to be able to find the right clocks irrespective of the
wait-boost.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Kenneth Graunke <kenneth@whitecape.org>
Cc: Stéphane Marchesin <stephane.marchesin@gmail.com>
Cc: Owen Taylor <otaylor@redhat.com>
Cc: "Meng, Mengmeng" <mengmeng.meng@intel.com>
Cc: "Zhuang, Lena" <lena.zhuang@intel.com>
Reviewed-by: Jesse Barnes <jbarnes@virtuousgeek.org>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-09-25 16:34:57 +00:00
|
|
|
} else { /* unknown event */
|
2015-04-07 15:20:29 +00:00
|
|
|
adj = 0;
|
drm/i915: Tweak RPS thresholds to more aggressively downclock
After applying wait-boost we often find ourselves stuck at higher clocks
than required. The current threshold value requires the GPU to be
continuously and completely idle for 313ms before it is dropped by one
bin. Conversely, we require the GPU to be busy for an average of 90% over
a 84ms period before we upclock. So the current thresholds almost never
downclock the GPU, and respond very slowly to sudden demands for more
power. It is easy to observe that we currently lock into the wrong bin
and both underperform in benchmarks and consume more power than optimal
(just by repeating the task and measuring the different results).
An alternative approach, as discussed in the bspec, is to use a
continuous threshold for upclocking, and an average value for downclocking.
This is good for quickly detecting and reacting to state changes within a
frame, however it fails with the common throttling method of waiting
upon the outstanding frame - at least it is difficult to choose a
threshold that works well at 15,000fps and at 60fps. So continue to use
average busy/idle loads to determine frequency change.
v2: Use 3 power zones to keep frequencies low in steady-state mostly
idle (e.g. scrolling, interactive 2D drawing), and frequencies high
for demanding games. In between those end-states, we use a
fast-reclocking algorithm to converge more quickly on the desired bin.
v3: Bug fixes - make sure we reset adj after switching power zones.
v4: Tune - drop the continuous busy thresholds as it prevents us from
choosing the right frequency for glxgears style swap benchmarks. Instead
the goal is to be able to find the right clocks irrespective of the
wait-boost.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Kenneth Graunke <kenneth@whitecape.org>
Cc: Stéphane Marchesin <stephane.marchesin@gmail.com>
Cc: Owen Taylor <otaylor@redhat.com>
Cc: "Meng, Mengmeng" <mengmeng.meng@intel.com>
Cc: "Zhuang, Lena" <lena.zhuang@intel.com>
Reviewed-by: Jesse Barnes <jbarnes@virtuousgeek.org>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-09-25 16:34:57 +00:00
|
|
|
}
|
2010-12-17 22:19:02 +00:00
|
|
|
|
2015-04-07 15:20:29 +00:00
|
|
|
dev_priv->rps.last_adj = adj;
|
|
|
|
|
2012-09-08 02:43:42 +00:00
|
|
|
/* sysfs frequency interfaces may have snuck in while servicing the
|
|
|
|
* interrupt
|
|
|
|
*/
|
2015-04-07 15:20:29 +00:00
|
|
|
new_delay += adj;
|
2015-05-21 20:01:47 +00:00
|
|
|
new_delay = clamp_t(int, new_delay, min, max);
|
2014-01-27 16:05:05 +00:00
|
|
|
|
2017-01-26 10:19:19 +00:00
|
|
|
if (intel_set_rps(dev_priv, new_delay)) {
|
|
|
|
DRM_DEBUG_DRIVER("Failed to set new GPU frequency\n");
|
|
|
|
dev_priv->rps.last_adj = 0;
|
|
|
|
}
|
2010-12-17 22:19:02 +00:00
|
|
|
|
2012-11-02 18:14:01 +00:00
|
|
|
mutex_unlock(&dev_priv->rps.hw_lock);
|
2017-03-09 21:12:32 +00:00
|
|
|
|
|
|
|
out:
|
|
|
|
/* Make sure not to corrupt PMIMR state used by ringbuffer on GEN6 */
|
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
|
|
|
if (dev_priv->rps.interrupts_enabled)
|
|
|
|
gen6_unmask_pm_irq(dev_priv, dev_priv->pm_rps_events);
|
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
2010-12-17 22:19:02 +00:00
|
|
|
}
|
|
|
|
|
2012-05-25 23:56:22 +00:00
|
|
|
|
|
|
|
/**
|
|
|
|
* ivybridge_parity_work - Workqueue called when a parity error interrupt
|
|
|
|
* occurred.
|
|
|
|
* @work: workqueue struct
|
|
|
|
*
|
|
|
|
* Doesn't actually do anything except notify userspace. As a consequence of
|
|
|
|
* this event, userspace should try to remap the bad rows since statistically
|
|
|
|
* it is likely the same row is more likely to go bad again.
|
|
|
|
*/
|
|
|
|
static void ivybridge_parity_work(struct work_struct *work)
|
|
|
|
{
|
2014-03-31 11:27:17 +00:00
|
|
|
struct drm_i915_private *dev_priv =
|
2017-04-28 07:58:39 +00:00
|
|
|
container_of(work, typeof(*dev_priv), l3_parity.error_work);
|
2012-05-25 23:56:22 +00:00
|
|
|
u32 error_status, row, bank, subbank;
|
2013-09-19 18:13:41 +00:00
|
|
|
char *parity_event[6];
|
2012-05-25 23:56:22 +00:00
|
|
|
uint32_t misccpctl;
|
2013-09-19 18:13:41 +00:00
|
|
|
uint8_t slice = 0;
|
2012-05-25 23:56:22 +00:00
|
|
|
|
|
|
|
/* We must turn off DOP level clock gating to access the L3 registers.
|
|
|
|
* In order to prevent a get/put style interface, acquire struct mutex
|
|
|
|
* any time we access those registers.
|
|
|
|
*/
|
2016-07-05 09:40:23 +00:00
|
|
|
mutex_lock(&dev_priv->drm.struct_mutex);
|
2012-05-25 23:56:22 +00:00
|
|
|
|
2013-09-19 18:13:41 +00:00
|
|
|
/* If we've screwed up tracking, just let the interrupt fire again */
|
|
|
|
if (WARN_ON(!dev_priv->l3_parity.which_slice))
|
|
|
|
goto out;
|
|
|
|
|
2012-05-25 23:56:22 +00:00
|
|
|
misccpctl = I915_READ(GEN7_MISCCPCTL);
|
|
|
|
I915_WRITE(GEN7_MISCCPCTL, misccpctl & ~GEN7_DOP_CLOCK_GATE_ENABLE);
|
|
|
|
POSTING_READ(GEN7_MISCCPCTL);
|
|
|
|
|
2013-09-19 18:13:41 +00:00
|
|
|
while ((slice = ffs(dev_priv->l3_parity.which_slice)) != 0) {
|
drm/i915: Type safe register read/write
Make I915_READ and I915_WRITE more type safe by wrapping the register
offset in a struct. This should eliminate most of the fumbles we've had
with misplaced parens.
This only takes care of normal mmio registers. We could extend the idea
to other register types and define each with its own struct. That way
you wouldn't be able to accidentally pass the wrong thing to a specific
register access function.
The gpio_reg setup is probably the ugliest thing left. But I figure I'd
just leave it for now, and wait for some divine inspiration to strike
before making it nice.
As for the generated code, it's actually a bit better sometimes. Eg.
looking at i915_irq_handler(), we can see the following change:
lea 0x70024(%rdx,%rax,1),%r9d
mov $0x1,%edx
- movslq %r9d,%r9
- mov %r9,%rsi
- mov %r9,-0x58(%rbp)
- callq *0xd8(%rbx)
+ mov %r9d,%esi
+ mov %r9d,-0x48(%rbp)
callq *0xd8(%rbx)
So previously gcc thought the register offset might be signed and
decided to sign extend it, just in case. The rest appears to be
mostly just minor shuffling of instructions.
v2: i915_mmio_reg_{offset,equal,valid}() helpers added
s/_REG/_MMIO/ in the register defines
mo more switch statements left to worry about
ring_emit stuff got sorted in a prep patch
cmd parser, lrc context and w/a batch buildup also in prep patch
vgpu stuff cleaned up and moved to a prep patch
all other unrelated changes split out
v3: Rebased due to BXT DSI/BLC, MOCS, etc.
v4: Rebased due to churn, s/i915_mmio_reg_t/i915_reg_t/
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/1447853606-2751-1-git-send-email-ville.syrjala@linux.intel.com
2015-11-18 13:33:26 +00:00
|
|
|
i915_reg_t reg;
|
2012-05-25 23:56:22 +00:00
|
|
|
|
2013-09-19 18:13:41 +00:00
|
|
|
slice--;
|
2016-04-07 08:08:05 +00:00
|
|
|
if (WARN_ON_ONCE(slice >= NUM_L3_SLICES(dev_priv)))
|
2013-09-19 18:13:41 +00:00
|
|
|
break;
|
2012-05-25 23:56:22 +00:00
|
|
|
|
2013-09-19 18:13:41 +00:00
|
|
|
dev_priv->l3_parity.which_slice &= ~(1<<slice);
|
2012-05-25 23:56:22 +00:00
|
|
|
|
2015-11-04 21:20:02 +00:00
|
|
|
reg = GEN7_L3CDERRST1(slice);
|
2012-05-25 23:56:22 +00:00
|
|
|
|
2013-09-19 18:13:41 +00:00
|
|
|
error_status = I915_READ(reg);
|
|
|
|
row = GEN7_PARITY_ERROR_ROW(error_status);
|
|
|
|
bank = GEN7_PARITY_ERROR_BANK(error_status);
|
|
|
|
subbank = GEN7_PARITY_ERROR_SUBBANK(error_status);
|
|
|
|
|
|
|
|
I915_WRITE(reg, GEN7_PARITY_ERROR_VALID | GEN7_L3CDERRST1_ENABLE);
|
|
|
|
POSTING_READ(reg);
|
|
|
|
|
|
|
|
parity_event[0] = I915_L3_PARITY_UEVENT "=1";
|
|
|
|
parity_event[1] = kasprintf(GFP_KERNEL, "ROW=%d", row);
|
|
|
|
parity_event[2] = kasprintf(GFP_KERNEL, "BANK=%d", bank);
|
|
|
|
parity_event[3] = kasprintf(GFP_KERNEL, "SUBBANK=%d", subbank);
|
|
|
|
parity_event[4] = kasprintf(GFP_KERNEL, "SLICE=%d", slice);
|
|
|
|
parity_event[5] = NULL;
|
|
|
|
|
2016-07-05 09:40:23 +00:00
|
|
|
kobject_uevent_env(&dev_priv->drm.primary->kdev->kobj,
|
2013-09-19 18:13:41 +00:00
|
|
|
KOBJ_CHANGE, parity_event);
|
2012-05-25 23:56:22 +00:00
|
|
|
|
2013-09-19 18:13:41 +00:00
|
|
|
DRM_DEBUG("Parity error: Slice = %d, Row = %d, Bank = %d, Sub bank = %d.\n",
|
|
|
|
slice, row, bank, subbank);
|
2012-05-25 23:56:22 +00:00
|
|
|
|
2013-09-19 18:13:41 +00:00
|
|
|
kfree(parity_event[4]);
|
|
|
|
kfree(parity_event[3]);
|
|
|
|
kfree(parity_event[2]);
|
|
|
|
kfree(parity_event[1]);
|
|
|
|
}
|
2012-05-25 23:56:22 +00:00
|
|
|
|
2013-09-19 18:13:41 +00:00
|
|
|
I915_WRITE(GEN7_MISCCPCTL, misccpctl);
|
2012-05-25 23:56:22 +00:00
|
|
|
|
2013-09-19 18:13:41 +00:00
|
|
|
out:
|
|
|
|
WARN_ON(dev_priv->l3_parity.which_slice);
|
2014-09-15 12:55:26 +00:00
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
2016-04-07 08:08:05 +00:00
|
|
|
gen5_enable_gt_irq(dev_priv, GT_PARITY_ERROR(dev_priv));
|
2014-09-15 12:55:26 +00:00
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
2013-09-19 18:13:41 +00:00
|
|
|
|
2016-07-05 09:40:23 +00:00
|
|
|
mutex_unlock(&dev_priv->drm.struct_mutex);
|
2012-05-25 23:56:22 +00:00
|
|
|
}
|
|
|
|
|
2016-04-13 18:19:57 +00:00
|
|
|
static void ivybridge_parity_error_irq_handler(struct drm_i915_private *dev_priv,
|
|
|
|
u32 iir)
|
2012-05-25 23:56:22 +00:00
|
|
|
{
|
2016-04-13 18:19:57 +00:00
|
|
|
if (!HAS_L3_DPF(dev_priv))
|
2012-05-25 23:56:22 +00:00
|
|
|
return;
|
|
|
|
|
2013-07-04 21:35:25 +00:00
|
|
|
spin_lock(&dev_priv->irq_lock);
|
2016-04-13 18:19:57 +00:00
|
|
|
gen5_disable_gt_irq(dev_priv, GT_PARITY_ERROR(dev_priv));
|
2013-07-04 21:35:25 +00:00
|
|
|
spin_unlock(&dev_priv->irq_lock);
|
2012-05-25 23:56:22 +00:00
|
|
|
|
2016-04-13 18:19:57 +00:00
|
|
|
iir &= GT_PARITY_ERROR(dev_priv);
|
2013-09-19 18:13:41 +00:00
|
|
|
if (iir & GT_RENDER_L3_PARITY_ERROR_INTERRUPT_S1)
|
|
|
|
dev_priv->l3_parity.which_slice |= 1 << 1;
|
|
|
|
|
|
|
|
if (iir & GT_RENDER_L3_PARITY_ERROR_INTERRUPT)
|
|
|
|
dev_priv->l3_parity.which_slice |= 1 << 0;
|
|
|
|
|
2012-11-02 18:55:07 +00:00
|
|
|
queue_work(dev_priv->wq, &dev_priv->l3_parity.error_work);
|
2012-05-25 23:56:22 +00:00
|
|
|
}
|
|
|
|
|
2016-04-13 18:19:57 +00:00
|
|
|
static void ilk_gt_irq_handler(struct drm_i915_private *dev_priv,
|
2013-07-12 22:56:30 +00:00
|
|
|
u32 gt_iir)
|
|
|
|
{
|
2016-07-01 16:23:21 +00:00
|
|
|
if (gt_iir & GT_RENDER_USER_INTERRUPT)
|
drm/i915: Allocate intel_engine_cs structure only for the enabled engines
With the possibility of addition of many more number of rings in future,
the drm_i915_private structure could bloat as an array, of type
intel_engine_cs, is embedded inside it.
struct intel_engine_cs engine[I915_NUM_ENGINES];
Though this is still fine as generally there is only a single instance of
drm_i915_private structure used, but not all of the possible rings would be
enabled or active on most of the platforms. Some memory can be saved by
allocating intel_engine_cs structure only for the enabled/active engines.
Currently the engine/ring ID is kept static and dev_priv->engine[] is simply
indexed using the enums defined in intel_engine_id.
To save memory and continue using the static engine/ring IDs, 'engine' is
defined as an array of pointers.
struct intel_engine_cs *engine[I915_NUM_ENGINES];
dev_priv->engine[engine_ID] will be NULL for disabled engine instances.
There is a text size reduction of 928 bytes, from 1028200 to 1027272, for
i915.o file (but for i915.ko file text size remain same as 1193131 bytes).
v2:
- Remove the engine iterator field added in drm_i915_private structure,
instead pass a local iterator variable to the for_each_engine**
macros. (Chris)
- Do away with intel_engine_initialized() and instead directly use the
NULL pointer check on engine pointer. (Chris)
v3:
- Remove for_each_engine_id() macro, as the updated macro for_each_engine()
can be used in place of it. (Chris)
- Protect the access to Render engine Fault register with a NULL check, as
engine specific init is done later in Driver load sequence.
v4:
- Use !!dev_priv->engine[VCS] style for the engine check in getparam. (Chris)
- Kill the superfluous init_engine_lists().
v5:
- Cleanup the intel_engines_init() & intel_engines_setup(), with respect to
allocation of intel_engine_cs structure. (Chris)
v6:
- Rebase.
v7:
- Optimize the for_each_engine_masked() macro. (Chris)
- Change the type of 'iter' local variable to enum intel_engine_id. (Chris)
- Rebase.
v8: Rebase.
v9: Rebase.
v10:
- For index calculation use engine ID instead of pointer based arithmetic in
intel_engine_sync_index() as engine pointers are not contiguous now (Chris)
- For appropriateness, rename local enum variable 'iter' to 'id'. (Joonas)
- Use for_each_engine macro for cleanup in intel_engines_init() and remove
check for NULL engine pointer in cleanup() routines. (Joonas)
v11: Rebase.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/1476378888-7372-1-git-send-email-akash.goel@intel.com
2016-10-13 17:14:48 +00:00
|
|
|
notify_ring(dev_priv->engine[RCS]);
|
2013-07-12 22:56:30 +00:00
|
|
|
if (gt_iir & ILK_BSD_USER_INTERRUPT)
|
drm/i915: Allocate intel_engine_cs structure only for the enabled engines
With the possibility of addition of many more number of rings in future,
the drm_i915_private structure could bloat as an array, of type
intel_engine_cs, is embedded inside it.
struct intel_engine_cs engine[I915_NUM_ENGINES];
Though this is still fine as generally there is only a single instance of
drm_i915_private structure used, but not all of the possible rings would be
enabled or active on most of the platforms. Some memory can be saved by
allocating intel_engine_cs structure only for the enabled/active engines.
Currently the engine/ring ID is kept static and dev_priv->engine[] is simply
indexed using the enums defined in intel_engine_id.
To save memory and continue using the static engine/ring IDs, 'engine' is
defined as an array of pointers.
struct intel_engine_cs *engine[I915_NUM_ENGINES];
dev_priv->engine[engine_ID] will be NULL for disabled engine instances.
There is a text size reduction of 928 bytes, from 1028200 to 1027272, for
i915.o file (but for i915.ko file text size remain same as 1193131 bytes).
v2:
- Remove the engine iterator field added in drm_i915_private structure,
instead pass a local iterator variable to the for_each_engine**
macros. (Chris)
- Do away with intel_engine_initialized() and instead directly use the
NULL pointer check on engine pointer. (Chris)
v3:
- Remove for_each_engine_id() macro, as the updated macro for_each_engine()
can be used in place of it. (Chris)
- Protect the access to Render engine Fault register with a NULL check, as
engine specific init is done later in Driver load sequence.
v4:
- Use !!dev_priv->engine[VCS] style for the engine check in getparam. (Chris)
- Kill the superfluous init_engine_lists().
v5:
- Cleanup the intel_engines_init() & intel_engines_setup(), with respect to
allocation of intel_engine_cs structure. (Chris)
v6:
- Rebase.
v7:
- Optimize the for_each_engine_masked() macro. (Chris)
- Change the type of 'iter' local variable to enum intel_engine_id. (Chris)
- Rebase.
v8: Rebase.
v9: Rebase.
v10:
- For index calculation use engine ID instead of pointer based arithmetic in
intel_engine_sync_index() as engine pointers are not contiguous now (Chris)
- For appropriateness, rename local enum variable 'iter' to 'id'. (Joonas)
- Use for_each_engine macro for cleanup in intel_engines_init() and remove
check for NULL engine pointer in cleanup() routines. (Joonas)
v11: Rebase.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/1476378888-7372-1-git-send-email-akash.goel@intel.com
2016-10-13 17:14:48 +00:00
|
|
|
notify_ring(dev_priv->engine[VCS]);
|
2013-07-12 22:56:30 +00:00
|
|
|
}
|
|
|
|
|
2016-04-13 18:19:57 +00:00
|
|
|
static void snb_gt_irq_handler(struct drm_i915_private *dev_priv,
|
2012-03-30 18:24:35 +00:00
|
|
|
u32 gt_iir)
|
|
|
|
{
|
2016-07-01 16:23:21 +00:00
|
|
|
if (gt_iir & GT_RENDER_USER_INTERRUPT)
|
drm/i915: Allocate intel_engine_cs structure only for the enabled engines
With the possibility of addition of many more number of rings in future,
the drm_i915_private structure could bloat as an array, of type
intel_engine_cs, is embedded inside it.
struct intel_engine_cs engine[I915_NUM_ENGINES];
Though this is still fine as generally there is only a single instance of
drm_i915_private structure used, but not all of the possible rings would be
enabled or active on most of the platforms. Some memory can be saved by
allocating intel_engine_cs structure only for the enabled/active engines.
Currently the engine/ring ID is kept static and dev_priv->engine[] is simply
indexed using the enums defined in intel_engine_id.
To save memory and continue using the static engine/ring IDs, 'engine' is
defined as an array of pointers.
struct intel_engine_cs *engine[I915_NUM_ENGINES];
dev_priv->engine[engine_ID] will be NULL for disabled engine instances.
There is a text size reduction of 928 bytes, from 1028200 to 1027272, for
i915.o file (but for i915.ko file text size remain same as 1193131 bytes).
v2:
- Remove the engine iterator field added in drm_i915_private structure,
instead pass a local iterator variable to the for_each_engine**
macros. (Chris)
- Do away with intel_engine_initialized() and instead directly use the
NULL pointer check on engine pointer. (Chris)
v3:
- Remove for_each_engine_id() macro, as the updated macro for_each_engine()
can be used in place of it. (Chris)
- Protect the access to Render engine Fault register with a NULL check, as
engine specific init is done later in Driver load sequence.
v4:
- Use !!dev_priv->engine[VCS] style for the engine check in getparam. (Chris)
- Kill the superfluous init_engine_lists().
v5:
- Cleanup the intel_engines_init() & intel_engines_setup(), with respect to
allocation of intel_engine_cs structure. (Chris)
v6:
- Rebase.
v7:
- Optimize the for_each_engine_masked() macro. (Chris)
- Change the type of 'iter' local variable to enum intel_engine_id. (Chris)
- Rebase.
v8: Rebase.
v9: Rebase.
v10:
- For index calculation use engine ID instead of pointer based arithmetic in
intel_engine_sync_index() as engine pointers are not contiguous now (Chris)
- For appropriateness, rename local enum variable 'iter' to 'id'. (Joonas)
- Use for_each_engine macro for cleanup in intel_engines_init() and remove
check for NULL engine pointer in cleanup() routines. (Joonas)
v11: Rebase.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/1476378888-7372-1-git-send-email-akash.goel@intel.com
2016-10-13 17:14:48 +00:00
|
|
|
notify_ring(dev_priv->engine[RCS]);
|
2013-05-29 02:22:29 +00:00
|
|
|
if (gt_iir & GT_BSD_USER_INTERRUPT)
|
drm/i915: Allocate intel_engine_cs structure only for the enabled engines
With the possibility of addition of many more number of rings in future,
the drm_i915_private structure could bloat as an array, of type
intel_engine_cs, is embedded inside it.
struct intel_engine_cs engine[I915_NUM_ENGINES];
Though this is still fine as generally there is only a single instance of
drm_i915_private structure used, but not all of the possible rings would be
enabled or active on most of the platforms. Some memory can be saved by
allocating intel_engine_cs structure only for the enabled/active engines.
Currently the engine/ring ID is kept static and dev_priv->engine[] is simply
indexed using the enums defined in intel_engine_id.
To save memory and continue using the static engine/ring IDs, 'engine' is
defined as an array of pointers.
struct intel_engine_cs *engine[I915_NUM_ENGINES];
dev_priv->engine[engine_ID] will be NULL for disabled engine instances.
There is a text size reduction of 928 bytes, from 1028200 to 1027272, for
i915.o file (but for i915.ko file text size remain same as 1193131 bytes).
v2:
- Remove the engine iterator field added in drm_i915_private structure,
instead pass a local iterator variable to the for_each_engine**
macros. (Chris)
- Do away with intel_engine_initialized() and instead directly use the
NULL pointer check on engine pointer. (Chris)
v3:
- Remove for_each_engine_id() macro, as the updated macro for_each_engine()
can be used in place of it. (Chris)
- Protect the access to Render engine Fault register with a NULL check, as
engine specific init is done later in Driver load sequence.
v4:
- Use !!dev_priv->engine[VCS] style for the engine check in getparam. (Chris)
- Kill the superfluous init_engine_lists().
v5:
- Cleanup the intel_engines_init() & intel_engines_setup(), with respect to
allocation of intel_engine_cs structure. (Chris)
v6:
- Rebase.
v7:
- Optimize the for_each_engine_masked() macro. (Chris)
- Change the type of 'iter' local variable to enum intel_engine_id. (Chris)
- Rebase.
v8: Rebase.
v9: Rebase.
v10:
- For index calculation use engine ID instead of pointer based arithmetic in
intel_engine_sync_index() as engine pointers are not contiguous now (Chris)
- For appropriateness, rename local enum variable 'iter' to 'id'. (Joonas)
- Use for_each_engine macro for cleanup in intel_engines_init() and remove
check for NULL engine pointer in cleanup() routines. (Joonas)
v11: Rebase.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/1476378888-7372-1-git-send-email-akash.goel@intel.com
2016-10-13 17:14:48 +00:00
|
|
|
notify_ring(dev_priv->engine[VCS]);
|
2013-05-29 02:22:29 +00:00
|
|
|
if (gt_iir & GT_BLT_USER_INTERRUPT)
|
drm/i915: Allocate intel_engine_cs structure only for the enabled engines
With the possibility of addition of many more number of rings in future,
the drm_i915_private structure could bloat as an array, of type
intel_engine_cs, is embedded inside it.
struct intel_engine_cs engine[I915_NUM_ENGINES];
Though this is still fine as generally there is only a single instance of
drm_i915_private structure used, but not all of the possible rings would be
enabled or active on most of the platforms. Some memory can be saved by
allocating intel_engine_cs structure only for the enabled/active engines.
Currently the engine/ring ID is kept static and dev_priv->engine[] is simply
indexed using the enums defined in intel_engine_id.
To save memory and continue using the static engine/ring IDs, 'engine' is
defined as an array of pointers.
struct intel_engine_cs *engine[I915_NUM_ENGINES];
dev_priv->engine[engine_ID] will be NULL for disabled engine instances.
There is a text size reduction of 928 bytes, from 1028200 to 1027272, for
i915.o file (but for i915.ko file text size remain same as 1193131 bytes).
v2:
- Remove the engine iterator field added in drm_i915_private structure,
instead pass a local iterator variable to the for_each_engine**
macros. (Chris)
- Do away with intel_engine_initialized() and instead directly use the
NULL pointer check on engine pointer. (Chris)
v3:
- Remove for_each_engine_id() macro, as the updated macro for_each_engine()
can be used in place of it. (Chris)
- Protect the access to Render engine Fault register with a NULL check, as
engine specific init is done later in Driver load sequence.
v4:
- Use !!dev_priv->engine[VCS] style for the engine check in getparam. (Chris)
- Kill the superfluous init_engine_lists().
v5:
- Cleanup the intel_engines_init() & intel_engines_setup(), with respect to
allocation of intel_engine_cs structure. (Chris)
v6:
- Rebase.
v7:
- Optimize the for_each_engine_masked() macro. (Chris)
- Change the type of 'iter' local variable to enum intel_engine_id. (Chris)
- Rebase.
v8: Rebase.
v9: Rebase.
v10:
- For index calculation use engine ID instead of pointer based arithmetic in
intel_engine_sync_index() as engine pointers are not contiguous now (Chris)
- For appropriateness, rename local enum variable 'iter' to 'id'. (Joonas)
- Use for_each_engine macro for cleanup in intel_engines_init() and remove
check for NULL engine pointer in cleanup() routines. (Joonas)
v11: Rebase.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/1476378888-7372-1-git-send-email-akash.goel@intel.com
2016-10-13 17:14:48 +00:00
|
|
|
notify_ring(dev_priv->engine[BCS]);
|
2012-03-30 18:24:35 +00:00
|
|
|
|
2013-05-29 02:22:29 +00:00
|
|
|
if (gt_iir & (GT_BLT_CS_ERROR_INTERRUPT |
|
|
|
|
GT_BSD_CS_ERROR_INTERRUPT |
|
2014-11-04 14:52:22 +00:00
|
|
|
GT_RENDER_CS_MASTER_ERROR_INTERRUPT))
|
|
|
|
DRM_DEBUG("Command parser error, gt_iir 0x%08x\n", gt_iir);
|
2012-05-25 23:56:22 +00:00
|
|
|
|
2016-04-13 18:19:57 +00:00
|
|
|
if (gt_iir & GT_PARITY_ERROR(dev_priv))
|
|
|
|
ivybridge_parity_error_irq_handler(dev_priv, gt_iir);
|
2012-03-30 18:24:35 +00:00
|
|
|
}
|
|
|
|
|
2017-05-17 12:10:06 +00:00
|
|
|
static void
|
2016-03-16 11:00:37 +00:00
|
|
|
gen8_cs_irq_handler(struct intel_engine_cs *engine, u32 iir, int test_shift)
|
2015-10-20 09:23:52 +00:00
|
|
|
{
|
2017-09-22 12:43:03 +00:00
|
|
|
struct intel_engine_execlists * const execlists = &engine->execlists;
|
2017-03-16 12:56:18 +00:00
|
|
|
bool tasklet = false;
|
2017-01-24 15:20:21 +00:00
|
|
|
|
|
|
|
if (iir & (GT_CONTEXT_SWITCH_INTERRUPT << test_shift)) {
|
drm/i915/execlists: Preemption!
When we write to ELSP, it triggers a context preemption at the earliest
arbitration point (3DPRIMITIVE, some PIPECONTROLs, a few other
operations and the explicit MI_ARB_CHECK). If this is to the same
context, it triggers a LITE_RESTORE where the RING_TAIL is merely
updated (used currently to chain requests from the same context
together, avoiding bubbles). However, if it is to a different context, a
full context-switch is performed and it will start to execute the new
context saving the image of the old for later execution.
Previously we avoided preemption by only submitting a new context when
the old was idle. But now we wish embrace it, and if the new request has
a higher priority than the currently executing request, we write to the
ELSP regardless, thus triggering preemption, but we tell the GPU to
switch to our special preemption context (not the target). In the
context-switch interrupt handler, we know that the previous contexts
have finished execution and so can unwind all the incomplete requests
and compute the new highest priority request to execute.
It would be feasible to avoid the switch-to-idle intermediate by
programming the ELSP with the target context. The difficulty is in
tracking which request that should be whilst maintaining the dependency
change, the error comes in with coalesced requests. As we only track the
most recent request and its priority, we may run into the issue of being
tricked in preempting a high priority request that was followed by a
low priority request from the same context (e.g. for PI); worse still
that earlier request may be our own dependency and the order then broken
by preemption. By injecting the switch-to-idle and then recomputing the
priority queue, we avoid the issue with tracking in-flight coalesced
requests. Having tried the preempt-to-busy approach, and failed to find
a way around the coalesced priority issue, Michal's original proposal to
inject an idle context (based on handling GuC preemption) succeeds.
The current heuristic for deciding when to preempt are only if the new
request is of higher priority, and has the privileged priority of
greater than 0. Note that the scheduler remains unfair!
v2: Disable for gen8 (bdw/bsw) as we need additional w/a for GPGPU.
Since, the feature is now conditional and not always available when we
have a scheduler, make it known via the HAS_SCHEDULER GETPARAM (now a
capability mask).
v3: Stylistic tweaks.
v4: Appease Joonas with a snippet of kerneldoc, only to fuel to fire of
the preempt vs preempting debate.
Suggested-by: Michal Winiarski <michal.winiarski@intel.com>
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Michal Winiarski <michal.winiarski@intel.com>
Cc: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Cc: Arkadiusz Hiler <arkadiusz.hiler@intel.com>
Cc: Mika Kuoppala <mika.kuoppala@intel.com>
Cc: Ben Widawsky <benjamin.widawsky@intel.com>
Cc: Zhenyu Wang <zhenyuw@linux.intel.com>
Cc: Zhi Wang <zhi.a.wang@intel.com>
Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20171003203453.15692-8-chris@chris-wilson.co.uk
2017-10-03 20:34:52 +00:00
|
|
|
__set_bit(ENGINE_IRQ_EXECLIST, &engine->irq_posted);
|
|
|
|
tasklet = true;
|
2017-01-24 15:20:21 +00:00
|
|
|
}
|
2017-03-16 12:56:18 +00:00
|
|
|
|
|
|
|
if (iir & (GT_RENDER_USER_INTERRUPT << test_shift)) {
|
|
|
|
notify_ring(engine);
|
2017-09-19 19:38:44 +00:00
|
|
|
tasklet |= i915_modparams.enable_guc_submission;
|
2017-03-16 12:56:18 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if (tasklet)
|
2017-09-22 12:43:03 +00:00
|
|
|
tasklet_hi_schedule(&execlists->irq_tasklet);
|
2015-10-20 09:23:52 +00:00
|
|
|
}
|
|
|
|
|
2016-04-13 18:19:58 +00:00
|
|
|
static irqreturn_t gen8_gt_irq_ack(struct drm_i915_private *dev_priv,
|
|
|
|
u32 master_ctl,
|
|
|
|
u32 gt_iir[4])
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
{
|
|
|
|
irqreturn_t ret = IRQ_NONE;
|
|
|
|
|
|
|
|
if (master_ctl & (GEN8_GT_RCS_IRQ | GEN8_GT_BCS_IRQ)) {
|
2016-04-13 18:19:58 +00:00
|
|
|
gt_iir[0] = I915_READ_FW(GEN8_GT_IIR(0));
|
|
|
|
if (gt_iir[0]) {
|
|
|
|
I915_WRITE_FW(GEN8_GT_IIR(0), gt_iir[0]);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
ret = IRQ_HANDLED;
|
|
|
|
} else
|
|
|
|
DRM_ERROR("The master control interrupt lied (GT0)!\n");
|
|
|
|
}
|
|
|
|
|
2014-04-17 02:37:38 +00:00
|
|
|
if (master_ctl & (GEN8_GT_VCS1_IRQ | GEN8_GT_VCS2_IRQ)) {
|
2016-04-13 18:19:58 +00:00
|
|
|
gt_iir[1] = I915_READ_FW(GEN8_GT_IIR(1));
|
|
|
|
if (gt_iir[1]) {
|
|
|
|
I915_WRITE_FW(GEN8_GT_IIR(1), gt_iir[1]);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
ret = IRQ_HANDLED;
|
2014-05-15 17:58:08 +00:00
|
|
|
} else
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
DRM_ERROR("The master control interrupt lied (GT1)!\n");
|
2014-05-15 17:58:08 +00:00
|
|
|
}
|
|
|
|
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
if (master_ctl & GEN8_GT_VECS_IRQ) {
|
2016-04-13 18:19:58 +00:00
|
|
|
gt_iir[3] = I915_READ_FW(GEN8_GT_IIR(3));
|
|
|
|
if (gt_iir[3]) {
|
|
|
|
I915_WRITE_FW(GEN8_GT_IIR(3), gt_iir[3]);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
ret = IRQ_HANDLED;
|
|
|
|
} else
|
|
|
|
DRM_ERROR("The master control interrupt lied (GT3)!\n");
|
|
|
|
}
|
|
|
|
|
2016-10-12 16:24:31 +00:00
|
|
|
if (master_ctl & (GEN8_GT_PM_IRQ | GEN8_GT_GUC_IRQ)) {
|
2016-04-13 18:19:58 +00:00
|
|
|
gt_iir[2] = I915_READ_FW(GEN8_GT_IIR(2));
|
2016-10-12 16:24:31 +00:00
|
|
|
if (gt_iir[2] & (dev_priv->pm_rps_events |
|
|
|
|
dev_priv->pm_guc_events)) {
|
2015-04-07 15:21:04 +00:00
|
|
|
I915_WRITE_FW(GEN8_GT_IIR(2),
|
2016-10-12 16:24:31 +00:00
|
|
|
gt_iir[2] & (dev_priv->pm_rps_events |
|
|
|
|
dev_priv->pm_guc_events));
|
2014-06-16 15:10:59 +00:00
|
|
|
ret = IRQ_HANDLED;
|
2014-05-15 17:58:08 +00:00
|
|
|
} else
|
|
|
|
DRM_ERROR("The master control interrupt lied (PM)!\n");
|
|
|
|
}
|
|
|
|
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2016-04-13 18:19:58 +00:00
|
|
|
static void gen8_gt_irq_handler(struct drm_i915_private *dev_priv,
|
|
|
|
u32 gt_iir[4])
|
|
|
|
{
|
|
|
|
if (gt_iir[0]) {
|
drm/i915: Allocate intel_engine_cs structure only for the enabled engines
With the possibility of addition of many more number of rings in future,
the drm_i915_private structure could bloat as an array, of type
intel_engine_cs, is embedded inside it.
struct intel_engine_cs engine[I915_NUM_ENGINES];
Though this is still fine as generally there is only a single instance of
drm_i915_private structure used, but not all of the possible rings would be
enabled or active on most of the platforms. Some memory can be saved by
allocating intel_engine_cs structure only for the enabled/active engines.
Currently the engine/ring ID is kept static and dev_priv->engine[] is simply
indexed using the enums defined in intel_engine_id.
To save memory and continue using the static engine/ring IDs, 'engine' is
defined as an array of pointers.
struct intel_engine_cs *engine[I915_NUM_ENGINES];
dev_priv->engine[engine_ID] will be NULL for disabled engine instances.
There is a text size reduction of 928 bytes, from 1028200 to 1027272, for
i915.o file (but for i915.ko file text size remain same as 1193131 bytes).
v2:
- Remove the engine iterator field added in drm_i915_private structure,
instead pass a local iterator variable to the for_each_engine**
macros. (Chris)
- Do away with intel_engine_initialized() and instead directly use the
NULL pointer check on engine pointer. (Chris)
v3:
- Remove for_each_engine_id() macro, as the updated macro for_each_engine()
can be used in place of it. (Chris)
- Protect the access to Render engine Fault register with a NULL check, as
engine specific init is done later in Driver load sequence.
v4:
- Use !!dev_priv->engine[VCS] style for the engine check in getparam. (Chris)
- Kill the superfluous init_engine_lists().
v5:
- Cleanup the intel_engines_init() & intel_engines_setup(), with respect to
allocation of intel_engine_cs structure. (Chris)
v6:
- Rebase.
v7:
- Optimize the for_each_engine_masked() macro. (Chris)
- Change the type of 'iter' local variable to enum intel_engine_id. (Chris)
- Rebase.
v8: Rebase.
v9: Rebase.
v10:
- For index calculation use engine ID instead of pointer based arithmetic in
intel_engine_sync_index() as engine pointers are not contiguous now (Chris)
- For appropriateness, rename local enum variable 'iter' to 'id'. (Joonas)
- Use for_each_engine macro for cleanup in intel_engines_init() and remove
check for NULL engine pointer in cleanup() routines. (Joonas)
v11: Rebase.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/1476378888-7372-1-git-send-email-akash.goel@intel.com
2016-10-13 17:14:48 +00:00
|
|
|
gen8_cs_irq_handler(dev_priv->engine[RCS],
|
2016-04-13 18:19:58 +00:00
|
|
|
gt_iir[0], GEN8_RCS_IRQ_SHIFT);
|
drm/i915: Allocate intel_engine_cs structure only for the enabled engines
With the possibility of addition of many more number of rings in future,
the drm_i915_private structure could bloat as an array, of type
intel_engine_cs, is embedded inside it.
struct intel_engine_cs engine[I915_NUM_ENGINES];
Though this is still fine as generally there is only a single instance of
drm_i915_private structure used, but not all of the possible rings would be
enabled or active on most of the platforms. Some memory can be saved by
allocating intel_engine_cs structure only for the enabled/active engines.
Currently the engine/ring ID is kept static and dev_priv->engine[] is simply
indexed using the enums defined in intel_engine_id.
To save memory and continue using the static engine/ring IDs, 'engine' is
defined as an array of pointers.
struct intel_engine_cs *engine[I915_NUM_ENGINES];
dev_priv->engine[engine_ID] will be NULL for disabled engine instances.
There is a text size reduction of 928 bytes, from 1028200 to 1027272, for
i915.o file (but for i915.ko file text size remain same as 1193131 bytes).
v2:
- Remove the engine iterator field added in drm_i915_private structure,
instead pass a local iterator variable to the for_each_engine**
macros. (Chris)
- Do away with intel_engine_initialized() and instead directly use the
NULL pointer check on engine pointer. (Chris)
v3:
- Remove for_each_engine_id() macro, as the updated macro for_each_engine()
can be used in place of it. (Chris)
- Protect the access to Render engine Fault register with a NULL check, as
engine specific init is done later in Driver load sequence.
v4:
- Use !!dev_priv->engine[VCS] style for the engine check in getparam. (Chris)
- Kill the superfluous init_engine_lists().
v5:
- Cleanup the intel_engines_init() & intel_engines_setup(), with respect to
allocation of intel_engine_cs structure. (Chris)
v6:
- Rebase.
v7:
- Optimize the for_each_engine_masked() macro. (Chris)
- Change the type of 'iter' local variable to enum intel_engine_id. (Chris)
- Rebase.
v8: Rebase.
v9: Rebase.
v10:
- For index calculation use engine ID instead of pointer based arithmetic in
intel_engine_sync_index() as engine pointers are not contiguous now (Chris)
- For appropriateness, rename local enum variable 'iter' to 'id'. (Joonas)
- Use for_each_engine macro for cleanup in intel_engines_init() and remove
check for NULL engine pointer in cleanup() routines. (Joonas)
v11: Rebase.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/1476378888-7372-1-git-send-email-akash.goel@intel.com
2016-10-13 17:14:48 +00:00
|
|
|
gen8_cs_irq_handler(dev_priv->engine[BCS],
|
2016-04-13 18:19:58 +00:00
|
|
|
gt_iir[0], GEN8_BCS_IRQ_SHIFT);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (gt_iir[1]) {
|
drm/i915: Allocate intel_engine_cs structure only for the enabled engines
With the possibility of addition of many more number of rings in future,
the drm_i915_private structure could bloat as an array, of type
intel_engine_cs, is embedded inside it.
struct intel_engine_cs engine[I915_NUM_ENGINES];
Though this is still fine as generally there is only a single instance of
drm_i915_private structure used, but not all of the possible rings would be
enabled or active on most of the platforms. Some memory can be saved by
allocating intel_engine_cs structure only for the enabled/active engines.
Currently the engine/ring ID is kept static and dev_priv->engine[] is simply
indexed using the enums defined in intel_engine_id.
To save memory and continue using the static engine/ring IDs, 'engine' is
defined as an array of pointers.
struct intel_engine_cs *engine[I915_NUM_ENGINES];
dev_priv->engine[engine_ID] will be NULL for disabled engine instances.
There is a text size reduction of 928 bytes, from 1028200 to 1027272, for
i915.o file (but for i915.ko file text size remain same as 1193131 bytes).
v2:
- Remove the engine iterator field added in drm_i915_private structure,
instead pass a local iterator variable to the for_each_engine**
macros. (Chris)
- Do away with intel_engine_initialized() and instead directly use the
NULL pointer check on engine pointer. (Chris)
v3:
- Remove for_each_engine_id() macro, as the updated macro for_each_engine()
can be used in place of it. (Chris)
- Protect the access to Render engine Fault register with a NULL check, as
engine specific init is done later in Driver load sequence.
v4:
- Use !!dev_priv->engine[VCS] style for the engine check in getparam. (Chris)
- Kill the superfluous init_engine_lists().
v5:
- Cleanup the intel_engines_init() & intel_engines_setup(), with respect to
allocation of intel_engine_cs structure. (Chris)
v6:
- Rebase.
v7:
- Optimize the for_each_engine_masked() macro. (Chris)
- Change the type of 'iter' local variable to enum intel_engine_id. (Chris)
- Rebase.
v8: Rebase.
v9: Rebase.
v10:
- For index calculation use engine ID instead of pointer based arithmetic in
intel_engine_sync_index() as engine pointers are not contiguous now (Chris)
- For appropriateness, rename local enum variable 'iter' to 'id'. (Joonas)
- Use for_each_engine macro for cleanup in intel_engines_init() and remove
check for NULL engine pointer in cleanup() routines. (Joonas)
v11: Rebase.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/1476378888-7372-1-git-send-email-akash.goel@intel.com
2016-10-13 17:14:48 +00:00
|
|
|
gen8_cs_irq_handler(dev_priv->engine[VCS],
|
2016-04-13 18:19:58 +00:00
|
|
|
gt_iir[1], GEN8_VCS1_IRQ_SHIFT);
|
drm/i915: Allocate intel_engine_cs structure only for the enabled engines
With the possibility of addition of many more number of rings in future,
the drm_i915_private structure could bloat as an array, of type
intel_engine_cs, is embedded inside it.
struct intel_engine_cs engine[I915_NUM_ENGINES];
Though this is still fine as generally there is only a single instance of
drm_i915_private structure used, but not all of the possible rings would be
enabled or active on most of the platforms. Some memory can be saved by
allocating intel_engine_cs structure only for the enabled/active engines.
Currently the engine/ring ID is kept static and dev_priv->engine[] is simply
indexed using the enums defined in intel_engine_id.
To save memory and continue using the static engine/ring IDs, 'engine' is
defined as an array of pointers.
struct intel_engine_cs *engine[I915_NUM_ENGINES];
dev_priv->engine[engine_ID] will be NULL for disabled engine instances.
There is a text size reduction of 928 bytes, from 1028200 to 1027272, for
i915.o file (but for i915.ko file text size remain same as 1193131 bytes).
v2:
- Remove the engine iterator field added in drm_i915_private structure,
instead pass a local iterator variable to the for_each_engine**
macros. (Chris)
- Do away with intel_engine_initialized() and instead directly use the
NULL pointer check on engine pointer. (Chris)
v3:
- Remove for_each_engine_id() macro, as the updated macro for_each_engine()
can be used in place of it. (Chris)
- Protect the access to Render engine Fault register with a NULL check, as
engine specific init is done later in Driver load sequence.
v4:
- Use !!dev_priv->engine[VCS] style for the engine check in getparam. (Chris)
- Kill the superfluous init_engine_lists().
v5:
- Cleanup the intel_engines_init() & intel_engines_setup(), with respect to
allocation of intel_engine_cs structure. (Chris)
v6:
- Rebase.
v7:
- Optimize the for_each_engine_masked() macro. (Chris)
- Change the type of 'iter' local variable to enum intel_engine_id. (Chris)
- Rebase.
v8: Rebase.
v9: Rebase.
v10:
- For index calculation use engine ID instead of pointer based arithmetic in
intel_engine_sync_index() as engine pointers are not contiguous now (Chris)
- For appropriateness, rename local enum variable 'iter' to 'id'. (Joonas)
- Use for_each_engine macro for cleanup in intel_engines_init() and remove
check for NULL engine pointer in cleanup() routines. (Joonas)
v11: Rebase.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/1476378888-7372-1-git-send-email-akash.goel@intel.com
2016-10-13 17:14:48 +00:00
|
|
|
gen8_cs_irq_handler(dev_priv->engine[VCS2],
|
2016-04-13 18:19:58 +00:00
|
|
|
gt_iir[1], GEN8_VCS2_IRQ_SHIFT);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (gt_iir[3])
|
drm/i915: Allocate intel_engine_cs structure only for the enabled engines
With the possibility of addition of many more number of rings in future,
the drm_i915_private structure could bloat as an array, of type
intel_engine_cs, is embedded inside it.
struct intel_engine_cs engine[I915_NUM_ENGINES];
Though this is still fine as generally there is only a single instance of
drm_i915_private structure used, but not all of the possible rings would be
enabled or active on most of the platforms. Some memory can be saved by
allocating intel_engine_cs structure only for the enabled/active engines.
Currently the engine/ring ID is kept static and dev_priv->engine[] is simply
indexed using the enums defined in intel_engine_id.
To save memory and continue using the static engine/ring IDs, 'engine' is
defined as an array of pointers.
struct intel_engine_cs *engine[I915_NUM_ENGINES];
dev_priv->engine[engine_ID] will be NULL for disabled engine instances.
There is a text size reduction of 928 bytes, from 1028200 to 1027272, for
i915.o file (but for i915.ko file text size remain same as 1193131 bytes).
v2:
- Remove the engine iterator field added in drm_i915_private structure,
instead pass a local iterator variable to the for_each_engine**
macros. (Chris)
- Do away with intel_engine_initialized() and instead directly use the
NULL pointer check on engine pointer. (Chris)
v3:
- Remove for_each_engine_id() macro, as the updated macro for_each_engine()
can be used in place of it. (Chris)
- Protect the access to Render engine Fault register with a NULL check, as
engine specific init is done later in Driver load sequence.
v4:
- Use !!dev_priv->engine[VCS] style for the engine check in getparam. (Chris)
- Kill the superfluous init_engine_lists().
v5:
- Cleanup the intel_engines_init() & intel_engines_setup(), with respect to
allocation of intel_engine_cs structure. (Chris)
v6:
- Rebase.
v7:
- Optimize the for_each_engine_masked() macro. (Chris)
- Change the type of 'iter' local variable to enum intel_engine_id. (Chris)
- Rebase.
v8: Rebase.
v9: Rebase.
v10:
- For index calculation use engine ID instead of pointer based arithmetic in
intel_engine_sync_index() as engine pointers are not contiguous now (Chris)
- For appropriateness, rename local enum variable 'iter' to 'id'. (Joonas)
- Use for_each_engine macro for cleanup in intel_engines_init() and remove
check for NULL engine pointer in cleanup() routines. (Joonas)
v11: Rebase.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/1476378888-7372-1-git-send-email-akash.goel@intel.com
2016-10-13 17:14:48 +00:00
|
|
|
gen8_cs_irq_handler(dev_priv->engine[VECS],
|
2016-04-13 18:19:58 +00:00
|
|
|
gt_iir[3], GEN8_VECS_IRQ_SHIFT);
|
|
|
|
|
|
|
|
if (gt_iir[2] & dev_priv->pm_rps_events)
|
|
|
|
gen6_rps_irq_handler(dev_priv, gt_iir[2]);
|
2016-10-12 16:24:31 +00:00
|
|
|
|
|
|
|
if (gt_iir[2] & dev_priv->pm_guc_events)
|
|
|
|
gen9_guc_irq_handler(dev_priv, gt_iir[2]);
|
2016-04-13 18:19:58 +00:00
|
|
|
}
|
|
|
|
|
2015-07-20 21:43:39 +00:00
|
|
|
static bool bxt_port_hotplug_long_detect(enum port port, u32 val)
|
|
|
|
{
|
|
|
|
switch (port) {
|
|
|
|
case PORT_A:
|
2015-08-27 20:56:00 +00:00
|
|
|
return val & PORTA_HOTPLUG_LONG_DETECT;
|
2015-07-20 21:43:39 +00:00
|
|
|
case PORT_B:
|
|
|
|
return val & PORTB_HOTPLUG_LONG_DETECT;
|
|
|
|
case PORT_C:
|
|
|
|
return val & PORTC_HOTPLUG_LONG_DETECT;
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-08-27 20:56:02 +00:00
|
|
|
static bool spt_port_hotplug2_long_detect(enum port port, u32 val)
|
|
|
|
{
|
|
|
|
switch (port) {
|
|
|
|
case PORT_E:
|
|
|
|
return val & PORTE_HOTPLUG_LONG_DETECT;
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-08-27 20:56:07 +00:00
|
|
|
static bool spt_port_hotplug_long_detect(enum port port, u32 val)
|
|
|
|
{
|
|
|
|
switch (port) {
|
|
|
|
case PORT_A:
|
|
|
|
return val & PORTA_HOTPLUG_LONG_DETECT;
|
|
|
|
case PORT_B:
|
|
|
|
return val & PORTB_HOTPLUG_LONG_DETECT;
|
|
|
|
case PORT_C:
|
|
|
|
return val & PORTC_HOTPLUG_LONG_DETECT;
|
|
|
|
case PORT_D:
|
|
|
|
return val & PORTD_HOTPLUG_LONG_DETECT;
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-08-27 20:56:03 +00:00
|
|
|
static bool ilk_port_hotplug_long_detect(enum port port, u32 val)
|
|
|
|
{
|
|
|
|
switch (port) {
|
|
|
|
case PORT_A:
|
|
|
|
return val & DIGITAL_PORTA_HOTPLUG_LONG_DETECT;
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-05-28 12:43:53 +00:00
|
|
|
static bool pch_port_hotplug_long_detect(enum port port, u32 val)
|
2014-06-18 01:29:35 +00:00
|
|
|
{
|
|
|
|
switch (port) {
|
|
|
|
case PORT_B:
|
2015-05-28 12:43:53 +00:00
|
|
|
return val & PORTB_HOTPLUG_LONG_DETECT;
|
2014-06-18 01:29:35 +00:00
|
|
|
case PORT_C:
|
2015-05-28 12:43:53 +00:00
|
|
|
return val & PORTC_HOTPLUG_LONG_DETECT;
|
2014-06-18 01:29:35 +00:00
|
|
|
case PORT_D:
|
2015-05-28 12:43:53 +00:00
|
|
|
return val & PORTD_HOTPLUG_LONG_DETECT;
|
|
|
|
default:
|
|
|
|
return false;
|
2014-06-18 01:29:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-05-28 12:43:53 +00:00
|
|
|
static bool i9xx_port_hotplug_long_detect(enum port port, u32 val)
|
2014-06-18 01:29:35 +00:00
|
|
|
{
|
|
|
|
switch (port) {
|
|
|
|
case PORT_B:
|
2015-05-28 12:43:53 +00:00
|
|
|
return val & PORTB_HOTPLUG_INT_LONG_PULSE;
|
2014-06-18 01:29:35 +00:00
|
|
|
case PORT_C:
|
2015-05-28 12:43:53 +00:00
|
|
|
return val & PORTC_HOTPLUG_INT_LONG_PULSE;
|
2014-06-18 01:29:35 +00:00
|
|
|
case PORT_D:
|
2015-05-28 12:43:53 +00:00
|
|
|
return val & PORTD_HOTPLUG_INT_LONG_PULSE;
|
|
|
|
default:
|
|
|
|
return false;
|
2014-06-18 01:29:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-08-28 18:26:27 +00:00
|
|
|
/*
|
|
|
|
* Get a bit mask of pins that have triggered, and which ones may be long.
|
|
|
|
* This can be called multiple times with the same masks to accumulate
|
|
|
|
* hotplug detection results from several registers.
|
|
|
|
*
|
|
|
|
* Note that the caller is expected to zero out the masks initially.
|
|
|
|
*/
|
2015-07-21 22:32:44 +00:00
|
|
|
static void intel_get_hpd_pins(u32 *pin_mask, u32 *long_mask,
|
2015-06-18 10:06:17 +00:00
|
|
|
u32 hotplug_trigger, u32 dig_hotplug_reg,
|
2015-07-21 22:32:44 +00:00
|
|
|
const u32 hpd[HPD_NUM_PINS],
|
|
|
|
bool long_pulse_detect(enum port port, u32 val))
|
2015-05-28 12:43:53 +00:00
|
|
|
{
|
2015-06-18 10:06:17 +00:00
|
|
|
enum port port;
|
2015-05-28 12:43:53 +00:00
|
|
|
int i;
|
|
|
|
|
|
|
|
for_each_hpd_pin(i) {
|
2015-06-18 10:06:17 +00:00
|
|
|
if ((hpd[i] & hotplug_trigger) == 0)
|
|
|
|
continue;
|
2015-05-28 12:43:53 +00:00
|
|
|
|
2015-06-18 10:06:17 +00:00
|
|
|
*pin_mask |= BIT(i);
|
|
|
|
|
2017-08-11 18:26:49 +00:00
|
|
|
port = intel_hpd_pin_to_port(i);
|
|
|
|
if (port == PORT_NONE)
|
2015-07-21 22:32:45 +00:00
|
|
|
continue;
|
|
|
|
|
2015-07-21 22:32:44 +00:00
|
|
|
if (long_pulse_detect(port, dig_hotplug_reg))
|
2015-06-18 10:06:17 +00:00
|
|
|
*long_mask |= BIT(i);
|
2015-05-28 12:43:53 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
DRM_DEBUG_DRIVER("hotplug event received, stat 0x%08x, dig 0x%08x, pins 0x%08x\n",
|
|
|
|
hotplug_trigger, dig_hotplug_reg, *pin_mask);
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void gmbus_irq_handler(struct drm_i915_private *dev_priv)
|
2012-12-01 12:53:44 +00:00
|
|
|
{
|
2012-12-01 12:53:45 +00:00
|
|
|
wake_up_all(&dev_priv->gmbus_wait_queue);
|
2012-12-01 12:53:44 +00:00
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void dp_aux_irq_handler(struct drm_i915_private *dev_priv)
|
2012-12-01 12:53:47 +00:00
|
|
|
{
|
drm/i915: irq-drive the dp aux communication
At least on the platforms that have a dp aux irq and also have it
enabled - vlvhsw should have one, too. But I don't have a machine to
test this on. Judging from docs there's no dp aux interrupt for gm45.
Also, I only have an ivb cpu edp machine, so the dp aux A code for
snb/ilk is untested.
For dpcd probing when nothing is connected it slashes about 5ms of cpu
time (cpu time is now negligible), which agrees with 3 * 5 400 usec
timeouts.
A previous version of this patch increases the time required to go
through the dp_detect cycle (which includes reading the edid) from
around 33 ms to around 40 ms. Experiments indicated that this is
purely due to the irq latency - the hw doesn't allow us to queue up
dp aux transactions and hence irq latency directly affects throughput.
gmbus is much better, there we have a 8 byte buffer, and we get the
irq once another 4 bytes can be queued up.
But by using the pm_qos interface to request the lowest possible cpu
wake-up latency this slowdown completely disappeared.
Since all our output detection logic is single-threaded with the
mode_config mutex right now anyway, I've decide not ot play fancy and
to just reuse the gmbus wait queue. But this would definitely prep the
way to run dp detection on different ports in parallel
v2: Add a timeout for dp aux transfers when using interrupts - the hw
_does_ prevent this with the hw-based 400 usec timeout, but if the
irq somehow doesn't arrive we're screwed. Lesson learned while
developing this ;-)
v3: While at it also convert the busy-loop to wait_for_atomic, so that
we don't run the risk of an infinite loop any more.
v4: Ensure we have the smallest possible irq latency by using the
pm_qos interface.
v5: Add a comment to the code to explain why we frob pm_qos. Suggested
by Chris Wilson.
v6: Disable dp irq for vlv, that's easier than trying to get at docs
and hw.
v7: Squash in a fix for Haswell that Paulo Zanoni tracked down - the
dp aux registers aren't at a fixed offset any more, but can be on the
PCH while the DP port is on the cpu die.
Reviewed-by: Imre Deak <imre.deak@intel.com> (v6)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2012-12-01 12:53:48 +00:00
|
|
|
wake_up_all(&dev_priv->gmbus_wait_queue);
|
2012-12-01 12:53:47 +00:00
|
|
|
}
|
|
|
|
|
2013-10-15 17:55:27 +00:00
|
|
|
#if defined(CONFIG_DEBUG_FS)
|
2016-05-06 13:48:28 +00:00
|
|
|
static void display_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
|
|
|
|
enum pipe pipe,
|
2013-10-18 14:37:07 +00:00
|
|
|
uint32_t crc0, uint32_t crc1,
|
|
|
|
uint32_t crc2, uint32_t crc3,
|
|
|
|
uint32_t crc4)
|
2013-10-15 17:55:27 +00:00
|
|
|
{
|
|
|
|
struct intel_pipe_crc *pipe_crc = &dev_priv->pipe_crc[pipe];
|
|
|
|
struct intel_pipe_crc_entry *entry;
|
2017-01-10 13:43:04 +00:00
|
|
|
struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
|
|
|
|
struct drm_driver *driver = dev_priv->drm.driver;
|
|
|
|
uint32_t crcs[5];
|
2013-10-15 17:55:30 +00:00
|
|
|
int head, tail;
|
2013-10-15 17:55:29 +00:00
|
|
|
|
2013-10-21 13:29:30 +00:00
|
|
|
spin_lock(&pipe_crc->lock);
|
2017-01-10 13:43:04 +00:00
|
|
|
if (pipe_crc->source) {
|
|
|
|
if (!pipe_crc->entries) {
|
|
|
|
spin_unlock(&pipe_crc->lock);
|
|
|
|
DRM_DEBUG_KMS("spurious interrupt\n");
|
|
|
|
return;
|
|
|
|
}
|
2013-10-21 13:29:30 +00:00
|
|
|
|
2017-01-10 13:43:04 +00:00
|
|
|
head = pipe_crc->head;
|
|
|
|
tail = pipe_crc->tail;
|
2013-10-15 17:55:29 +00:00
|
|
|
|
2017-01-10 13:43:04 +00:00
|
|
|
if (CIRC_SPACE(head, tail, INTEL_PIPE_CRC_ENTRIES_NR) < 1) {
|
|
|
|
spin_unlock(&pipe_crc->lock);
|
|
|
|
DRM_ERROR("CRC buffer overflowing\n");
|
|
|
|
return;
|
|
|
|
}
|
2013-10-15 17:55:29 +00:00
|
|
|
|
2017-01-10 13:43:04 +00:00
|
|
|
entry = &pipe_crc->entries[head];
|
2013-10-15 17:55:27 +00:00
|
|
|
|
2017-01-10 13:43:04 +00:00
|
|
|
entry->frame = driver->get_vblank_counter(&dev_priv->drm, pipe);
|
|
|
|
entry->crc[0] = crc0;
|
|
|
|
entry->crc[1] = crc1;
|
|
|
|
entry->crc[2] = crc2;
|
|
|
|
entry->crc[3] = crc3;
|
|
|
|
entry->crc[4] = crc4;
|
2013-10-15 17:55:29 +00:00
|
|
|
|
2017-01-10 13:43:04 +00:00
|
|
|
head = (head + 1) & (INTEL_PIPE_CRC_ENTRIES_NR - 1);
|
|
|
|
pipe_crc->head = head;
|
2013-10-21 13:29:30 +00:00
|
|
|
|
2017-01-10 13:43:04 +00:00
|
|
|
spin_unlock(&pipe_crc->lock);
|
2013-10-15 17:55:40 +00:00
|
|
|
|
2017-01-10 13:43:04 +00:00
|
|
|
wake_up_interruptible(&pipe_crc->wq);
|
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* For some not yet identified reason, the first CRC is
|
|
|
|
* bonkers. So let's just wait for the next vblank and read
|
|
|
|
* out the buggy result.
|
|
|
|
*
|
2017-09-28 00:20:40 +00:00
|
|
|
* On GEN8+ sometimes the second CRC is bonkers as well, so
|
2017-01-10 13:43:04 +00:00
|
|
|
* don't trust that one either.
|
|
|
|
*/
|
|
|
|
if (pipe_crc->skipped == 0 ||
|
2017-09-28 00:20:40 +00:00
|
|
|
(INTEL_GEN(dev_priv) >= 8 && pipe_crc->skipped == 1)) {
|
2017-01-10 13:43:04 +00:00
|
|
|
pipe_crc->skipped++;
|
|
|
|
spin_unlock(&pipe_crc->lock);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
spin_unlock(&pipe_crc->lock);
|
|
|
|
crcs[0] = crc0;
|
|
|
|
crcs[1] = crc1;
|
|
|
|
crcs[2] = crc2;
|
|
|
|
crcs[3] = crc3;
|
|
|
|
crcs[4] = crc4;
|
2017-01-10 13:43:05 +00:00
|
|
|
drm_crtc_add_crc_entry(&crtc->base, true,
|
2017-05-24 14:51:47 +00:00
|
|
|
drm_crtc_accurate_vblank_count(&crtc->base),
|
2017-01-10 13:43:05 +00:00
|
|
|
crcs);
|
2017-01-10 13:43:04 +00:00
|
|
|
}
|
2013-10-15 17:55:27 +00:00
|
|
|
}
|
2013-10-18 14:37:07 +00:00
|
|
|
#else
|
|
|
|
static inline void
|
2016-05-06 13:48:28 +00:00
|
|
|
display_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
|
|
|
|
enum pipe pipe,
|
2013-10-18 14:37:07 +00:00
|
|
|
uint32_t crc0, uint32_t crc1,
|
|
|
|
uint32_t crc2, uint32_t crc3,
|
|
|
|
uint32_t crc4) {}
|
|
|
|
#endif
|
|
|
|
|
2013-10-16 20:55:46 +00:00
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void hsw_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
|
|
|
|
enum pipe pipe)
|
2013-10-16 20:55:52 +00:00
|
|
|
{
|
2016-05-06 13:48:28 +00:00
|
|
|
display_pipe_crc_irq_handler(dev_priv, pipe,
|
2013-10-18 14:37:07 +00:00
|
|
|
I915_READ(PIPE_CRC_RES_1_IVB(pipe)),
|
|
|
|
0, 0, 0, 0);
|
2013-10-16 20:55:52 +00:00
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void ivb_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
|
|
|
|
enum pipe pipe)
|
2013-10-16 20:55:46 +00:00
|
|
|
{
|
2016-05-06 13:48:28 +00:00
|
|
|
display_pipe_crc_irq_handler(dev_priv, pipe,
|
2013-10-18 14:37:07 +00:00
|
|
|
I915_READ(PIPE_CRC_RES_1_IVB(pipe)),
|
|
|
|
I915_READ(PIPE_CRC_RES_2_IVB(pipe)),
|
|
|
|
I915_READ(PIPE_CRC_RES_3_IVB(pipe)),
|
|
|
|
I915_READ(PIPE_CRC_RES_4_IVB(pipe)),
|
|
|
|
I915_READ(PIPE_CRC_RES_5_IVB(pipe)));
|
2013-10-16 20:55:46 +00:00
|
|
|
}
|
2013-10-16 20:55:48 +00:00
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void i9xx_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
|
|
|
|
enum pipe pipe)
|
2013-10-16 20:55:48 +00:00
|
|
|
{
|
2013-10-16 20:55:53 +00:00
|
|
|
uint32_t res1, res2;
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
if (INTEL_GEN(dev_priv) >= 3)
|
2013-10-16 20:55:53 +00:00
|
|
|
res1 = I915_READ(PIPE_CRC_RES_RES1_I915(pipe));
|
|
|
|
else
|
|
|
|
res1 = 0;
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv))
|
2013-10-16 20:55:53 +00:00
|
|
|
res2 = I915_READ(PIPE_CRC_RES_RES2_G4X(pipe));
|
|
|
|
else
|
|
|
|
res2 = 0;
|
2013-10-16 20:55:48 +00:00
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
display_pipe_crc_irq_handler(dev_priv, pipe,
|
2013-10-18 14:37:07 +00:00
|
|
|
I915_READ(PIPE_CRC_RES_RED(pipe)),
|
|
|
|
I915_READ(PIPE_CRC_RES_GREEN(pipe)),
|
|
|
|
I915_READ(PIPE_CRC_RES_BLUE(pipe)),
|
|
|
|
res1, res2);
|
2013-10-16 20:55:48 +00:00
|
|
|
}
|
2013-10-15 17:55:27 +00:00
|
|
|
|
2013-08-15 14:51:32 +00:00
|
|
|
/* The RPS events need forcewake, so we add them to a work queue and mask their
|
|
|
|
* IMR bits until the work is done. Other interrupts can be processed without
|
|
|
|
* the work queue. */
|
|
|
|
static void gen6_rps_irq_handler(struct drm_i915_private *dev_priv, u32 pm_iir)
|
2013-05-29 02:22:24 +00:00
|
|
|
{
|
2014-03-15 14:53:22 +00:00
|
|
|
if (pm_iir & dev_priv->pm_rps_events) {
|
2013-07-04 21:35:28 +00:00
|
|
|
spin_lock(&dev_priv->irq_lock);
|
2016-10-12 16:24:30 +00:00
|
|
|
gen6_mask_pm_irq(dev_priv, pm_iir & dev_priv->pm_rps_events);
|
drm/i915: sanitize rps irq disabling
When disabling the RPS interrupts there is a tricky dependency between
the thread disabling the interrupts, the RPS interrupt handler and the
corresponding RPS work. The RPS work can reenable the interrupts, so
there is no straightforward order in the disabling thread to (1) make
sure that any RPS work is flushed and to (2) disable all RPS
interrupts. Currently this is solved by masking the interrupts using two
separate mask registers (first level display IMR and PM IMR) and doing
the disabling when all first level interrupts are disabled.
This works, but the requirement to run with all first level interrupts
disabled is unnecessary making the suspend / unload time ordering of RPS
disabling wrt. other unitialization steps difficult and error prone.
Removing this restriction allows us to disable RPS early during suspend
/ unload and forget about it for the rest of the sequence. By adding a
more explicit method for avoiding the above race, it also becomes easier
to prove its correctness. Finally currently we can hit the WARN in
snb_update_pm_irq(), when a final RPS work runs with the first level
interrupts already disabled. This won't lead to any problem (due to the
separate interrupt masks), but with the change in this and the next
patch we can get rid of the WARN, while leaving it in place for other
scenarios.
To address the above points, add a new RPS interrupts_enabled flag and
use this during RPS disabling to avoid requeuing the RPS work and
reenabling of the RPS interrupts. Since the interrupt disabling happens
now in intel_suspend_gt_powersave(), we will disable RPS interrupts
explicitly during suspend (and not just through the first level mask),
but there is no problem doing so, it's also more consistent and allows
us to unify more of the RPS disabling during suspend and unload time in
the next patch.
v2/v3:
- rebase on patch "drm/i915: move rps irq disable one level up" in the
patchset
Signed-off-by: Imre Deak <imre.deak@intel.com>
Reviewed-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2014-11-19 13:30:04 +00:00
|
|
|
if (dev_priv->rps.interrupts_enabled) {
|
|
|
|
dev_priv->rps.pm_iir |= pm_iir & dev_priv->pm_rps_events;
|
2016-07-04 07:08:36 +00:00
|
|
|
schedule_work(&dev_priv->rps.work);
|
drm/i915: sanitize rps irq disabling
When disabling the RPS interrupts there is a tricky dependency between
the thread disabling the interrupts, the RPS interrupt handler and the
corresponding RPS work. The RPS work can reenable the interrupts, so
there is no straightforward order in the disabling thread to (1) make
sure that any RPS work is flushed and to (2) disable all RPS
interrupts. Currently this is solved by masking the interrupts using two
separate mask registers (first level display IMR and PM IMR) and doing
the disabling when all first level interrupts are disabled.
This works, but the requirement to run with all first level interrupts
disabled is unnecessary making the suspend / unload time ordering of RPS
disabling wrt. other unitialization steps difficult and error prone.
Removing this restriction allows us to disable RPS early during suspend
/ unload and forget about it for the rest of the sequence. By adding a
more explicit method for avoiding the above race, it also becomes easier
to prove its correctness. Finally currently we can hit the WARN in
snb_update_pm_irq(), when a final RPS work runs with the first level
interrupts already disabled. This won't lead to any problem (due to the
separate interrupt masks), but with the change in this and the next
patch we can get rid of the WARN, while leaving it in place for other
scenarios.
To address the above points, add a new RPS interrupts_enabled flag and
use this during RPS disabling to avoid requeuing the RPS work and
reenabling of the RPS interrupts. Since the interrupt disabling happens
now in intel_suspend_gt_powersave(), we will disable RPS interrupts
explicitly during suspend (and not just through the first level mask),
but there is no problem doing so, it's also more consistent and allows
us to unify more of the RPS disabling during suspend and unload time in
the next patch.
v2/v3:
- rebase on patch "drm/i915: move rps irq disable one level up" in the
patchset
Signed-off-by: Imre Deak <imre.deak@intel.com>
Reviewed-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2014-11-19 13:30:04 +00:00
|
|
|
}
|
2013-07-04 21:35:28 +00:00
|
|
|
spin_unlock(&dev_priv->irq_lock);
|
2013-05-29 02:22:24 +00:00
|
|
|
}
|
|
|
|
|
2017-07-18 18:28:00 +00:00
|
|
|
if (INTEL_GEN(dev_priv) >= 8)
|
2014-11-05 18:48:37 +00:00
|
|
|
return;
|
|
|
|
|
2016-04-07 08:08:05 +00:00
|
|
|
if (HAS_VEBOX(dev_priv)) {
|
2013-08-15 14:51:32 +00:00
|
|
|
if (pm_iir & PM_VEBOX_USER_INTERRUPT)
|
drm/i915: Allocate intel_engine_cs structure only for the enabled engines
With the possibility of addition of many more number of rings in future,
the drm_i915_private structure could bloat as an array, of type
intel_engine_cs, is embedded inside it.
struct intel_engine_cs engine[I915_NUM_ENGINES];
Though this is still fine as generally there is only a single instance of
drm_i915_private structure used, but not all of the possible rings would be
enabled or active on most of the platforms. Some memory can be saved by
allocating intel_engine_cs structure only for the enabled/active engines.
Currently the engine/ring ID is kept static and dev_priv->engine[] is simply
indexed using the enums defined in intel_engine_id.
To save memory and continue using the static engine/ring IDs, 'engine' is
defined as an array of pointers.
struct intel_engine_cs *engine[I915_NUM_ENGINES];
dev_priv->engine[engine_ID] will be NULL for disabled engine instances.
There is a text size reduction of 928 bytes, from 1028200 to 1027272, for
i915.o file (but for i915.ko file text size remain same as 1193131 bytes).
v2:
- Remove the engine iterator field added in drm_i915_private structure,
instead pass a local iterator variable to the for_each_engine**
macros. (Chris)
- Do away with intel_engine_initialized() and instead directly use the
NULL pointer check on engine pointer. (Chris)
v3:
- Remove for_each_engine_id() macro, as the updated macro for_each_engine()
can be used in place of it. (Chris)
- Protect the access to Render engine Fault register with a NULL check, as
engine specific init is done later in Driver load sequence.
v4:
- Use !!dev_priv->engine[VCS] style for the engine check in getparam. (Chris)
- Kill the superfluous init_engine_lists().
v5:
- Cleanup the intel_engines_init() & intel_engines_setup(), with respect to
allocation of intel_engine_cs structure. (Chris)
v6:
- Rebase.
v7:
- Optimize the for_each_engine_masked() macro. (Chris)
- Change the type of 'iter' local variable to enum intel_engine_id. (Chris)
- Rebase.
v8: Rebase.
v9: Rebase.
v10:
- For index calculation use engine ID instead of pointer based arithmetic in
intel_engine_sync_index() as engine pointers are not contiguous now (Chris)
- For appropriateness, rename local enum variable 'iter' to 'id'. (Joonas)
- Use for_each_engine macro for cleanup in intel_engines_init() and remove
check for NULL engine pointer in cleanup() routines. (Joonas)
v11: Rebase.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/1476378888-7372-1-git-send-email-akash.goel@intel.com
2016-10-13 17:14:48 +00:00
|
|
|
notify_ring(dev_priv->engine[VECS]);
|
2013-05-29 02:22:31 +00:00
|
|
|
|
2014-11-04 14:52:22 +00:00
|
|
|
if (pm_iir & PM_VEBOX_CS_ERROR_INTERRUPT)
|
|
|
|
DRM_DEBUG("Command parser error, pm_iir 0x%08x\n", pm_iir);
|
2013-05-29 02:22:31 +00:00
|
|
|
}
|
2013-05-29 02:22:24 +00:00
|
|
|
}
|
|
|
|
|
2016-10-12 16:24:31 +00:00
|
|
|
static void gen9_guc_irq_handler(struct drm_i915_private *dev_priv, u32 gt_iir)
|
|
|
|
{
|
|
|
|
if (gt_iir & GEN9_GUC_TO_HOST_INT_EVENT) {
|
drm/i915: Handle log buffer flush interrupt event from GuC
GuC ukernel sends an interrupt to Host to flush the log buffer
and expects Host to correspondingly update the read pointer
information in the state structure, once it has consumed the
log buffer contents by copying them to a file or buffer.
Even if Host couldn't copy the contents, it can still update the
read pointer so that logging state is not disturbed on GuC side.
v2:
- Use a dedicated workqueue for handling flush interrupt. (Tvrtko)
- Reduce the overall log buffer copying time by skipping the copy of
crash buffer area for regular cases and copying only the state
structure data in first page.
v3:
- Create a vmalloc mapping of log buffer. (Chris)
- Cover the flush acknowledgment under rpm get & put.(Chris)
- Revert the change of skipping the copy of crash dump area, as
not really needed, will be covered by subsequent patch.
v4:
- Destroy the wq under the same condition in which it was created,
pass dev_piv pointer instead of dev to newly added GuC function,
add more comments & rename variable for clarity. (Tvrtko)
v5:
- Allocate & destroy the dedicated wq, for handling flush interrupt,
from the setup/teardown routines of GuC logging. (Chris)
- Validate the log buffer size value retrieved from state structure
and do some minor cleanup. (Tvrtko)
- Fix error/warnings reported by checkpatch. (Tvrtko)
- Rebase.
v6:
- Remove the interrupts_enabled check from guc_capture_logs_work, need
to process that last work item also, queued just before disabling the
interrupt as log buffer flush interrupt handling is a bit different
case where GuC is actually expecting an ACK from host, which should be
provided to keep the logging going.
Sync against the work will be done by caller disabling the interrupt.
- Don't sample the log buffer size value from state structure, directly
use the expected value to move the pointer & do the copy and that cannot
go wrong (out of bounds) as Driver only allocated the log buffer and the
relay buffers. Driver should refrain from interpreting the log packet,
as much possible and let Userspace parser detect the anomaly. (Chris)
v7:
- Use switch statement instead of 'if else' for retrieving the GuC log
buffer size. (Tvrtko)
- Refactored the log buffer copying function and shortended the name of
couple of variables for better readability. (Tvrtko)
v8:
- Make the dedicated wq as a high priority one to further reduce the
turnaround time of handing log buffer flush event from GuC.
Signed-off-by: Sagar Arun Kamble <sagar.a.kamble@intel.com>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
2016-10-12 16:24:32 +00:00
|
|
|
/* Sample the log buffer flush related bits & clear them out now
|
|
|
|
* itself from the message identity register to minimize the
|
|
|
|
* probability of losing a flush interrupt, when there are back
|
|
|
|
* to back flush interrupts.
|
|
|
|
* There can be a new flush interrupt, for different log buffer
|
|
|
|
* type (like for ISR), whilst Host is handling one (for DPC).
|
|
|
|
* Since same bit is used in message register for ISR & DPC, it
|
|
|
|
* could happen that GuC sets the bit for 2nd interrupt but Host
|
|
|
|
* clears out the bit on handling the 1st interrupt.
|
|
|
|
*/
|
|
|
|
u32 msg, flush;
|
|
|
|
|
|
|
|
msg = I915_READ(SOFT_SCRATCH(15));
|
2016-11-25 17:59:34 +00:00
|
|
|
flush = msg & (INTEL_GUC_RECV_MSG_CRASH_DUMP_POSTED |
|
|
|
|
INTEL_GUC_RECV_MSG_FLUSH_LOG_BUFFER);
|
drm/i915: Handle log buffer flush interrupt event from GuC
GuC ukernel sends an interrupt to Host to flush the log buffer
and expects Host to correspondingly update the read pointer
information in the state structure, once it has consumed the
log buffer contents by copying them to a file or buffer.
Even if Host couldn't copy the contents, it can still update the
read pointer so that logging state is not disturbed on GuC side.
v2:
- Use a dedicated workqueue for handling flush interrupt. (Tvrtko)
- Reduce the overall log buffer copying time by skipping the copy of
crash buffer area for regular cases and copying only the state
structure data in first page.
v3:
- Create a vmalloc mapping of log buffer. (Chris)
- Cover the flush acknowledgment under rpm get & put.(Chris)
- Revert the change of skipping the copy of crash dump area, as
not really needed, will be covered by subsequent patch.
v4:
- Destroy the wq under the same condition in which it was created,
pass dev_piv pointer instead of dev to newly added GuC function,
add more comments & rename variable for clarity. (Tvrtko)
v5:
- Allocate & destroy the dedicated wq, for handling flush interrupt,
from the setup/teardown routines of GuC logging. (Chris)
- Validate the log buffer size value retrieved from state structure
and do some minor cleanup. (Tvrtko)
- Fix error/warnings reported by checkpatch. (Tvrtko)
- Rebase.
v6:
- Remove the interrupts_enabled check from guc_capture_logs_work, need
to process that last work item also, queued just before disabling the
interrupt as log buffer flush interrupt handling is a bit different
case where GuC is actually expecting an ACK from host, which should be
provided to keep the logging going.
Sync against the work will be done by caller disabling the interrupt.
- Don't sample the log buffer size value from state structure, directly
use the expected value to move the pointer & do the copy and that cannot
go wrong (out of bounds) as Driver only allocated the log buffer and the
relay buffers. Driver should refrain from interpreting the log packet,
as much possible and let Userspace parser detect the anomaly. (Chris)
v7:
- Use switch statement instead of 'if else' for retrieving the GuC log
buffer size. (Tvrtko)
- Refactored the log buffer copying function and shortended the name of
couple of variables for better readability. (Tvrtko)
v8:
- Make the dedicated wq as a high priority one to further reduce the
turnaround time of handing log buffer flush event from GuC.
Signed-off-by: Sagar Arun Kamble <sagar.a.kamble@intel.com>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
2016-10-12 16:24:32 +00:00
|
|
|
if (flush) {
|
|
|
|
/* Clear the message bits that are handled */
|
|
|
|
I915_WRITE(SOFT_SCRATCH(15), msg & ~flush);
|
|
|
|
|
|
|
|
/* Handle flush interrupt in bottom half */
|
2017-03-22 17:39:48 +00:00
|
|
|
queue_work(dev_priv->guc.log.runtime.flush_wq,
|
|
|
|
&dev_priv->guc.log.runtime.flush_work);
|
2016-10-12 16:24:36 +00:00
|
|
|
|
|
|
|
dev_priv->guc.log.flush_interrupt_count++;
|
drm/i915: Handle log buffer flush interrupt event from GuC
GuC ukernel sends an interrupt to Host to flush the log buffer
and expects Host to correspondingly update the read pointer
information in the state structure, once it has consumed the
log buffer contents by copying them to a file or buffer.
Even if Host couldn't copy the contents, it can still update the
read pointer so that logging state is not disturbed on GuC side.
v2:
- Use a dedicated workqueue for handling flush interrupt. (Tvrtko)
- Reduce the overall log buffer copying time by skipping the copy of
crash buffer area for regular cases and copying only the state
structure data in first page.
v3:
- Create a vmalloc mapping of log buffer. (Chris)
- Cover the flush acknowledgment under rpm get & put.(Chris)
- Revert the change of skipping the copy of crash dump area, as
not really needed, will be covered by subsequent patch.
v4:
- Destroy the wq under the same condition in which it was created,
pass dev_piv pointer instead of dev to newly added GuC function,
add more comments & rename variable for clarity. (Tvrtko)
v5:
- Allocate & destroy the dedicated wq, for handling flush interrupt,
from the setup/teardown routines of GuC logging. (Chris)
- Validate the log buffer size value retrieved from state structure
and do some minor cleanup. (Tvrtko)
- Fix error/warnings reported by checkpatch. (Tvrtko)
- Rebase.
v6:
- Remove the interrupts_enabled check from guc_capture_logs_work, need
to process that last work item also, queued just before disabling the
interrupt as log buffer flush interrupt handling is a bit different
case where GuC is actually expecting an ACK from host, which should be
provided to keep the logging going.
Sync against the work will be done by caller disabling the interrupt.
- Don't sample the log buffer size value from state structure, directly
use the expected value to move the pointer & do the copy and that cannot
go wrong (out of bounds) as Driver only allocated the log buffer and the
relay buffers. Driver should refrain from interpreting the log packet,
as much possible and let Userspace parser detect the anomaly. (Chris)
v7:
- Use switch statement instead of 'if else' for retrieving the GuC log
buffer size. (Tvrtko)
- Refactored the log buffer copying function and shortended the name of
couple of variables for better readability. (Tvrtko)
v8:
- Make the dedicated wq as a high priority one to further reduce the
turnaround time of handing log buffer flush event from GuC.
Signed-off-by: Sagar Arun Kamble <sagar.a.kamble@intel.com>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
2016-10-12 16:24:32 +00:00
|
|
|
} else {
|
|
|
|
/* Not clearing of unhandled event bits won't result in
|
|
|
|
* re-triggering of the interrupt.
|
|
|
|
*/
|
|
|
|
}
|
2016-10-12 16:24:31 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-08-18 18:36:51 +00:00
|
|
|
static void i9xx_pipestat_irq_reset(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
|
|
|
enum pipe pipe;
|
|
|
|
|
|
|
|
for_each_pipe(dev_priv, pipe) {
|
|
|
|
I915_WRITE(PIPESTAT(pipe),
|
|
|
|
PIPESTAT_INT_STATUS_MASK |
|
|
|
|
PIPE_FIFO_UNDERRUN_STATUS);
|
|
|
|
|
|
|
|
dev_priv->pipestat_irq_mask[pipe] = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-08-18 18:36:59 +00:00
|
|
|
static void i9xx_pipestat_irq_ack(struct drm_i915_private *dev_priv,
|
|
|
|
u32 iir, u32 pipe_stats[I915_MAX_PIPES])
|
2014-02-04 19:35:46 +00:00
|
|
|
{
|
|
|
|
int pipe;
|
|
|
|
|
2014-02-04 19:35:47 +00:00
|
|
|
spin_lock(&dev_priv->irq_lock);
|
2016-02-18 19:54:26 +00:00
|
|
|
|
|
|
|
if (!dev_priv->display_irqs_enabled) {
|
|
|
|
spin_unlock(&dev_priv->irq_lock);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2014-08-18 12:49:10 +00:00
|
|
|
for_each_pipe(dev_priv, pipe) {
|
drm/i915: Type safe register read/write
Make I915_READ and I915_WRITE more type safe by wrapping the register
offset in a struct. This should eliminate most of the fumbles we've had
with misplaced parens.
This only takes care of normal mmio registers. We could extend the idea
to other register types and define each with its own struct. That way
you wouldn't be able to accidentally pass the wrong thing to a specific
register access function.
The gpio_reg setup is probably the ugliest thing left. But I figure I'd
just leave it for now, and wait for some divine inspiration to strike
before making it nice.
As for the generated code, it's actually a bit better sometimes. Eg.
looking at i915_irq_handler(), we can see the following change:
lea 0x70024(%rdx,%rax,1),%r9d
mov $0x1,%edx
- movslq %r9d,%r9
- mov %r9,%rsi
- mov %r9,-0x58(%rbp)
- callq *0xd8(%rbx)
+ mov %r9d,%esi
+ mov %r9d,-0x48(%rbp)
callq *0xd8(%rbx)
So previously gcc thought the register offset might be signed and
decided to sign extend it, just in case. The rest appears to be
mostly just minor shuffling of instructions.
v2: i915_mmio_reg_{offset,equal,valid}() helpers added
s/_REG/_MMIO/ in the register defines
mo more switch statements left to worry about
ring_emit stuff got sorted in a prep patch
cmd parser, lrc context and w/a batch buildup also in prep patch
vgpu stuff cleaned up and moved to a prep patch
all other unrelated changes split out
v3: Rebased due to BXT DSI/BLC, MOCS, etc.
v4: Rebased due to churn, s/i915_mmio_reg_t/i915_reg_t/
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/1447853606-2751-1-git-send-email-ville.syrjala@linux.intel.com
2015-11-18 13:33:26 +00:00
|
|
|
i915_reg_t reg;
|
2017-09-14 15:17:31 +00:00
|
|
|
u32 status_mask, enable_mask, iir_bit = 0;
|
2014-02-10 16:42:49 +00:00
|
|
|
|
2014-02-12 16:55:36 +00:00
|
|
|
/*
|
|
|
|
* PIPESTAT bits get signalled even when the interrupt is
|
|
|
|
* disabled with the mask bits, and some of the status bits do
|
|
|
|
* not generate interrupts at all (like the underrun bit). Hence
|
|
|
|
* we need to be careful that we only handle what we want to
|
|
|
|
* handle.
|
|
|
|
*/
|
2014-09-30 08:56:49 +00:00
|
|
|
|
|
|
|
/* fifo underruns are filterered in the underrun handler. */
|
2017-09-14 15:17:31 +00:00
|
|
|
status_mask = PIPE_FIFO_UNDERRUN_STATUS;
|
2014-02-12 16:55:36 +00:00
|
|
|
|
|
|
|
switch (pipe) {
|
|
|
|
case PIPE_A:
|
|
|
|
iir_bit = I915_DISPLAY_PIPE_A_EVENT_INTERRUPT;
|
|
|
|
break;
|
|
|
|
case PIPE_B:
|
|
|
|
iir_bit = I915_DISPLAY_PIPE_B_EVENT_INTERRUPT;
|
|
|
|
break;
|
2014-04-09 10:28:49 +00:00
|
|
|
case PIPE_C:
|
|
|
|
iir_bit = I915_DISPLAY_PIPE_C_EVENT_INTERRUPT;
|
|
|
|
break;
|
2014-02-12 16:55:36 +00:00
|
|
|
}
|
|
|
|
if (iir & iir_bit)
|
2017-09-14 15:17:31 +00:00
|
|
|
status_mask |= dev_priv->pipestat_irq_mask[pipe];
|
2014-02-12 16:55:36 +00:00
|
|
|
|
2017-09-14 15:17:31 +00:00
|
|
|
if (!status_mask)
|
2014-02-10 16:42:49 +00:00
|
|
|
continue;
|
|
|
|
|
|
|
|
reg = PIPESTAT(pipe);
|
2017-09-14 15:17:31 +00:00
|
|
|
pipe_stats[pipe] = I915_READ(reg) & status_mask;
|
|
|
|
enable_mask = i915_pipestat_enable_mask(dev_priv, pipe);
|
2014-02-04 19:35:46 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Clear the PIPE*STAT regs before the IIR
|
|
|
|
*/
|
2017-09-14 15:17:31 +00:00
|
|
|
if (pipe_stats[pipe])
|
|
|
|
I915_WRITE(reg, enable_mask | pipe_stats[pipe]);
|
2014-02-04 19:35:46 +00:00
|
|
|
}
|
2014-02-04 19:35:47 +00:00
|
|
|
spin_unlock(&dev_priv->irq_lock);
|
2016-04-13 18:19:55 +00:00
|
|
|
}
|
|
|
|
|
2017-08-18 18:36:59 +00:00
|
|
|
static void i8xx_pipestat_irq_handler(struct drm_i915_private *dev_priv,
|
|
|
|
u16 iir, u32 pipe_stats[I915_MAX_PIPES])
|
|
|
|
{
|
|
|
|
enum pipe pipe;
|
|
|
|
|
|
|
|
for_each_pipe(dev_priv, pipe) {
|
|
|
|
if (pipe_stats[pipe] & PIPE_VBLANK_INTERRUPT_STATUS)
|
|
|
|
drm_handle_vblank(&dev_priv->drm, pipe);
|
|
|
|
|
|
|
|
if (pipe_stats[pipe] & PIPE_CRC_DONE_INTERRUPT_STATUS)
|
|
|
|
i9xx_pipe_crc_irq_handler(dev_priv, pipe);
|
|
|
|
|
|
|
|
if (pipe_stats[pipe] & PIPE_FIFO_UNDERRUN_STATUS)
|
|
|
|
intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void i915_pipestat_irq_handler(struct drm_i915_private *dev_priv,
|
|
|
|
u32 iir, u32 pipe_stats[I915_MAX_PIPES])
|
|
|
|
{
|
|
|
|
bool blc_event = false;
|
|
|
|
enum pipe pipe;
|
|
|
|
|
|
|
|
for_each_pipe(dev_priv, pipe) {
|
|
|
|
if (pipe_stats[pipe] & PIPE_VBLANK_INTERRUPT_STATUS)
|
|
|
|
drm_handle_vblank(&dev_priv->drm, pipe);
|
|
|
|
|
|
|
|
if (pipe_stats[pipe] & PIPE_LEGACY_BLC_EVENT_STATUS)
|
|
|
|
blc_event = true;
|
|
|
|
|
|
|
|
if (pipe_stats[pipe] & PIPE_CRC_DONE_INTERRUPT_STATUS)
|
|
|
|
i9xx_pipe_crc_irq_handler(dev_priv, pipe);
|
|
|
|
|
|
|
|
if (pipe_stats[pipe] & PIPE_FIFO_UNDERRUN_STATUS)
|
|
|
|
intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (blc_event || (iir & I915_ASLE_INTERRUPT))
|
|
|
|
intel_opregion_asle_intr(dev_priv);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void i965_pipestat_irq_handler(struct drm_i915_private *dev_priv,
|
|
|
|
u32 iir, u32 pipe_stats[I915_MAX_PIPES])
|
|
|
|
{
|
|
|
|
bool blc_event = false;
|
|
|
|
enum pipe pipe;
|
|
|
|
|
|
|
|
for_each_pipe(dev_priv, pipe) {
|
|
|
|
if (pipe_stats[pipe] & PIPE_START_VBLANK_INTERRUPT_STATUS)
|
|
|
|
drm_handle_vblank(&dev_priv->drm, pipe);
|
|
|
|
|
|
|
|
if (pipe_stats[pipe] & PIPE_LEGACY_BLC_EVENT_STATUS)
|
|
|
|
blc_event = true;
|
|
|
|
|
|
|
|
if (pipe_stats[pipe] & PIPE_CRC_DONE_INTERRUPT_STATUS)
|
|
|
|
i9xx_pipe_crc_irq_handler(dev_priv, pipe);
|
|
|
|
|
|
|
|
if (pipe_stats[pipe] & PIPE_FIFO_UNDERRUN_STATUS)
|
|
|
|
intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (blc_event || (iir & I915_ASLE_INTERRUPT))
|
|
|
|
intel_opregion_asle_intr(dev_priv);
|
|
|
|
|
|
|
|
if (pipe_stats[0] & PIPE_GMBUS_INTERRUPT_STATUS)
|
|
|
|
gmbus_irq_handler(dev_priv);
|
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void valleyview_pipestat_irq_handler(struct drm_i915_private *dev_priv,
|
2016-04-13 18:19:55 +00:00
|
|
|
u32 pipe_stats[I915_MAX_PIPES])
|
|
|
|
{
|
|
|
|
enum pipe pipe;
|
2014-02-04 19:35:46 +00:00
|
|
|
|
2014-08-18 12:49:10 +00:00
|
|
|
for_each_pipe(dev_priv, pipe) {
|
2017-07-20 17:57:51 +00:00
|
|
|
if (pipe_stats[pipe] & PIPE_START_VBLANK_INTERRUPT_STATUS)
|
|
|
|
drm_handle_vblank(&dev_priv->drm, pipe);
|
2014-02-04 19:35:46 +00:00
|
|
|
|
|
|
|
if (pipe_stats[pipe] & PIPE_CRC_DONE_INTERRUPT_STATUS)
|
2016-05-06 13:48:28 +00:00
|
|
|
i9xx_pipe_crc_irq_handler(dev_priv, pipe);
|
2014-02-04 19:35:46 +00:00
|
|
|
|
2014-09-30 08:56:48 +00:00
|
|
|
if (pipe_stats[pipe] & PIPE_FIFO_UNDERRUN_STATUS)
|
|
|
|
intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
|
2014-02-04 19:35:46 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if (pipe_stats[0] & PIPE_GMBUS_INTERRUPT_STATUS)
|
2016-05-06 13:48:28 +00:00
|
|
|
gmbus_irq_handler(dev_priv);
|
2014-02-04 19:35:46 +00:00
|
|
|
}
|
|
|
|
|
2016-04-13 18:19:54 +00:00
|
|
|
static u32 i9xx_hpd_irq_ack(struct drm_i915_private *dev_priv)
|
2014-04-01 07:54:36 +00:00
|
|
|
{
|
|
|
|
u32 hotplug_status = I915_READ(PORT_HOTPLUG_STAT);
|
|
|
|
|
2016-04-13 18:19:54 +00:00
|
|
|
if (hotplug_status)
|
|
|
|
I915_WRITE(PORT_HOTPLUG_STAT, hotplug_status);
|
2014-04-01 07:54:36 +00:00
|
|
|
|
2016-04-13 18:19:54 +00:00
|
|
|
return hotplug_status;
|
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void i9xx_hpd_irq_handler(struct drm_i915_private *dev_priv,
|
2016-04-13 18:19:54 +00:00
|
|
|
u32 hotplug_status)
|
|
|
|
{
|
|
|
|
u32 pin_mask = 0, long_mask = 0;
|
2014-04-01 07:54:36 +00:00
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
if (IS_G4X(dev_priv) || IS_VALLEYVIEW(dev_priv) ||
|
|
|
|
IS_CHERRYVIEW(dev_priv)) {
|
2015-05-27 12:03:39 +00:00
|
|
|
u32 hotplug_trigger = hotplug_status & HOTPLUG_INT_STATUS_G4X;
|
2014-04-01 07:54:36 +00:00
|
|
|
|
drm/i915: Don't call intel_get_hpd_pins() when there's no hotplug interrupt
On GMCH plaforms we are now getting the following spew on aux
interrupts:
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_dp_aux_ch] dp_aux_ch timeout status 0x71450064
Prevent it by not calling intel_get_hpd_pins() unless one of the HPD
interrupt bits are actually set.
I already fixed similar annoyance once with
4bca26d0a6518d51a9abe64fbde4b12f04c74053 drm/i915: Use HOTPLUG_INT_STATUS_G4X on VLV/CHV
but another source for it got added in
fd63e2a972c670887e5e8a08440111d3812c0996 drm/i915: combine i9xx_get_hpd_pins and pch_get_hpd_pins
due to pch_get_hpd_pins() being chosen over i9xx_get_hpd_pins() to
serve as the new unified piece of code. pch_get_hpd_pins() had the debug
print, and i9xx_get_hpd_pins() didn't.
Cc: Imre Deak <imre.deak@intel.com>
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2015-08-28 19:59:08 +00:00
|
|
|
if (hotplug_trigger) {
|
|
|
|
intel_get_hpd_pins(&pin_mask, &long_mask, hotplug_trigger,
|
|
|
|
hotplug_trigger, hpd_status_g4x,
|
|
|
|
i9xx_port_hotplug_long_detect);
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
|
drm/i915: Don't call intel_get_hpd_pins() when there's no hotplug interrupt
On GMCH plaforms we are now getting the following spew on aux
interrupts:
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_dp_aux_ch] dp_aux_ch timeout status 0x71450064
Prevent it by not calling intel_get_hpd_pins() unless one of the HPD
interrupt bits are actually set.
I already fixed similar annoyance once with
4bca26d0a6518d51a9abe64fbde4b12f04c74053 drm/i915: Use HOTPLUG_INT_STATUS_G4X on VLV/CHV
but another source for it got added in
fd63e2a972c670887e5e8a08440111d3812c0996 drm/i915: combine i9xx_get_hpd_pins and pch_get_hpd_pins
due to pch_get_hpd_pins() being chosen over i9xx_get_hpd_pins() to
serve as the new unified piece of code. pch_get_hpd_pins() had the debug
print, and i9xx_get_hpd_pins() didn't.
Cc: Imre Deak <imre.deak@intel.com>
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2015-08-28 19:59:08 +00:00
|
|
|
}
|
2015-05-27 12:03:40 +00:00
|
|
|
|
|
|
|
if (hotplug_status & DP_AUX_CHANNEL_MASK_INT_STATUS_G4X)
|
2016-05-06 13:48:28 +00:00
|
|
|
dp_aux_irq_handler(dev_priv);
|
2015-05-27 12:03:39 +00:00
|
|
|
} else {
|
|
|
|
u32 hotplug_trigger = hotplug_status & HOTPLUG_INT_STATUS_I915;
|
2014-04-01 07:54:36 +00:00
|
|
|
|
drm/i915: Don't call intel_get_hpd_pins() when there's no hotplug interrupt
On GMCH plaforms we are now getting the following spew on aux
interrupts:
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_dp_aux_ch] dp_aux_ch timeout status 0x71450064
Prevent it by not calling intel_get_hpd_pins() unless one of the HPD
interrupt bits are actually set.
I already fixed similar annoyance once with
4bca26d0a6518d51a9abe64fbde4b12f04c74053 drm/i915: Use HOTPLUG_INT_STATUS_G4X on VLV/CHV
but another source for it got added in
fd63e2a972c670887e5e8a08440111d3812c0996 drm/i915: combine i9xx_get_hpd_pins and pch_get_hpd_pins
due to pch_get_hpd_pins() being chosen over i9xx_get_hpd_pins() to
serve as the new unified piece of code. pch_get_hpd_pins() had the debug
print, and i9xx_get_hpd_pins() didn't.
Cc: Imre Deak <imre.deak@intel.com>
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2015-08-28 19:59:08 +00:00
|
|
|
if (hotplug_trigger) {
|
|
|
|
intel_get_hpd_pins(&pin_mask, &long_mask, hotplug_trigger,
|
2015-09-30 06:47:41 +00:00
|
|
|
hotplug_trigger, hpd_status_i915,
|
drm/i915: Don't call intel_get_hpd_pins() when there's no hotplug interrupt
On GMCH plaforms we are now getting the following spew on aux
interrupts:
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_dp_aux_ch] dp_aux_ch timeout status 0x71450064
Prevent it by not calling intel_get_hpd_pins() unless one of the HPD
interrupt bits are actually set.
I already fixed similar annoyance once with
4bca26d0a6518d51a9abe64fbde4b12f04c74053 drm/i915: Use HOTPLUG_INT_STATUS_G4X on VLV/CHV
but another source for it got added in
fd63e2a972c670887e5e8a08440111d3812c0996 drm/i915: combine i9xx_get_hpd_pins and pch_get_hpd_pins
due to pch_get_hpd_pins() being chosen over i9xx_get_hpd_pins() to
serve as the new unified piece of code. pch_get_hpd_pins() had the debug
print, and i9xx_get_hpd_pins() didn't.
Cc: Imre Deak <imre.deak@intel.com>
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2015-08-28 19:59:08 +00:00
|
|
|
i9xx_port_hotplug_long_detect);
|
2016-05-06 13:48:28 +00:00
|
|
|
intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
|
drm/i915: Don't call intel_get_hpd_pins() when there's no hotplug interrupt
On GMCH plaforms we are now getting the following spew on aux
interrupts:
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_get_hpd_pins] hotplug event received, stat 0x00000000, dig 0x00000000, pins 0x00000000
[drm:intel_dp_aux_ch] dp_aux_ch timeout status 0x71450064
Prevent it by not calling intel_get_hpd_pins() unless one of the HPD
interrupt bits are actually set.
I already fixed similar annoyance once with
4bca26d0a6518d51a9abe64fbde4b12f04c74053 drm/i915: Use HOTPLUG_INT_STATUS_G4X on VLV/CHV
but another source for it got added in
fd63e2a972c670887e5e8a08440111d3812c0996 drm/i915: combine i9xx_get_hpd_pins and pch_get_hpd_pins
due to pch_get_hpd_pins() being chosen over i9xx_get_hpd_pins() to
serve as the new unified piece of code. pch_get_hpd_pins() had the debug
print, and i9xx_get_hpd_pins() didn't.
Cc: Imre Deak <imre.deak@intel.com>
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2015-08-28 19:59:08 +00:00
|
|
|
}
|
2014-06-16 15:10:58 +00:00
|
|
|
}
|
2014-04-01 07:54:36 +00:00
|
|
|
}
|
|
|
|
|
2012-10-02 13:10:55 +00:00
|
|
|
static irqreturn_t valleyview_irq_handler(int irq, void *arg)
|
2012-03-28 20:39:38 +00:00
|
|
|
{
|
2014-05-12 17:17:55 +00:00
|
|
|
struct drm_device *dev = arg;
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2012-03-28 20:39:38 +00:00
|
|
|
irqreturn_t ret = IRQ_NONE;
|
|
|
|
|
2015-02-24 09:14:30 +00:00
|
|
|
if (!intel_irqs_enabled(dev_priv))
|
|
|
|
return IRQ_NONE;
|
|
|
|
|
2015-12-16 00:52:19 +00:00
|
|
|
/* IRQs are synced during runtime_suspend, we don't require a wakeref */
|
|
|
|
disable_rpm_wakeref_asserts(dev_priv);
|
|
|
|
|
2016-04-13 18:19:52 +00:00
|
|
|
do {
|
2016-04-13 18:19:53 +00:00
|
|
|
u32 iir, gt_iir, pm_iir;
|
2016-04-13 18:19:55 +00:00
|
|
|
u32 pipe_stats[I915_MAX_PIPES] = {};
|
2016-04-13 18:19:54 +00:00
|
|
|
u32 hotplug_status = 0;
|
2016-04-13 18:19:51 +00:00
|
|
|
u32 ier = 0;
|
2014-06-16 15:10:58 +00:00
|
|
|
|
2012-03-28 20:39:38 +00:00
|
|
|
gt_iir = I915_READ(GTIIR);
|
|
|
|
pm_iir = I915_READ(GEN6_PMIIR);
|
2014-06-16 15:10:58 +00:00
|
|
|
iir = I915_READ(VLV_IIR);
|
2012-03-28 20:39:38 +00:00
|
|
|
|
|
|
|
if (gt_iir == 0 && pm_iir == 0 && iir == 0)
|
2016-04-13 18:19:52 +00:00
|
|
|
break;
|
2012-03-28 20:39:38 +00:00
|
|
|
|
|
|
|
ret = IRQ_HANDLED;
|
|
|
|
|
2016-04-13 18:19:51 +00:00
|
|
|
/*
|
|
|
|
* Theory on interrupt generation, based on empirical evidence:
|
|
|
|
*
|
|
|
|
* x = ((VLV_IIR & VLV_IER) ||
|
|
|
|
* (((GT_IIR & GT_IER) || (GEN6_PMIIR & GEN6_PMIER)) &&
|
|
|
|
* (VLV_MASTER_IER & MASTER_INTERRUPT_ENABLE)));
|
|
|
|
*
|
|
|
|
* A CPU interrupt will only be raised when 'x' has a 0->1 edge.
|
|
|
|
* Hence we clear MASTER_INTERRUPT_ENABLE and VLV_IER to
|
|
|
|
* guarantee the CPU interrupt will be raised again even if we
|
|
|
|
* don't end up clearing all the VLV_IIR, GT_IIR, GEN6_PMIIR
|
|
|
|
* bits this time around.
|
|
|
|
*/
|
2016-04-13 18:19:50 +00:00
|
|
|
I915_WRITE(VLV_MASTER_IER, 0);
|
2016-04-13 18:19:51 +00:00
|
|
|
ier = I915_READ(VLV_IER);
|
|
|
|
I915_WRITE(VLV_IER, 0);
|
2016-04-13 18:19:50 +00:00
|
|
|
|
|
|
|
if (gt_iir)
|
|
|
|
I915_WRITE(GTIIR, gt_iir);
|
|
|
|
if (pm_iir)
|
|
|
|
I915_WRITE(GEN6_PMIIR, pm_iir);
|
|
|
|
|
2016-04-13 18:19:49 +00:00
|
|
|
if (iir & I915_DISPLAY_PORT_INTERRUPT)
|
2016-04-13 18:19:54 +00:00
|
|
|
hotplug_status = i9xx_hpd_irq_ack(dev_priv);
|
2016-04-13 18:19:49 +00:00
|
|
|
|
2014-06-16 15:10:58 +00:00
|
|
|
/* Call regardless, as some status bits might not be
|
|
|
|
* signalled in iir */
|
2017-08-18 18:36:59 +00:00
|
|
|
i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats);
|
2016-04-13 18:19:49 +00:00
|
|
|
|
2017-01-24 22:57:49 +00:00
|
|
|
if (iir & (I915_LPE_PIPE_A_INTERRUPT |
|
|
|
|
I915_LPE_PIPE_B_INTERRUPT))
|
|
|
|
intel_lpe_audio_irq_handler(dev_priv);
|
|
|
|
|
2016-04-13 18:19:49 +00:00
|
|
|
/*
|
|
|
|
* VLV_IIR is single buffered, and reflects the level
|
|
|
|
* from PIPESTAT/PORT_HOTPLUG_STAT, hence clear it last.
|
|
|
|
*/
|
|
|
|
if (iir)
|
|
|
|
I915_WRITE(VLV_IIR, iir);
|
2016-04-13 18:19:50 +00:00
|
|
|
|
2016-04-13 18:19:51 +00:00
|
|
|
I915_WRITE(VLV_IER, ier);
|
2016-04-13 18:19:50 +00:00
|
|
|
I915_WRITE(VLV_MASTER_IER, MASTER_INTERRUPT_ENABLE);
|
|
|
|
POSTING_READ(VLV_MASTER_IER);
|
2016-04-13 18:19:54 +00:00
|
|
|
|
2016-04-13 18:19:56 +00:00
|
|
|
if (gt_iir)
|
2016-04-13 18:19:57 +00:00
|
|
|
snb_gt_irq_handler(dev_priv, gt_iir);
|
2016-04-13 18:19:56 +00:00
|
|
|
if (pm_iir)
|
|
|
|
gen6_rps_irq_handler(dev_priv, pm_iir);
|
|
|
|
|
2016-04-13 18:19:54 +00:00
|
|
|
if (hotplug_status)
|
2016-05-06 13:48:28 +00:00
|
|
|
i9xx_hpd_irq_handler(dev_priv, hotplug_status);
|
2016-04-13 18:19:55 +00:00
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
valleyview_pipestat_irq_handler(dev_priv, pipe_stats);
|
2016-04-13 18:19:52 +00:00
|
|
|
} while (0);
|
2012-03-28 20:39:38 +00:00
|
|
|
|
2015-12-16 00:52:19 +00:00
|
|
|
enable_rpm_wakeref_asserts(dev_priv);
|
|
|
|
|
2012-03-28 20:39:38 +00:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2014-04-09 17:40:52 +00:00
|
|
|
static irqreturn_t cherryview_irq_handler(int irq, void *arg)
|
|
|
|
{
|
2014-05-12 17:17:55 +00:00
|
|
|
struct drm_device *dev = arg;
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2014-04-09 17:40:52 +00:00
|
|
|
irqreturn_t ret = IRQ_NONE;
|
|
|
|
|
2015-02-24 09:14:30 +00:00
|
|
|
if (!intel_irqs_enabled(dev_priv))
|
|
|
|
return IRQ_NONE;
|
|
|
|
|
2015-12-16 00:52:19 +00:00
|
|
|
/* IRQs are synced during runtime_suspend, we don't require a wakeref */
|
|
|
|
disable_rpm_wakeref_asserts(dev_priv);
|
|
|
|
|
2016-03-14 09:01:57 +00:00
|
|
|
do {
|
2016-04-13 18:19:53 +00:00
|
|
|
u32 master_ctl, iir;
|
2016-04-13 18:19:58 +00:00
|
|
|
u32 gt_iir[4] = {};
|
2016-04-13 18:19:55 +00:00
|
|
|
u32 pipe_stats[I915_MAX_PIPES] = {};
|
2016-04-13 18:19:54 +00:00
|
|
|
u32 hotplug_status = 0;
|
2016-04-13 18:19:51 +00:00
|
|
|
u32 ier = 0;
|
|
|
|
|
2014-04-09 10:28:50 +00:00
|
|
|
master_ctl = I915_READ(GEN8_MASTER_IRQ) & ~GEN8_MASTER_IRQ_CONTROL;
|
|
|
|
iir = I915_READ(VLV_IIR);
|
2014-04-09 17:40:52 +00:00
|
|
|
|
2014-04-09 10:28:50 +00:00
|
|
|
if (master_ctl == 0 && iir == 0)
|
|
|
|
break;
|
2014-04-09 17:40:52 +00:00
|
|
|
|
2014-06-16 15:11:00 +00:00
|
|
|
ret = IRQ_HANDLED;
|
|
|
|
|
2016-04-13 18:19:51 +00:00
|
|
|
/*
|
|
|
|
* Theory on interrupt generation, based on empirical evidence:
|
|
|
|
*
|
|
|
|
* x = ((VLV_IIR & VLV_IER) ||
|
|
|
|
* ((GEN8_MASTER_IRQ & ~GEN8_MASTER_IRQ_CONTROL) &&
|
|
|
|
* (GEN8_MASTER_IRQ & GEN8_MASTER_IRQ_CONTROL)));
|
|
|
|
*
|
|
|
|
* A CPU interrupt will only be raised when 'x' has a 0->1 edge.
|
|
|
|
* Hence we clear GEN8_MASTER_IRQ_CONTROL and VLV_IER to
|
|
|
|
* guarantee the CPU interrupt will be raised again even if we
|
|
|
|
* don't end up clearing all the VLV_IIR and GEN8_MASTER_IRQ_CONTROL
|
|
|
|
* bits this time around.
|
|
|
|
*/
|
2014-04-09 10:28:50 +00:00
|
|
|
I915_WRITE(GEN8_MASTER_IRQ, 0);
|
2016-04-13 18:19:51 +00:00
|
|
|
ier = I915_READ(VLV_IER);
|
|
|
|
I915_WRITE(VLV_IER, 0);
|
2014-04-09 17:40:52 +00:00
|
|
|
|
2016-04-13 18:19:58 +00:00
|
|
|
gen8_gt_irq_ack(dev_priv, master_ctl, gt_iir);
|
2014-04-09 17:40:52 +00:00
|
|
|
|
2016-04-13 18:19:49 +00:00
|
|
|
if (iir & I915_DISPLAY_PORT_INTERRUPT)
|
2016-04-13 18:19:54 +00:00
|
|
|
hotplug_status = i9xx_hpd_irq_ack(dev_priv);
|
2016-04-13 18:19:49 +00:00
|
|
|
|
2014-06-16 15:11:00 +00:00
|
|
|
/* Call regardless, as some status bits might not be
|
|
|
|
* signalled in iir */
|
2017-08-18 18:36:59 +00:00
|
|
|
i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats);
|
2014-04-09 17:40:52 +00:00
|
|
|
|
2017-01-24 22:57:49 +00:00
|
|
|
if (iir & (I915_LPE_PIPE_A_INTERRUPT |
|
|
|
|
I915_LPE_PIPE_B_INTERRUPT |
|
|
|
|
I915_LPE_PIPE_C_INTERRUPT))
|
|
|
|
intel_lpe_audio_irq_handler(dev_priv);
|
|
|
|
|
2016-04-13 18:19:49 +00:00
|
|
|
/*
|
|
|
|
* VLV_IIR is single buffered, and reflects the level
|
|
|
|
* from PIPESTAT/PORT_HOTPLUG_STAT, hence clear it last.
|
|
|
|
*/
|
|
|
|
if (iir)
|
|
|
|
I915_WRITE(VLV_IIR, iir);
|
|
|
|
|
2016-04-13 18:19:51 +00:00
|
|
|
I915_WRITE(VLV_IER, ier);
|
2016-04-13 18:19:47 +00:00
|
|
|
I915_WRITE(GEN8_MASTER_IRQ, GEN8_MASTER_IRQ_CONTROL);
|
2014-04-09 10:28:50 +00:00
|
|
|
POSTING_READ(GEN8_MASTER_IRQ);
|
2016-04-13 18:19:54 +00:00
|
|
|
|
2016-04-13 18:19:58 +00:00
|
|
|
gen8_gt_irq_handler(dev_priv, gt_iir);
|
|
|
|
|
2016-04-13 18:19:54 +00:00
|
|
|
if (hotplug_status)
|
2016-05-06 13:48:28 +00:00
|
|
|
i9xx_hpd_irq_handler(dev_priv, hotplug_status);
|
2016-04-13 18:19:55 +00:00
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
valleyview_pipestat_irq_handler(dev_priv, pipe_stats);
|
2016-03-14 09:01:57 +00:00
|
|
|
} while (0);
|
2014-04-09 10:28:49 +00:00
|
|
|
|
2015-12-16 00:52:19 +00:00
|
|
|
enable_rpm_wakeref_asserts(dev_priv);
|
|
|
|
|
2014-04-09 17:40:52 +00:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void ibx_hpd_irq_handler(struct drm_i915_private *dev_priv,
|
|
|
|
u32 hotplug_trigger,
|
2015-08-27 20:56:10 +00:00
|
|
|
const u32 hpd[HPD_NUM_PINS])
|
|
|
|
{
|
|
|
|
u32 dig_hotplug_reg, pin_mask = 0, long_mask = 0;
|
|
|
|
|
2015-11-25 14:47:22 +00:00
|
|
|
/*
|
|
|
|
* Somehow the PCH doesn't seem to really ack the interrupt to the CPU
|
|
|
|
* unless we touch the hotplug register, even if hotplug_trigger is
|
|
|
|
* zero. Not acking leads to "The master control interrupt lied (SDE)!"
|
|
|
|
* errors.
|
|
|
|
*/
|
2015-08-27 20:56:10 +00:00
|
|
|
dig_hotplug_reg = I915_READ(PCH_PORT_HOTPLUG);
|
2015-11-25 14:47:22 +00:00
|
|
|
if (!hotplug_trigger) {
|
|
|
|
u32 mask = PORTA_HOTPLUG_STATUS_MASK |
|
|
|
|
PORTD_HOTPLUG_STATUS_MASK |
|
|
|
|
PORTC_HOTPLUG_STATUS_MASK |
|
|
|
|
PORTB_HOTPLUG_STATUS_MASK;
|
|
|
|
dig_hotplug_reg &= ~mask;
|
|
|
|
}
|
|
|
|
|
2015-08-27 20:56:10 +00:00
|
|
|
I915_WRITE(PCH_PORT_HOTPLUG, dig_hotplug_reg);
|
2015-11-25 14:47:22 +00:00
|
|
|
if (!hotplug_trigger)
|
|
|
|
return;
|
2015-08-27 20:56:10 +00:00
|
|
|
|
|
|
|
intel_get_hpd_pins(&pin_mask, &long_mask, hotplug_trigger,
|
|
|
|
dig_hotplug_reg, hpd,
|
|
|
|
pch_port_hotplug_long_detect);
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
|
2015-08-27 20:56:10 +00:00
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void ibx_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
|
2011-01-04 23:09:39 +00:00
|
|
|
{
|
2011-02-07 20:26:52 +00:00
|
|
|
int pipe;
|
2013-04-16 11:36:54 +00:00
|
|
|
u32 hotplug_trigger = pch_iir & SDE_HOTPLUG_MASK;
|
2014-06-18 01:29:35 +00:00
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
ibx_hpd_irq_handler(dev_priv, hotplug_trigger, hpd_ibx);
|
2013-06-27 15:52:14 +00:00
|
|
|
|
2013-04-17 14:48:48 +00:00
|
|
|
if (pch_iir & SDE_AUDIO_POWER_MASK) {
|
|
|
|
int port = ffs((pch_iir & SDE_AUDIO_POWER_MASK) >>
|
|
|
|
SDE_AUDIO_POWER_SHIFT);
|
2011-01-04 23:09:39 +00:00
|
|
|
DRM_DEBUG_DRIVER("PCH audio power change on port %d\n",
|
2013-04-17 14:48:48 +00:00
|
|
|
port_name(port));
|
|
|
|
}
|
2011-01-04 23:09:39 +00:00
|
|
|
|
2012-12-01 12:53:47 +00:00
|
|
|
if (pch_iir & SDE_AUX_MASK)
|
2016-05-06 13:48:28 +00:00
|
|
|
dp_aux_irq_handler(dev_priv);
|
2012-12-01 12:53:47 +00:00
|
|
|
|
2011-01-04 23:09:39 +00:00
|
|
|
if (pch_iir & SDE_GMBUS)
|
2016-05-06 13:48:28 +00:00
|
|
|
gmbus_irq_handler(dev_priv);
|
2011-01-04 23:09:39 +00:00
|
|
|
|
|
|
|
if (pch_iir & SDE_AUDIO_HDCP_MASK)
|
|
|
|
DRM_DEBUG_DRIVER("PCH HDCP audio interrupt\n");
|
|
|
|
|
|
|
|
if (pch_iir & SDE_AUDIO_TRANS_MASK)
|
|
|
|
DRM_DEBUG_DRIVER("PCH transcoder audio interrupt\n");
|
|
|
|
|
|
|
|
if (pch_iir & SDE_POISON)
|
|
|
|
DRM_ERROR("PCH poison interrupt\n");
|
|
|
|
|
2011-02-07 20:26:52 +00:00
|
|
|
if (pch_iir & SDE_FDI_MASK)
|
2014-08-18 12:49:10 +00:00
|
|
|
for_each_pipe(dev_priv, pipe)
|
2011-02-07 20:26:52 +00:00
|
|
|
DRM_DEBUG_DRIVER(" pipe %c FDI IIR: 0x%08x\n",
|
|
|
|
pipe_name(pipe),
|
|
|
|
I915_READ(FDI_RX_IIR(pipe)));
|
2011-01-04 23:09:39 +00:00
|
|
|
|
|
|
|
if (pch_iir & (SDE_TRANSB_CRC_DONE | SDE_TRANSA_CRC_DONE))
|
|
|
|
DRM_DEBUG_DRIVER("PCH transcoder CRC done interrupt\n");
|
|
|
|
|
|
|
|
if (pch_iir & (SDE_TRANSB_CRC_ERR | SDE_TRANSA_CRC_ERR))
|
|
|
|
DRM_DEBUG_DRIVER("PCH transcoder CRC error interrupt\n");
|
|
|
|
|
|
|
|
if (pch_iir & SDE_TRANSA_FIFO_UNDER)
|
2017-07-17 18:14:03 +00:00
|
|
|
intel_pch_fifo_underrun_irq_handler(dev_priv, PIPE_A);
|
drm/i915: report Gen5+ CPU and PCH FIFO underruns
In this commit we enable both CPU and PCH FIFO underrun reporting and
start reporting them. We follow a few rules:
- after we receive one of these errors, we mask the interrupt, so
we won't get an "interrupt storm" and we also won't flood dmesg;
- at each mode set we enable the interrupts again, so we'll see each
message at most once per mode set;
- in the specific places where we need to ignore the errors, we
completely mask the interrupts.
The downside of this patch is that since we're completely disabling
(masking) the interrupts instead of just not printing error messages,
we will mask more than just what we want on IVB/HSW CPU interrupts
(due to GEN7_ERR_INT) and on CPT/PPT/LPT PCHs (due to SERR_INT). So
when we decide to mask PCH FIFO underruns for pipe A on CPT, we'll
also be masking PCH FIFO underruns for pipe B, because both are
reported by SERR_INT, which has to be either completely enabled or
completely disabled (in othe words, there's no way to disable/enable
specific bits of GEN7_ERR_INT and SERR_INT).
V2: Rename some functions and variables, downgrade messages to
DRM_DEBUG_DRIVER and rebase.
Signed-off-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Reviewed-by: Imre Deak <imre.deak@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-04-12 20:57:57 +00:00
|
|
|
|
|
|
|
if (pch_iir & SDE_TRANSB_FIFO_UNDER)
|
2017-07-17 18:14:03 +00:00
|
|
|
intel_pch_fifo_underrun_irq_handler(dev_priv, PIPE_B);
|
drm/i915: report Gen5+ CPU and PCH FIFO underruns
In this commit we enable both CPU and PCH FIFO underrun reporting and
start reporting them. We follow a few rules:
- after we receive one of these errors, we mask the interrupt, so
we won't get an "interrupt storm" and we also won't flood dmesg;
- at each mode set we enable the interrupts again, so we'll see each
message at most once per mode set;
- in the specific places where we need to ignore the errors, we
completely mask the interrupts.
The downside of this patch is that since we're completely disabling
(masking) the interrupts instead of just not printing error messages,
we will mask more than just what we want on IVB/HSW CPU interrupts
(due to GEN7_ERR_INT) and on CPT/PPT/LPT PCHs (due to SERR_INT). So
when we decide to mask PCH FIFO underruns for pipe A on CPT, we'll
also be masking PCH FIFO underruns for pipe B, because both are
reported by SERR_INT, which has to be either completely enabled or
completely disabled (in othe words, there's no way to disable/enable
specific bits of GEN7_ERR_INT and SERR_INT).
V2: Rename some functions and variables, downgrade messages to
DRM_DEBUG_DRIVER and rebase.
Signed-off-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Reviewed-by: Imre Deak <imre.deak@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-04-12 20:57:57 +00:00
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void ivb_err_int_handler(struct drm_i915_private *dev_priv)
|
drm/i915: report Gen5+ CPU and PCH FIFO underruns
In this commit we enable both CPU and PCH FIFO underrun reporting and
start reporting them. We follow a few rules:
- after we receive one of these errors, we mask the interrupt, so
we won't get an "interrupt storm" and we also won't flood dmesg;
- at each mode set we enable the interrupts again, so we'll see each
message at most once per mode set;
- in the specific places where we need to ignore the errors, we
completely mask the interrupts.
The downside of this patch is that since we're completely disabling
(masking) the interrupts instead of just not printing error messages,
we will mask more than just what we want on IVB/HSW CPU interrupts
(due to GEN7_ERR_INT) and on CPT/PPT/LPT PCHs (due to SERR_INT). So
when we decide to mask PCH FIFO underruns for pipe A on CPT, we'll
also be masking PCH FIFO underruns for pipe B, because both are
reported by SERR_INT, which has to be either completely enabled or
completely disabled (in othe words, there's no way to disable/enable
specific bits of GEN7_ERR_INT and SERR_INT).
V2: Rename some functions and variables, downgrade messages to
DRM_DEBUG_DRIVER and rebase.
Signed-off-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Reviewed-by: Imre Deak <imre.deak@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-04-12 20:57:57 +00:00
|
|
|
{
|
|
|
|
u32 err_int = I915_READ(GEN7_ERR_INT);
|
2013-10-16 20:55:52 +00:00
|
|
|
enum pipe pipe;
|
drm/i915: report Gen5+ CPU and PCH FIFO underruns
In this commit we enable both CPU and PCH FIFO underrun reporting and
start reporting them. We follow a few rules:
- after we receive one of these errors, we mask the interrupt, so
we won't get an "interrupt storm" and we also won't flood dmesg;
- at each mode set we enable the interrupts again, so we'll see each
message at most once per mode set;
- in the specific places where we need to ignore the errors, we
completely mask the interrupts.
The downside of this patch is that since we're completely disabling
(masking) the interrupts instead of just not printing error messages,
we will mask more than just what we want on IVB/HSW CPU interrupts
(due to GEN7_ERR_INT) and on CPT/PPT/LPT PCHs (due to SERR_INT). So
when we decide to mask PCH FIFO underruns for pipe A on CPT, we'll
also be masking PCH FIFO underruns for pipe B, because both are
reported by SERR_INT, which has to be either completely enabled or
completely disabled (in othe words, there's no way to disable/enable
specific bits of GEN7_ERR_INT and SERR_INT).
V2: Rename some functions and variables, downgrade messages to
DRM_DEBUG_DRIVER and rebase.
Signed-off-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Reviewed-by: Imre Deak <imre.deak@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-04-12 20:57:57 +00:00
|
|
|
|
2013-04-12 20:57:58 +00:00
|
|
|
if (err_int & ERR_INT_POISON)
|
|
|
|
DRM_ERROR("Poison interrupt\n");
|
|
|
|
|
2014-08-18 12:49:10 +00:00
|
|
|
for_each_pipe(dev_priv, pipe) {
|
2014-09-30 08:56:48 +00:00
|
|
|
if (err_int & ERR_INT_FIFO_UNDERRUN(pipe))
|
|
|
|
intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
|
2013-10-15 17:55:27 +00:00
|
|
|
|
2013-10-16 20:55:52 +00:00
|
|
|
if (err_int & ERR_INT_PIPE_CRC_DONE(pipe)) {
|
2016-05-06 13:48:28 +00:00
|
|
|
if (IS_IVYBRIDGE(dev_priv))
|
|
|
|
ivb_pipe_crc_irq_handler(dev_priv, pipe);
|
2013-10-16 20:55:52 +00:00
|
|
|
else
|
2016-05-06 13:48:28 +00:00
|
|
|
hsw_pipe_crc_irq_handler(dev_priv, pipe);
|
2013-10-16 20:55:52 +00:00
|
|
|
}
|
|
|
|
}
|
2013-10-15 17:55:27 +00:00
|
|
|
|
drm/i915: report Gen5+ CPU and PCH FIFO underruns
In this commit we enable both CPU and PCH FIFO underrun reporting and
start reporting them. We follow a few rules:
- after we receive one of these errors, we mask the interrupt, so
we won't get an "interrupt storm" and we also won't flood dmesg;
- at each mode set we enable the interrupts again, so we'll see each
message at most once per mode set;
- in the specific places where we need to ignore the errors, we
completely mask the interrupts.
The downside of this patch is that since we're completely disabling
(masking) the interrupts instead of just not printing error messages,
we will mask more than just what we want on IVB/HSW CPU interrupts
(due to GEN7_ERR_INT) and on CPT/PPT/LPT PCHs (due to SERR_INT). So
when we decide to mask PCH FIFO underruns for pipe A on CPT, we'll
also be masking PCH FIFO underruns for pipe B, because both are
reported by SERR_INT, which has to be either completely enabled or
completely disabled (in othe words, there's no way to disable/enable
specific bits of GEN7_ERR_INT and SERR_INT).
V2: Rename some functions and variables, downgrade messages to
DRM_DEBUG_DRIVER and rebase.
Signed-off-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Reviewed-by: Imre Deak <imre.deak@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-04-12 20:57:57 +00:00
|
|
|
I915_WRITE(GEN7_ERR_INT, err_int);
|
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void cpt_serr_int_handler(struct drm_i915_private *dev_priv)
|
drm/i915: report Gen5+ CPU and PCH FIFO underruns
In this commit we enable both CPU and PCH FIFO underrun reporting and
start reporting them. We follow a few rules:
- after we receive one of these errors, we mask the interrupt, so
we won't get an "interrupt storm" and we also won't flood dmesg;
- at each mode set we enable the interrupts again, so we'll see each
message at most once per mode set;
- in the specific places where we need to ignore the errors, we
completely mask the interrupts.
The downside of this patch is that since we're completely disabling
(masking) the interrupts instead of just not printing error messages,
we will mask more than just what we want on IVB/HSW CPU interrupts
(due to GEN7_ERR_INT) and on CPT/PPT/LPT PCHs (due to SERR_INT). So
when we decide to mask PCH FIFO underruns for pipe A on CPT, we'll
also be masking PCH FIFO underruns for pipe B, because both are
reported by SERR_INT, which has to be either completely enabled or
completely disabled (in othe words, there's no way to disable/enable
specific bits of GEN7_ERR_INT and SERR_INT).
V2: Rename some functions and variables, downgrade messages to
DRM_DEBUG_DRIVER and rebase.
Signed-off-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Reviewed-by: Imre Deak <imre.deak@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-04-12 20:57:57 +00:00
|
|
|
{
|
|
|
|
u32 serr_int = I915_READ(SERR_INT);
|
|
|
|
|
2013-04-12 20:57:58 +00:00
|
|
|
if (serr_int & SERR_INT_POISON)
|
|
|
|
DRM_ERROR("PCH poison interrupt\n");
|
|
|
|
|
drm/i915: report Gen5+ CPU and PCH FIFO underruns
In this commit we enable both CPU and PCH FIFO underrun reporting and
start reporting them. We follow a few rules:
- after we receive one of these errors, we mask the interrupt, so
we won't get an "interrupt storm" and we also won't flood dmesg;
- at each mode set we enable the interrupts again, so we'll see each
message at most once per mode set;
- in the specific places where we need to ignore the errors, we
completely mask the interrupts.
The downside of this patch is that since we're completely disabling
(masking) the interrupts instead of just not printing error messages,
we will mask more than just what we want on IVB/HSW CPU interrupts
(due to GEN7_ERR_INT) and on CPT/PPT/LPT PCHs (due to SERR_INT). So
when we decide to mask PCH FIFO underruns for pipe A on CPT, we'll
also be masking PCH FIFO underruns for pipe B, because both are
reported by SERR_INT, which has to be either completely enabled or
completely disabled (in othe words, there's no way to disable/enable
specific bits of GEN7_ERR_INT and SERR_INT).
V2: Rename some functions and variables, downgrade messages to
DRM_DEBUG_DRIVER and rebase.
Signed-off-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Reviewed-by: Imre Deak <imre.deak@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-04-12 20:57:57 +00:00
|
|
|
if (serr_int & SERR_INT_TRANS_A_FIFO_UNDERRUN)
|
2017-07-17 18:14:03 +00:00
|
|
|
intel_pch_fifo_underrun_irq_handler(dev_priv, PIPE_A);
|
drm/i915: report Gen5+ CPU and PCH FIFO underruns
In this commit we enable both CPU and PCH FIFO underrun reporting and
start reporting them. We follow a few rules:
- after we receive one of these errors, we mask the interrupt, so
we won't get an "interrupt storm" and we also won't flood dmesg;
- at each mode set we enable the interrupts again, so we'll see each
message at most once per mode set;
- in the specific places where we need to ignore the errors, we
completely mask the interrupts.
The downside of this patch is that since we're completely disabling
(masking) the interrupts instead of just not printing error messages,
we will mask more than just what we want on IVB/HSW CPU interrupts
(due to GEN7_ERR_INT) and on CPT/PPT/LPT PCHs (due to SERR_INT). So
when we decide to mask PCH FIFO underruns for pipe A on CPT, we'll
also be masking PCH FIFO underruns for pipe B, because both are
reported by SERR_INT, which has to be either completely enabled or
completely disabled (in othe words, there's no way to disable/enable
specific bits of GEN7_ERR_INT and SERR_INT).
V2: Rename some functions and variables, downgrade messages to
DRM_DEBUG_DRIVER and rebase.
Signed-off-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Reviewed-by: Imre Deak <imre.deak@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-04-12 20:57:57 +00:00
|
|
|
|
|
|
|
if (serr_int & SERR_INT_TRANS_B_FIFO_UNDERRUN)
|
2017-07-17 18:14:03 +00:00
|
|
|
intel_pch_fifo_underrun_irq_handler(dev_priv, PIPE_B);
|
drm/i915: report Gen5+ CPU and PCH FIFO underruns
In this commit we enable both CPU and PCH FIFO underrun reporting and
start reporting them. We follow a few rules:
- after we receive one of these errors, we mask the interrupt, so
we won't get an "interrupt storm" and we also won't flood dmesg;
- at each mode set we enable the interrupts again, so we'll see each
message at most once per mode set;
- in the specific places where we need to ignore the errors, we
completely mask the interrupts.
The downside of this patch is that since we're completely disabling
(masking) the interrupts instead of just not printing error messages,
we will mask more than just what we want on IVB/HSW CPU interrupts
(due to GEN7_ERR_INT) and on CPT/PPT/LPT PCHs (due to SERR_INT). So
when we decide to mask PCH FIFO underruns for pipe A on CPT, we'll
also be masking PCH FIFO underruns for pipe B, because both are
reported by SERR_INT, which has to be either completely enabled or
completely disabled (in othe words, there's no way to disable/enable
specific bits of GEN7_ERR_INT and SERR_INT).
V2: Rename some functions and variables, downgrade messages to
DRM_DEBUG_DRIVER and rebase.
Signed-off-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Reviewed-by: Imre Deak <imre.deak@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-04-12 20:57:57 +00:00
|
|
|
|
|
|
|
if (serr_int & SERR_INT_TRANS_C_FIFO_UNDERRUN)
|
2017-07-17 18:14:03 +00:00
|
|
|
intel_pch_fifo_underrun_irq_handler(dev_priv, PIPE_C);
|
drm/i915: report Gen5+ CPU and PCH FIFO underruns
In this commit we enable both CPU and PCH FIFO underrun reporting and
start reporting them. We follow a few rules:
- after we receive one of these errors, we mask the interrupt, so
we won't get an "interrupt storm" and we also won't flood dmesg;
- at each mode set we enable the interrupts again, so we'll see each
message at most once per mode set;
- in the specific places where we need to ignore the errors, we
completely mask the interrupts.
The downside of this patch is that since we're completely disabling
(masking) the interrupts instead of just not printing error messages,
we will mask more than just what we want on IVB/HSW CPU interrupts
(due to GEN7_ERR_INT) and on CPT/PPT/LPT PCHs (due to SERR_INT). So
when we decide to mask PCH FIFO underruns for pipe A on CPT, we'll
also be masking PCH FIFO underruns for pipe B, because both are
reported by SERR_INT, which has to be either completely enabled or
completely disabled (in othe words, there's no way to disable/enable
specific bits of GEN7_ERR_INT and SERR_INT).
V2: Rename some functions and variables, downgrade messages to
DRM_DEBUG_DRIVER and rebase.
Signed-off-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Reviewed-by: Imre Deak <imre.deak@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-04-12 20:57:57 +00:00
|
|
|
|
|
|
|
I915_WRITE(SERR_INT, serr_int);
|
2011-01-04 23:09:39 +00:00
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void cpt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
|
2012-06-06 19:45:44 +00:00
|
|
|
{
|
|
|
|
int pipe;
|
2015-08-27 20:56:02 +00:00
|
|
|
u32 hotplug_trigger = pch_iir & SDE_HOTPLUG_MASK_CPT;
|
2014-06-18 01:29:35 +00:00
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
ibx_hpd_irq_handler(dev_priv, hotplug_trigger, hpd_cpt);
|
2013-06-27 15:52:14 +00:00
|
|
|
|
2013-04-17 14:48:48 +00:00
|
|
|
if (pch_iir & SDE_AUDIO_POWER_MASK_CPT) {
|
|
|
|
int port = ffs((pch_iir & SDE_AUDIO_POWER_MASK_CPT) >>
|
|
|
|
SDE_AUDIO_POWER_SHIFT_CPT);
|
|
|
|
DRM_DEBUG_DRIVER("PCH audio power change on port %c\n",
|
|
|
|
port_name(port));
|
|
|
|
}
|
2012-06-06 19:45:44 +00:00
|
|
|
|
|
|
|
if (pch_iir & SDE_AUX_MASK_CPT)
|
2016-05-06 13:48:28 +00:00
|
|
|
dp_aux_irq_handler(dev_priv);
|
2012-06-06 19:45:44 +00:00
|
|
|
|
|
|
|
if (pch_iir & SDE_GMBUS_CPT)
|
2016-05-06 13:48:28 +00:00
|
|
|
gmbus_irq_handler(dev_priv);
|
2012-06-06 19:45:44 +00:00
|
|
|
|
|
|
|
if (pch_iir & SDE_AUDIO_CP_REQ_CPT)
|
|
|
|
DRM_DEBUG_DRIVER("Audio CP request interrupt\n");
|
|
|
|
|
|
|
|
if (pch_iir & SDE_AUDIO_CP_CHG_CPT)
|
|
|
|
DRM_DEBUG_DRIVER("Audio CP change interrupt\n");
|
|
|
|
|
|
|
|
if (pch_iir & SDE_FDI_MASK_CPT)
|
2014-08-18 12:49:10 +00:00
|
|
|
for_each_pipe(dev_priv, pipe)
|
2012-06-06 19:45:44 +00:00
|
|
|
DRM_DEBUG_DRIVER(" pipe %c FDI IIR: 0x%08x\n",
|
|
|
|
pipe_name(pipe),
|
|
|
|
I915_READ(FDI_RX_IIR(pipe)));
|
drm/i915: report Gen5+ CPU and PCH FIFO underruns
In this commit we enable both CPU and PCH FIFO underrun reporting and
start reporting them. We follow a few rules:
- after we receive one of these errors, we mask the interrupt, so
we won't get an "interrupt storm" and we also won't flood dmesg;
- at each mode set we enable the interrupts again, so we'll see each
message at most once per mode set;
- in the specific places where we need to ignore the errors, we
completely mask the interrupts.
The downside of this patch is that since we're completely disabling
(masking) the interrupts instead of just not printing error messages,
we will mask more than just what we want on IVB/HSW CPU interrupts
(due to GEN7_ERR_INT) and on CPT/PPT/LPT PCHs (due to SERR_INT). So
when we decide to mask PCH FIFO underruns for pipe A on CPT, we'll
also be masking PCH FIFO underruns for pipe B, because both are
reported by SERR_INT, which has to be either completely enabled or
completely disabled (in othe words, there's no way to disable/enable
specific bits of GEN7_ERR_INT and SERR_INT).
V2: Rename some functions and variables, downgrade messages to
DRM_DEBUG_DRIVER and rebase.
Signed-off-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Reviewed-by: Imre Deak <imre.deak@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-04-12 20:57:57 +00:00
|
|
|
|
|
|
|
if (pch_iir & SDE_ERROR_CPT)
|
2016-05-06 13:48:28 +00:00
|
|
|
cpt_serr_int_handler(dev_priv);
|
2012-06-06 19:45:44 +00:00
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void spt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
|
2015-08-27 20:56:02 +00:00
|
|
|
{
|
|
|
|
u32 hotplug_trigger = pch_iir & SDE_HOTPLUG_MASK_SPT &
|
|
|
|
~SDE_PORTE_HOTPLUG_SPT;
|
|
|
|
u32 hotplug2_trigger = pch_iir & SDE_PORTE_HOTPLUG_SPT;
|
|
|
|
u32 pin_mask = 0, long_mask = 0;
|
|
|
|
|
|
|
|
if (hotplug_trigger) {
|
|
|
|
u32 dig_hotplug_reg;
|
|
|
|
|
|
|
|
dig_hotplug_reg = I915_READ(PCH_PORT_HOTPLUG);
|
|
|
|
I915_WRITE(PCH_PORT_HOTPLUG, dig_hotplug_reg);
|
|
|
|
|
|
|
|
intel_get_hpd_pins(&pin_mask, &long_mask, hotplug_trigger,
|
|
|
|
dig_hotplug_reg, hpd_spt,
|
2015-08-27 20:56:07 +00:00
|
|
|
spt_port_hotplug_long_detect);
|
2015-08-27 20:56:02 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if (hotplug2_trigger) {
|
|
|
|
u32 dig_hotplug_reg;
|
|
|
|
|
|
|
|
dig_hotplug_reg = I915_READ(PCH_PORT_HOTPLUG2);
|
|
|
|
I915_WRITE(PCH_PORT_HOTPLUG2, dig_hotplug_reg);
|
|
|
|
|
|
|
|
intel_get_hpd_pins(&pin_mask, &long_mask, hotplug2_trigger,
|
|
|
|
dig_hotplug_reg, hpd_spt,
|
|
|
|
spt_port_hotplug2_long_detect);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (pin_mask)
|
2016-05-06 13:48:28 +00:00
|
|
|
intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
|
2015-08-27 20:56:02 +00:00
|
|
|
|
|
|
|
if (pch_iir & SDE_GMBUS_CPT)
|
2016-05-06 13:48:28 +00:00
|
|
|
gmbus_irq_handler(dev_priv);
|
2015-08-27 20:56:02 +00:00
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void ilk_hpd_irq_handler(struct drm_i915_private *dev_priv,
|
|
|
|
u32 hotplug_trigger,
|
2015-08-27 20:56:10 +00:00
|
|
|
const u32 hpd[HPD_NUM_PINS])
|
|
|
|
{
|
|
|
|
u32 dig_hotplug_reg, pin_mask = 0, long_mask = 0;
|
|
|
|
|
|
|
|
dig_hotplug_reg = I915_READ(DIGITAL_PORT_HOTPLUG_CNTRL);
|
|
|
|
I915_WRITE(DIGITAL_PORT_HOTPLUG_CNTRL, dig_hotplug_reg);
|
|
|
|
|
|
|
|
intel_get_hpd_pins(&pin_mask, &long_mask, hotplug_trigger,
|
|
|
|
dig_hotplug_reg, hpd,
|
|
|
|
ilk_port_hotplug_long_detect);
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
|
2015-08-27 20:56:10 +00:00
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void ilk_display_irq_handler(struct drm_i915_private *dev_priv,
|
|
|
|
u32 de_iir)
|
2013-07-12 19:35:10 +00:00
|
|
|
{
|
2013-10-21 16:04:36 +00:00
|
|
|
enum pipe pipe;
|
2015-08-27 20:56:03 +00:00
|
|
|
u32 hotplug_trigger = de_iir & DE_DP_A_HOTPLUG;
|
|
|
|
|
2015-08-27 20:56:10 +00:00
|
|
|
if (hotplug_trigger)
|
2016-05-06 13:48:28 +00:00
|
|
|
ilk_hpd_irq_handler(dev_priv, hotplug_trigger, hpd_ilk);
|
2013-07-12 19:35:10 +00:00
|
|
|
|
|
|
|
if (de_iir & DE_AUX_CHANNEL_A)
|
2016-05-06 13:48:28 +00:00
|
|
|
dp_aux_irq_handler(dev_priv);
|
2013-07-12 19:35:10 +00:00
|
|
|
|
|
|
|
if (de_iir & DE_GSE)
|
2016-05-06 13:48:28 +00:00
|
|
|
intel_opregion_asle_intr(dev_priv);
|
2013-07-12 19:35:10 +00:00
|
|
|
|
|
|
|
if (de_iir & DE_POISON)
|
|
|
|
DRM_ERROR("Poison interrupt\n");
|
|
|
|
|
2014-08-18 12:49:10 +00:00
|
|
|
for_each_pipe(dev_priv, pipe) {
|
2017-07-20 17:57:51 +00:00
|
|
|
if (de_iir & DE_PIPE_VBLANK(pipe))
|
|
|
|
drm_handle_vblank(&dev_priv->drm, pipe);
|
2013-10-16 20:55:48 +00:00
|
|
|
|
2013-10-21 16:04:36 +00:00
|
|
|
if (de_iir & DE_PIPE_FIFO_UNDERRUN(pipe))
|
2014-09-30 08:56:48 +00:00
|
|
|
intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
|
2013-10-16 20:55:48 +00:00
|
|
|
|
2013-10-21 16:04:36 +00:00
|
|
|
if (de_iir & DE_PIPE_CRC_DONE(pipe))
|
2016-05-06 13:48:28 +00:00
|
|
|
i9xx_pipe_crc_irq_handler(dev_priv, pipe);
|
2013-07-12 19:35:10 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* check event from PCH */
|
|
|
|
if (de_iir & DE_PCH_EVENT) {
|
|
|
|
u32 pch_iir = I915_READ(SDEIIR);
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
if (HAS_PCH_CPT(dev_priv))
|
|
|
|
cpt_irq_handler(dev_priv, pch_iir);
|
2013-07-12 19:35:10 +00:00
|
|
|
else
|
2016-05-06 13:48:28 +00:00
|
|
|
ibx_irq_handler(dev_priv, pch_iir);
|
2013-07-12 19:35:10 +00:00
|
|
|
|
|
|
|
/* should clear PCH hotplug event before clear CPU irq */
|
|
|
|
I915_WRITE(SDEIIR, pch_iir);
|
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
if (IS_GEN5(dev_priv) && de_iir & DE_PCU_EVENT)
|
|
|
|
ironlake_rps_change_irq_handler(dev_priv);
|
2013-07-12 19:35:10 +00:00
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void ivb_display_irq_handler(struct drm_i915_private *dev_priv,
|
|
|
|
u32 de_iir)
|
2013-07-12 19:35:11 +00:00
|
|
|
{
|
2014-03-03 17:31:46 +00:00
|
|
|
enum pipe pipe;
|
2015-08-27 20:56:04 +00:00
|
|
|
u32 hotplug_trigger = de_iir & DE_DP_A_HOTPLUG_IVB;
|
|
|
|
|
2015-08-27 20:56:10 +00:00
|
|
|
if (hotplug_trigger)
|
2016-05-06 13:48:28 +00:00
|
|
|
ilk_hpd_irq_handler(dev_priv, hotplug_trigger, hpd_ivb);
|
2013-07-12 19:35:11 +00:00
|
|
|
|
|
|
|
if (de_iir & DE_ERR_INT_IVB)
|
2016-05-06 13:48:28 +00:00
|
|
|
ivb_err_int_handler(dev_priv);
|
2013-07-12 19:35:11 +00:00
|
|
|
|
|
|
|
if (de_iir & DE_AUX_CHANNEL_A_IVB)
|
2016-05-06 13:48:28 +00:00
|
|
|
dp_aux_irq_handler(dev_priv);
|
2013-07-12 19:35:11 +00:00
|
|
|
|
|
|
|
if (de_iir & DE_GSE_IVB)
|
2016-05-06 13:48:28 +00:00
|
|
|
intel_opregion_asle_intr(dev_priv);
|
2013-07-12 19:35:11 +00:00
|
|
|
|
2014-08-18 12:49:10 +00:00
|
|
|
for_each_pipe(dev_priv, pipe) {
|
2017-07-20 17:57:51 +00:00
|
|
|
if (de_iir & (DE_PIPE_VBLANK_IVB(pipe)))
|
|
|
|
drm_handle_vblank(&dev_priv->drm, pipe);
|
2013-07-12 19:35:11 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* check event from PCH */
|
2016-05-06 13:48:28 +00:00
|
|
|
if (!HAS_PCH_NOP(dev_priv) && (de_iir & DE_PCH_EVENT_IVB)) {
|
2013-07-12 19:35:11 +00:00
|
|
|
u32 pch_iir = I915_READ(SDEIIR);
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
cpt_irq_handler(dev_priv, pch_iir);
|
2013-07-12 19:35:11 +00:00
|
|
|
|
|
|
|
/* clear PCH hotplug event before clear CPU irq */
|
|
|
|
I915_WRITE(SDEIIR, pch_iir);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-06-16 15:10:57 +00:00
|
|
|
/*
|
|
|
|
* To handle irqs with the minimum potential races with fresh interrupts, we:
|
|
|
|
* 1 - Disable Master Interrupt Control.
|
|
|
|
* 2 - Find the source(s) of the interrupt.
|
|
|
|
* 3 - Clear the Interrupt Identity bits (IIR).
|
|
|
|
* 4 - Process the interrupt(s) that had bits set in the IIRs.
|
|
|
|
* 5 - Re-enable Master Interrupt Control.
|
|
|
|
*/
|
2013-07-12 22:56:30 +00:00
|
|
|
static irqreturn_t ironlake_irq_handler(int irq, void *arg)
|
2011-04-06 19:13:38 +00:00
|
|
|
{
|
2014-05-12 17:17:55 +00:00
|
|
|
struct drm_device *dev = arg;
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2013-07-12 22:56:30 +00:00
|
|
|
u32 de_iir, gt_iir, de_ier, sde_ier = 0;
|
2012-05-09 20:45:44 +00:00
|
|
|
irqreturn_t ret = IRQ_NONE;
|
2011-04-06 19:13:38 +00:00
|
|
|
|
2015-02-24 09:14:30 +00:00
|
|
|
if (!intel_irqs_enabled(dev_priv))
|
|
|
|
return IRQ_NONE;
|
|
|
|
|
2015-12-16 00:52:19 +00:00
|
|
|
/* IRQs are synced during runtime_suspend, we don't require a wakeref */
|
|
|
|
disable_rpm_wakeref_asserts(dev_priv);
|
|
|
|
|
2011-04-06 19:13:38 +00:00
|
|
|
/* disable master interrupt before clearing iir */
|
|
|
|
de_ier = I915_READ(DEIER);
|
|
|
|
I915_WRITE(DEIER, de_ier & ~DE_MASTER_IRQ_CONTROL);
|
2013-07-12 19:35:14 +00:00
|
|
|
POSTING_READ(DEIER);
|
2011-04-06 19:13:38 +00:00
|
|
|
|
drm/i915: also disable south interrupts when handling them
From the docs:
"IIR can queue up to two interrupt events. When the IIR is cleared,
it will set itself again after one clock if a second event was
stored."
"Only the rising edge of the PCH Display interrupt will cause the
North Display IIR (DEIIR) PCH Display Interrupt even bit to be set,
so all PCH Display Interrupts, including back to back interrupts,
must be cleared before a new PCH Display interrupt can cause DEIIR
to be set".
The current code works fine because we don't get many interrupts, but
if we enable the PCH FIFO underrun interrupts we'll start getting so
many interrupts that at some point new PCH interrupts won't cause
DEIIR to be set.
The initial implementation I tried was to turn the code that checks
SDEIIR into a loop, but we can still get interrupts even after the
loop is done (and before the irq handler finishes), so we have to
either disable the interrupts or mask them. In the end I concluded
that just disabling the PCH interrupts is enough, you don't even need
the loop, so this is what this patch implements. I've tested it and it
passes the 2 "PCH FIFO underrun interrupt storms" I can reproduce:
the "ironlake_crtc_disable" case and the "wrong watermarks" case.
In other words, here's how to reproduce the problem fixed by this
patch:
1 - Enable PCH FIFO underrun interrupts (SERR_INT on SNB+)
2 - Boot the machine
3 - While booting we'll get tons of PCH FIFO underrun interrupts
4 - Plug a new monitor
5 - Run xrandr, notice it won't detect the new monitor
6 - Read SDEIIR and notice it's not 0 while DEIIR is 0
Q: Can't we just clear DEIIR before SDEIIR?
A: It doesn't work. SDEIIR has to be completely cleared (including the
interrupts stored on its back queue) before it can flip DEIIR's bit to
1 again, and even while you're clearing it you'll be getting more and
more interrupts.
Q: Why does it work by just disabling+enabling the south interrupts?
A: Because when we re-enable them, if there's something on the SDEIIR
register (maybe an interrupt stored on the queue), the re-enabling
will make DEIIR's bit flip to 1, and since we'll already have
interrupts enabled we'll get another interrupt, then run our irq
handler again to process the "back" interrupts.
v2: Even bigger commit message, added code comments.
Note that this fixes missed dp aux irqs which have been reported for
3.9-rc1. This regression has been introduced by switching to
irq-driven dp aux transactions with
commit 9ee32fea5fe810ec06af3a15e4c65478de56d4f5
Author: Daniel Vetter <daniel.vetter@ffwll.ch>
Date: Sat Dec 1 13:53:48 2012 +0100
drm/i915: irq-drive the dp aux communication
References: http://www.mail-archive.com/intel-gfx@lists.freedesktop.org/msg18588.html
References: https://lkml.org/lkml/2013/2/26/769
Tested-by: Imre Deak <imre.deak@intel.com>
Reported-by: Sedat Dilek <sedat.dilek@gmail.com>
Reported-by: Linus Torvalds <torvalds@linux-foundation.org>
Signed-off-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
[danvet: Pimp commit message with references for the dp aux irq
timeout regression this fixes.]
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-02-22 20:05:28 +00:00
|
|
|
/* Disable south interrupts. We'll only write to SDEIIR once, so further
|
|
|
|
* interrupts will will be stored on its back queue, and then we'll be
|
|
|
|
* able to process them after we restore SDEIER (as soon as we restore
|
|
|
|
* it, we'll get an interrupt if SDEIIR still has something to process
|
|
|
|
* due to its back queue). */
|
2016-05-06 13:48:28 +00:00
|
|
|
if (!HAS_PCH_NOP(dev_priv)) {
|
2013-04-05 20:12:41 +00:00
|
|
|
sde_ier = I915_READ(SDEIER);
|
|
|
|
I915_WRITE(SDEIER, 0);
|
|
|
|
POSTING_READ(SDEIER);
|
|
|
|
}
|
drm/i915: also disable south interrupts when handling them
From the docs:
"IIR can queue up to two interrupt events. When the IIR is cleared,
it will set itself again after one clock if a second event was
stored."
"Only the rising edge of the PCH Display interrupt will cause the
North Display IIR (DEIIR) PCH Display Interrupt even bit to be set,
so all PCH Display Interrupts, including back to back interrupts,
must be cleared before a new PCH Display interrupt can cause DEIIR
to be set".
The current code works fine because we don't get many interrupts, but
if we enable the PCH FIFO underrun interrupts we'll start getting so
many interrupts that at some point new PCH interrupts won't cause
DEIIR to be set.
The initial implementation I tried was to turn the code that checks
SDEIIR into a loop, but we can still get interrupts even after the
loop is done (and before the irq handler finishes), so we have to
either disable the interrupts or mask them. In the end I concluded
that just disabling the PCH interrupts is enough, you don't even need
the loop, so this is what this patch implements. I've tested it and it
passes the 2 "PCH FIFO underrun interrupt storms" I can reproduce:
the "ironlake_crtc_disable" case and the "wrong watermarks" case.
In other words, here's how to reproduce the problem fixed by this
patch:
1 - Enable PCH FIFO underrun interrupts (SERR_INT on SNB+)
2 - Boot the machine
3 - While booting we'll get tons of PCH FIFO underrun interrupts
4 - Plug a new monitor
5 - Run xrandr, notice it won't detect the new monitor
6 - Read SDEIIR and notice it's not 0 while DEIIR is 0
Q: Can't we just clear DEIIR before SDEIIR?
A: It doesn't work. SDEIIR has to be completely cleared (including the
interrupts stored on its back queue) before it can flip DEIIR's bit to
1 again, and even while you're clearing it you'll be getting more and
more interrupts.
Q: Why does it work by just disabling+enabling the south interrupts?
A: Because when we re-enable them, if there's something on the SDEIIR
register (maybe an interrupt stored on the queue), the re-enabling
will make DEIIR's bit flip to 1, and since we'll already have
interrupts enabled we'll get another interrupt, then run our irq
handler again to process the "back" interrupts.
v2: Even bigger commit message, added code comments.
Note that this fixes missed dp aux irqs which have been reported for
3.9-rc1. This regression has been introduced by switching to
irq-driven dp aux transactions with
commit 9ee32fea5fe810ec06af3a15e4c65478de56d4f5
Author: Daniel Vetter <daniel.vetter@ffwll.ch>
Date: Sat Dec 1 13:53:48 2012 +0100
drm/i915: irq-drive the dp aux communication
References: http://www.mail-archive.com/intel-gfx@lists.freedesktop.org/msg18588.html
References: https://lkml.org/lkml/2013/2/26/769
Tested-by: Imre Deak <imre.deak@intel.com>
Reported-by: Sedat Dilek <sedat.dilek@gmail.com>
Reported-by: Linus Torvalds <torvalds@linux-foundation.org>
Signed-off-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
[danvet: Pimp commit message with references for the dp aux irq
timeout regression this fixes.]
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-02-22 20:05:28 +00:00
|
|
|
|
2014-06-16 15:10:57 +00:00
|
|
|
/* Find, clear, then process each source of interrupt */
|
|
|
|
|
2011-04-06 19:13:38 +00:00
|
|
|
gt_iir = I915_READ(GTIIR);
|
2012-05-09 20:45:44 +00:00
|
|
|
if (gt_iir) {
|
2014-06-16 15:10:57 +00:00
|
|
|
I915_WRITE(GTIIR, gt_iir);
|
|
|
|
ret = IRQ_HANDLED;
|
2016-05-06 13:48:28 +00:00
|
|
|
if (INTEL_GEN(dev_priv) >= 6)
|
2016-04-13 18:19:57 +00:00
|
|
|
snb_gt_irq_handler(dev_priv, gt_iir);
|
2013-07-19 21:57:55 +00:00
|
|
|
else
|
2016-04-13 18:19:57 +00:00
|
|
|
ilk_gt_irq_handler(dev_priv, gt_iir);
|
2011-04-06 19:13:38 +00:00
|
|
|
}
|
|
|
|
|
2012-05-09 20:45:44 +00:00
|
|
|
de_iir = I915_READ(DEIIR);
|
|
|
|
if (de_iir) {
|
2014-06-16 15:10:57 +00:00
|
|
|
I915_WRITE(DEIIR, de_iir);
|
|
|
|
ret = IRQ_HANDLED;
|
2016-05-06 13:48:28 +00:00
|
|
|
if (INTEL_GEN(dev_priv) >= 7)
|
|
|
|
ivb_display_irq_handler(dev_priv, de_iir);
|
2013-07-12 22:56:30 +00:00
|
|
|
else
|
2016-05-06 13:48:28 +00:00
|
|
|
ilk_display_irq_handler(dev_priv, de_iir);
|
2011-04-06 19:13:38 +00:00
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
if (INTEL_GEN(dev_priv) >= 6) {
|
2013-07-12 22:56:30 +00:00
|
|
|
u32 pm_iir = I915_READ(GEN6_PMIIR);
|
|
|
|
if (pm_iir) {
|
|
|
|
I915_WRITE(GEN6_PMIIR, pm_iir);
|
|
|
|
ret = IRQ_HANDLED;
|
2014-06-16 15:10:57 +00:00
|
|
|
gen6_rps_irq_handler(dev_priv, pm_iir);
|
2013-07-12 22:56:30 +00:00
|
|
|
}
|
2012-05-09 20:45:44 +00:00
|
|
|
}
|
2011-04-06 19:13:38 +00:00
|
|
|
|
|
|
|
I915_WRITE(DEIER, de_ier);
|
|
|
|
POSTING_READ(DEIER);
|
2016-05-06 13:48:28 +00:00
|
|
|
if (!HAS_PCH_NOP(dev_priv)) {
|
2013-04-05 20:12:41 +00:00
|
|
|
I915_WRITE(SDEIER, sde_ier);
|
|
|
|
POSTING_READ(SDEIER);
|
|
|
|
}
|
2011-04-06 19:13:38 +00:00
|
|
|
|
2015-12-16 00:52:19 +00:00
|
|
|
/* IRQs are synced during runtime_suspend, we don't require a wakeref */
|
|
|
|
enable_rpm_wakeref_asserts(dev_priv);
|
|
|
|
|
2011-04-06 19:13:38 +00:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void bxt_hpd_irq_handler(struct drm_i915_private *dev_priv,
|
|
|
|
u32 hotplug_trigger,
|
2015-08-27 20:56:10 +00:00
|
|
|
const u32 hpd[HPD_NUM_PINS])
|
2014-08-22 12:10:41 +00:00
|
|
|
{
|
2015-08-27 20:56:09 +00:00
|
|
|
u32 dig_hotplug_reg, pin_mask = 0, long_mask = 0;
|
2014-08-22 12:10:41 +00:00
|
|
|
|
2015-08-27 20:56:11 +00:00
|
|
|
dig_hotplug_reg = I915_READ(PCH_PORT_HOTPLUG);
|
|
|
|
I915_WRITE(PCH_PORT_HOTPLUG, dig_hotplug_reg);
|
2014-08-22 12:10:41 +00:00
|
|
|
|
2015-08-27 20:56:09 +00:00
|
|
|
intel_get_hpd_pins(&pin_mask, &long_mask, hotplug_trigger,
|
2015-08-27 20:56:10 +00:00
|
|
|
dig_hotplug_reg, hpd,
|
2015-08-27 20:56:09 +00:00
|
|
|
bxt_port_hotplug_long_detect);
|
2015-08-27 20:56:10 +00:00
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
|
2014-08-22 12:10:41 +00:00
|
|
|
}
|
|
|
|
|
2016-01-12 16:04:07 +00:00
|
|
|
static irqreturn_t
|
|
|
|
gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl)
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
{
|
|
|
|
irqreturn_t ret = IRQ_NONE;
|
2016-01-12 16:04:07 +00:00
|
|
|
u32 iir;
|
2013-11-07 10:05:40 +00:00
|
|
|
enum pipe pipe;
|
2014-11-13 17:51:48 +00:00
|
|
|
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
if (master_ctl & GEN8_DE_MISC_IRQ) {
|
2016-01-12 16:04:06 +00:00
|
|
|
iir = I915_READ(GEN8_DE_MISC_IIR);
|
|
|
|
if (iir) {
|
|
|
|
I915_WRITE(GEN8_DE_MISC_IIR, iir);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
ret = IRQ_HANDLED;
|
2016-01-12 16:04:06 +00:00
|
|
|
if (iir & GEN8_DE_MISC_GSE)
|
2016-05-06 13:48:28 +00:00
|
|
|
intel_opregion_asle_intr(dev_priv);
|
2014-06-16 15:10:59 +00:00
|
|
|
else
|
|
|
|
DRM_ERROR("Unexpected DE Misc interrupt\n");
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
}
|
2014-06-16 15:10:59 +00:00
|
|
|
else
|
|
|
|
DRM_ERROR("The master control interrupt lied (DE MISC)!\n");
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
}
|
|
|
|
|
2013-11-07 13:49:55 +00:00
|
|
|
if (master_ctl & GEN8_DE_PORT_IRQ) {
|
2016-01-12 16:04:06 +00:00
|
|
|
iir = I915_READ(GEN8_DE_PORT_IIR);
|
|
|
|
if (iir) {
|
|
|
|
u32 tmp_mask;
|
2014-08-22 12:10:41 +00:00
|
|
|
bool found = false;
|
2015-08-27 20:56:09 +00:00
|
|
|
|
2016-01-12 16:04:06 +00:00
|
|
|
I915_WRITE(GEN8_DE_PORT_IIR, iir);
|
2013-11-07 13:49:55 +00:00
|
|
|
ret = IRQ_HANDLED;
|
2014-11-13 17:51:48 +00:00
|
|
|
|
2016-01-12 16:04:06 +00:00
|
|
|
tmp_mask = GEN8_AUX_CHANNEL_A;
|
2017-07-18 18:28:00 +00:00
|
|
|
if (INTEL_GEN(dev_priv) >= 9)
|
2016-01-12 16:04:06 +00:00
|
|
|
tmp_mask |= GEN9_AUX_CHANNEL_B |
|
|
|
|
GEN9_AUX_CHANNEL_C |
|
|
|
|
GEN9_AUX_CHANNEL_D;
|
|
|
|
|
|
|
|
if (iir & tmp_mask) {
|
2016-05-06 13:48:28 +00:00
|
|
|
dp_aux_irq_handler(dev_priv);
|
2014-08-22 12:10:41 +00:00
|
|
|
found = true;
|
|
|
|
}
|
|
|
|
|
2016-12-02 08:23:49 +00:00
|
|
|
if (IS_GEN9_LP(dev_priv)) {
|
2016-01-12 16:04:06 +00:00
|
|
|
tmp_mask = iir & BXT_DE_PORT_HOTPLUG_MASK;
|
|
|
|
if (tmp_mask) {
|
2016-05-06 13:48:28 +00:00
|
|
|
bxt_hpd_irq_handler(dev_priv, tmp_mask,
|
|
|
|
hpd_bxt);
|
2016-01-12 16:04:06 +00:00
|
|
|
found = true;
|
|
|
|
}
|
|
|
|
} else if (IS_BROADWELL(dev_priv)) {
|
|
|
|
tmp_mask = iir & GEN8_PORT_DP_A_HOTPLUG;
|
|
|
|
if (tmp_mask) {
|
2016-05-06 13:48:28 +00:00
|
|
|
ilk_hpd_irq_handler(dev_priv,
|
|
|
|
tmp_mask, hpd_bdw);
|
2016-01-12 16:04:06 +00:00
|
|
|
found = true;
|
|
|
|
}
|
2014-08-22 12:10:41 +00:00
|
|
|
}
|
|
|
|
|
2016-12-02 08:23:49 +00:00
|
|
|
if (IS_GEN9_LP(dev_priv) && (iir & BXT_DE_PORT_GMBUS)) {
|
2016-05-06 13:48:28 +00:00
|
|
|
gmbus_irq_handler(dev_priv);
|
2014-08-22 12:10:43 +00:00
|
|
|
found = true;
|
|
|
|
}
|
|
|
|
|
2014-08-22 12:10:41 +00:00
|
|
|
if (!found)
|
2014-06-16 15:10:59 +00:00
|
|
|
DRM_ERROR("Unexpected DE Port interrupt\n");
|
2013-11-07 13:49:55 +00:00
|
|
|
}
|
2014-06-16 15:10:59 +00:00
|
|
|
else
|
|
|
|
DRM_ERROR("The master control interrupt lied (DE PORT)!\n");
|
2013-11-07 13:49:55 +00:00
|
|
|
}
|
|
|
|
|
2014-08-18 12:49:10 +00:00
|
|
|
for_each_pipe(dev_priv, pipe) {
|
2017-07-20 17:57:51 +00:00
|
|
|
u32 fault_errors;
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
|
2013-11-07 10:05:40 +00:00
|
|
|
if (!(master_ctl & GEN8_DE_PIPE_IRQ(pipe)))
|
|
|
|
continue;
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
|
2016-01-12 16:04:06 +00:00
|
|
|
iir = I915_READ(GEN8_DE_PIPE_IIR(pipe));
|
|
|
|
if (!iir) {
|
|
|
|
DRM_ERROR("The master control interrupt lied (DE PIPE)!\n");
|
|
|
|
continue;
|
|
|
|
}
|
2014-03-20 20:45:01 +00:00
|
|
|
|
2016-01-12 16:04:06 +00:00
|
|
|
ret = IRQ_HANDLED;
|
|
|
|
I915_WRITE(GEN8_DE_PIPE_IIR(pipe), iir);
|
2014-06-16 15:10:59 +00:00
|
|
|
|
2017-07-20 17:57:51 +00:00
|
|
|
if (iir & GEN8_PIPE_VBLANK)
|
|
|
|
drm_handle_vblank(&dev_priv->drm, pipe);
|
2014-06-16 15:10:59 +00:00
|
|
|
|
2016-01-12 16:04:06 +00:00
|
|
|
if (iir & GEN8_PIPE_CDCLK_CRC_DONE)
|
2016-05-06 13:48:28 +00:00
|
|
|
hsw_pipe_crc_irq_handler(dev_priv, pipe);
|
2014-06-16 15:10:59 +00:00
|
|
|
|
2016-01-12 16:04:06 +00:00
|
|
|
if (iir & GEN8_PIPE_FIFO_UNDERRUN)
|
|
|
|
intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
|
2014-03-20 20:45:01 +00:00
|
|
|
|
2016-01-12 16:04:06 +00:00
|
|
|
fault_errors = iir;
|
2017-07-18 18:28:00 +00:00
|
|
|
if (INTEL_GEN(dev_priv) >= 9)
|
2016-01-12 16:04:06 +00:00
|
|
|
fault_errors &= GEN9_DE_PIPE_IRQ_FAULT_ERRORS;
|
|
|
|
else
|
|
|
|
fault_errors &= GEN8_DE_PIPE_IRQ_FAULT_ERRORS;
|
2014-03-20 20:45:01 +00:00
|
|
|
|
2016-01-12 16:04:06 +00:00
|
|
|
if (fault_errors)
|
2016-10-27 12:48:32 +00:00
|
|
|
DRM_ERROR("Fault errors on pipe %c: 0x%08x\n",
|
2016-01-12 16:04:06 +00:00
|
|
|
pipe_name(pipe),
|
|
|
|
fault_errors);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
if (HAS_PCH_SPLIT(dev_priv) && !HAS_PCH_NOP(dev_priv) &&
|
2014-08-22 12:10:42 +00:00
|
|
|
master_ctl & GEN8_DE_PCH_IRQ) {
|
2013-11-07 10:05:43 +00:00
|
|
|
/*
|
|
|
|
* FIXME(BDW): Assume for now that the new interrupt handling
|
|
|
|
* scheme also closed the SDE interrupt handling race we've seen
|
|
|
|
* on older pch-split platforms. But this needs testing.
|
|
|
|
*/
|
2016-01-12 16:04:06 +00:00
|
|
|
iir = I915_READ(SDEIIR);
|
|
|
|
if (iir) {
|
|
|
|
I915_WRITE(SDEIIR, iir);
|
2013-11-07 10:05:43 +00:00
|
|
|
ret = IRQ_HANDLED;
|
2015-08-27 20:56:02 +00:00
|
|
|
|
2017-06-02 20:06:39 +00:00
|
|
|
if (HAS_PCH_SPT(dev_priv) || HAS_PCH_KBP(dev_priv) ||
|
|
|
|
HAS_PCH_CNP(dev_priv))
|
2016-05-06 13:48:28 +00:00
|
|
|
spt_irq_handler(dev_priv, iir);
|
2015-08-27 20:56:02 +00:00
|
|
|
else
|
2016-05-06 13:48:28 +00:00
|
|
|
cpt_irq_handler(dev_priv, iir);
|
2016-01-07 08:29:10 +00:00
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* Like on previous PCH there seems to be something
|
|
|
|
* fishy going on with forwarding PCH interrupts.
|
|
|
|
*/
|
|
|
|
DRM_DEBUG_DRIVER("The master control interrupt lied (SDE)!\n");
|
|
|
|
}
|
2013-11-07 10:05:43 +00:00
|
|
|
}
|
|
|
|
|
2016-01-12 16:04:07 +00:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static irqreturn_t gen8_irq_handler(int irq, void *arg)
|
|
|
|
{
|
|
|
|
struct drm_device *dev = arg;
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2016-01-12 16:04:07 +00:00
|
|
|
u32 master_ctl;
|
2016-04-13 18:19:58 +00:00
|
|
|
u32 gt_iir[4] = {};
|
2016-01-12 16:04:07 +00:00
|
|
|
irqreturn_t ret;
|
|
|
|
|
|
|
|
if (!intel_irqs_enabled(dev_priv))
|
|
|
|
return IRQ_NONE;
|
|
|
|
|
|
|
|
master_ctl = I915_READ_FW(GEN8_MASTER_IRQ);
|
|
|
|
master_ctl &= ~GEN8_MASTER_IRQ_CONTROL;
|
|
|
|
if (!master_ctl)
|
|
|
|
return IRQ_NONE;
|
|
|
|
|
|
|
|
I915_WRITE_FW(GEN8_MASTER_IRQ, 0);
|
|
|
|
|
|
|
|
/* IRQs are synced during runtime_suspend, we don't require a wakeref */
|
|
|
|
disable_rpm_wakeref_asserts(dev_priv);
|
|
|
|
|
|
|
|
/* Find, clear, then process each source of interrupt */
|
2016-04-13 18:19:58 +00:00
|
|
|
ret = gen8_gt_irq_ack(dev_priv, master_ctl, gt_iir);
|
|
|
|
gen8_gt_irq_handler(dev_priv, gt_iir);
|
2016-01-12 16:04:07 +00:00
|
|
|
ret |= gen8_de_irq_handler(dev_priv, master_ctl);
|
|
|
|
|
2015-04-07 15:21:04 +00:00
|
|
|
I915_WRITE_FW(GEN8_MASTER_IRQ, GEN8_MASTER_IRQ_CONTROL);
|
|
|
|
POSTING_READ_FW(GEN8_MASTER_IRQ);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
|
2015-12-16 00:52:19 +00:00
|
|
|
enable_rpm_wakeref_asserts(dev_priv);
|
|
|
|
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2017-06-22 10:56:25 +00:00
|
|
|
struct wedge_me {
|
|
|
|
struct delayed_work work;
|
|
|
|
struct drm_i915_private *i915;
|
|
|
|
const char *name;
|
|
|
|
};
|
|
|
|
|
|
|
|
static void wedge_me(struct work_struct *work)
|
|
|
|
{
|
|
|
|
struct wedge_me *w = container_of(work, typeof(*w), work.work);
|
|
|
|
|
|
|
|
dev_err(w->i915->drm.dev,
|
|
|
|
"%s timed out, cancelling all in-flight rendering.\n",
|
|
|
|
w->name);
|
|
|
|
i915_gem_set_wedged(w->i915);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void __init_wedge(struct wedge_me *w,
|
|
|
|
struct drm_i915_private *i915,
|
|
|
|
long timeout,
|
|
|
|
const char *name)
|
|
|
|
{
|
|
|
|
w->i915 = i915;
|
|
|
|
w->name = name;
|
|
|
|
|
|
|
|
INIT_DELAYED_WORK_ONSTACK(&w->work, wedge_me);
|
|
|
|
schedule_delayed_work(&w->work, timeout);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void __fini_wedge(struct wedge_me *w)
|
|
|
|
{
|
|
|
|
cancel_delayed_work_sync(&w->work);
|
|
|
|
destroy_delayed_work_on_stack(&w->work);
|
|
|
|
w->i915 = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
#define i915_wedge_on_timeout(W, DEV, TIMEOUT) \
|
|
|
|
for (__init_wedge((W), (DEV), (TIMEOUT), __func__); \
|
|
|
|
(W)->i915; \
|
|
|
|
__fini_wedge((W)))
|
|
|
|
|
2009-07-11 20:48:03 +00:00
|
|
|
/**
|
2017-06-20 09:57:43 +00:00
|
|
|
* i915_reset_device - do process context error handling work
|
2016-06-03 13:02:17 +00:00
|
|
|
* @dev_priv: i915 device private
|
2009-07-11 20:48:03 +00:00
|
|
|
*
|
|
|
|
* Fire an error uevent so userspace can see that a hang or error
|
|
|
|
* was detected.
|
|
|
|
*/
|
2017-06-20 09:57:43 +00:00
|
|
|
static void i915_reset_device(struct drm_i915_private *dev_priv)
|
2009-07-11 20:48:03 +00:00
|
|
|
{
|
2016-07-05 09:40:23 +00:00
|
|
|
struct kobject *kobj = &dev_priv->drm.primary->kdev->kobj;
|
2013-07-19 16:16:42 +00:00
|
|
|
char *error_event[] = { I915_ERROR_UEVENT "=1", NULL };
|
|
|
|
char *reset_event[] = { I915_RESET_UEVENT "=1", NULL };
|
|
|
|
char *reset_done_event[] = { I915_ERROR_UEVENT "=0", NULL };
|
2017-06-22 10:56:25 +00:00
|
|
|
struct wedge_me w;
|
2009-07-11 20:48:03 +00:00
|
|
|
|
2016-05-06 14:40:21 +00:00
|
|
|
kobject_uevent_env(kobj, KOBJ_CHANGE, error_event);
|
2009-09-14 21:48:46 +00:00
|
|
|
|
2016-09-09 13:11:47 +00:00
|
|
|
DRM_DEBUG_DRIVER("resetting chip\n");
|
|
|
|
kobject_uevent_env(kobj, KOBJ_CHANGE, reset_event);
|
|
|
|
|
2017-06-22 10:56:25 +00:00
|
|
|
/* Use a watchdog to ensure that our reset completes */
|
|
|
|
i915_wedge_on_timeout(&w, dev_priv, 5*HZ) {
|
|
|
|
intel_prepare_reset(dev_priv);
|
2014-11-24 16:28:11 +00:00
|
|
|
|
2017-06-22 10:56:25 +00:00
|
|
|
/* Signal that locked waiters should reset the GPU */
|
|
|
|
set_bit(I915_RESET_HANDOFF, &dev_priv->gpu_error.flags);
|
|
|
|
wake_up_all(&dev_priv->gpu_error.wait_queue);
|
2017-03-16 17:13:02 +00:00
|
|
|
|
2017-06-22 10:56:25 +00:00
|
|
|
/* Wait for anyone holding the lock to wakeup, without
|
|
|
|
* blocking indefinitely on struct_mutex.
|
2016-09-09 13:11:52 +00:00
|
|
|
*/
|
2017-06-22 10:56:25 +00:00
|
|
|
do {
|
|
|
|
if (mutex_trylock(&dev_priv->drm.struct_mutex)) {
|
2017-07-21 12:32:37 +00:00
|
|
|
i915_reset(dev_priv, 0);
|
2017-06-22 10:56:25 +00:00
|
|
|
mutex_unlock(&dev_priv->drm.struct_mutex);
|
|
|
|
}
|
|
|
|
} while (wait_on_bit_timeout(&dev_priv->gpu_error.flags,
|
|
|
|
I915_RESET_HANDOFF,
|
|
|
|
TASK_UNINTERRUPTIBLE,
|
|
|
|
1));
|
2013-09-08 19:57:13 +00:00
|
|
|
|
2017-06-22 10:56:25 +00:00
|
|
|
intel_finish_reset(dev_priv);
|
|
|
|
}
|
2014-04-22 22:09:04 +00:00
|
|
|
|
2016-09-09 13:11:52 +00:00
|
|
|
if (!test_bit(I915_WEDGED, &dev_priv->gpu_error.flags))
|
2016-09-09 13:11:47 +00:00
|
|
|
kobject_uevent_env(kobj,
|
|
|
|
KOBJ_CHANGE, reset_done_event);
|
2009-07-11 20:48:03 +00:00
|
|
|
}
|
|
|
|
|
2016-10-19 12:52:03 +00:00
|
|
|
static void i915_clear_error_registers(struct drm_i915_private *dev_priv)
|
2009-07-11 20:48:03 +00:00
|
|
|
{
|
2016-10-19 12:52:03 +00:00
|
|
|
u32 eir;
|
2009-07-11 20:48:03 +00:00
|
|
|
|
2016-10-19 12:52:03 +00:00
|
|
|
if (!IS_GEN2(dev_priv))
|
|
|
|
I915_WRITE(PGTBL_ER, I915_READ(PGTBL_ER));
|
2009-07-11 20:48:03 +00:00
|
|
|
|
2016-10-19 12:52:03 +00:00
|
|
|
if (INTEL_GEN(dev_priv) < 4)
|
|
|
|
I915_WRITE(IPEIR, I915_READ(IPEIR));
|
|
|
|
else
|
|
|
|
I915_WRITE(IPEIR_I965, I915_READ(IPEIR_I965));
|
2009-07-11 20:48:03 +00:00
|
|
|
|
2016-10-19 12:52:03 +00:00
|
|
|
I915_WRITE(EIR, I915_READ(EIR));
|
2009-07-11 20:48:03 +00:00
|
|
|
eir = I915_READ(EIR);
|
|
|
|
if (eir) {
|
|
|
|
/*
|
|
|
|
* some errors might have become stuck,
|
|
|
|
* mask them.
|
|
|
|
*/
|
2016-10-19 12:52:03 +00:00
|
|
|
DRM_DEBUG_DRIVER("EIR stuck: 0x%08x, masking\n", eir);
|
2009-07-11 20:48:03 +00:00
|
|
|
I915_WRITE(EMR, I915_READ(EMR) | eir);
|
|
|
|
I915_WRITE(IIR, I915_RENDER_COMMAND_PARSER_ERROR_INTERRUPT);
|
|
|
|
}
|
2010-05-27 12:18:12 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
2015-01-28 15:03:14 +00:00
|
|
|
* i915_handle_error - handle a gpu error
|
2016-06-03 13:02:17 +00:00
|
|
|
* @dev_priv: i915 device private
|
2016-03-18 20:07:55 +00:00
|
|
|
* @engine_mask: mask representing engines that are hung
|
2017-01-12 04:18:08 +00:00
|
|
|
* @fmt: Error message format string
|
|
|
|
*
|
2015-10-08 07:57:49 +00:00
|
|
|
* Do some basic checking of register state at error time and
|
2010-05-27 12:18:12 +00:00
|
|
|
* dump it to the syslog. Also call i915_capture_error_state() to make
|
|
|
|
* sure we get a record and make it available in debugfs. Fire a uevent
|
|
|
|
* so userspace knows something bad happened (should trigger collection
|
|
|
|
* of a ring dump etc.).
|
|
|
|
*/
|
2016-05-06 14:40:21 +00:00
|
|
|
void i915_handle_error(struct drm_i915_private *dev_priv,
|
|
|
|
u32 engine_mask,
|
2014-02-25 15:11:26 +00:00
|
|
|
const char *fmt, ...)
|
2010-05-27 12:18:12 +00:00
|
|
|
{
|
drm/i915: Modify error handler for per engine hang recovery
This is a preparatory patch which modifies error handler to do per engine
hang recovery. The actual patch which implements this sequence follows
later in the series. The aim is to prepare existing recovery function to
adapt to this new function where applicable (which fails at this point
because core implementation is lacking) and continue recovery using legacy
full gpu reset.
A helper function is also added to query the availability of engine
reset. A subsequent patch will add the capability to query which type
of reset is present (engine -> full -> no-reset) via the get-param
ioctl.
It has been decided that the error events that are used to notify user of
reset will only be sent in case if full chip reset. In case of just
single (or multiple) engine resets, userspace won't be notified by these
events.
Note that this implementation of engine reset is for i915 directly
submitting to the ELSP, where the driver manages the hang detection,
recovery and resubmission. With GuC submission these tasks are shared
between driver and firmware; i915 will still responsible for detecting a
hang, and when it does it will have to request GuC to reset that Engine and
remind the firmware about the outstanding submissions. This will be
added in different patch.
v2: rebase, advertise engine reset availability in platform definition,
add note about GuC submission.
v3: s/*engine_reset*/*reset_engine*/. (Chris)
Handle reset as 2 level resets, by first going to engine only and fall
backing to full/chip reset as needed, i.e. reset_engine will need the
struct_mutex.
v4: Pass the engine mask to i915_reset. (Chris)
v5: Rebase, update selftests.
v6: Rebase, prepare for mutex-less reset engine.
v7: Pass reset_engine mask as a function parameter, and iterate over the
engine mask for reset_engine. (Chris)
v8: Use i915.reset >=2 in has_reset_engine; remove redundant reset
logging; add a reset-engine-in-progress flag to prevent concurrent
resets, and avoid dual purposing of reset-backoff. (Chris)
v9: Support reset of different engines in parallel (Chris)
v10: Handle reset-engine flag locking better (Chris)
v11: Squash in reporting of per-engine-reset availability.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@linux.intel.com>
Signed-off-by: Ian Lister <ian.lister@intel.com>
Signed-off-by: Tomas Elf <tomas.elf@intel.com>
Signed-off-by: Arun Siluvery <arun.siluvery@linux.intel.com>
Signed-off-by: Michel Thierry <michel.thierry@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/20170615201828.23144-4-michel.thierry@intel.com
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/20170620095751.13127-5-chris@chris-wilson.co.uk
2017-06-20 09:57:46 +00:00
|
|
|
struct intel_engine_cs *engine;
|
|
|
|
unsigned int tmp;
|
2014-02-25 15:11:26 +00:00
|
|
|
va_list args;
|
|
|
|
char error_msg[80];
|
2010-05-27 12:18:12 +00:00
|
|
|
|
2014-02-25 15:11:26 +00:00
|
|
|
va_start(args, fmt);
|
|
|
|
vscnprintf(error_msg, sizeof(error_msg), fmt, args);
|
|
|
|
va_end(args);
|
|
|
|
|
2017-03-14 17:18:40 +00:00
|
|
|
/*
|
|
|
|
* In most cases it's guaranteed that we get here with an RPM
|
|
|
|
* reference held, for example because there is a pending GPU
|
|
|
|
* request that won't finish until the reset is done. This
|
|
|
|
* isn't the case at least when we get here by doing a
|
|
|
|
* simulated reset via debugfs, so get an RPM reference.
|
|
|
|
*/
|
|
|
|
intel_runtime_pm_get(dev_priv);
|
|
|
|
|
2016-05-06 14:40:21 +00:00
|
|
|
i915_capture_error_state(dev_priv, engine_mask, error_msg);
|
2016-10-19 12:52:03 +00:00
|
|
|
i915_clear_error_registers(dev_priv);
|
2009-07-11 20:48:03 +00:00
|
|
|
|
drm/i915: Modify error handler for per engine hang recovery
This is a preparatory patch which modifies error handler to do per engine
hang recovery. The actual patch which implements this sequence follows
later in the series. The aim is to prepare existing recovery function to
adapt to this new function where applicable (which fails at this point
because core implementation is lacking) and continue recovery using legacy
full gpu reset.
A helper function is also added to query the availability of engine
reset. A subsequent patch will add the capability to query which type
of reset is present (engine -> full -> no-reset) via the get-param
ioctl.
It has been decided that the error events that are used to notify user of
reset will only be sent in case if full chip reset. In case of just
single (or multiple) engine resets, userspace won't be notified by these
events.
Note that this implementation of engine reset is for i915 directly
submitting to the ELSP, where the driver manages the hang detection,
recovery and resubmission. With GuC submission these tasks are shared
between driver and firmware; i915 will still responsible for detecting a
hang, and when it does it will have to request GuC to reset that Engine and
remind the firmware about the outstanding submissions. This will be
added in different patch.
v2: rebase, advertise engine reset availability in platform definition,
add note about GuC submission.
v3: s/*engine_reset*/*reset_engine*/. (Chris)
Handle reset as 2 level resets, by first going to engine only and fall
backing to full/chip reset as needed, i.e. reset_engine will need the
struct_mutex.
v4: Pass the engine mask to i915_reset. (Chris)
v5: Rebase, update selftests.
v6: Rebase, prepare for mutex-less reset engine.
v7: Pass reset_engine mask as a function parameter, and iterate over the
engine mask for reset_engine. (Chris)
v8: Use i915.reset >=2 in has_reset_engine; remove redundant reset
logging; add a reset-engine-in-progress flag to prevent concurrent
resets, and avoid dual purposing of reset-backoff. (Chris)
v9: Support reset of different engines in parallel (Chris)
v10: Handle reset-engine flag locking better (Chris)
v11: Squash in reporting of per-engine-reset availability.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@linux.intel.com>
Signed-off-by: Ian Lister <ian.lister@intel.com>
Signed-off-by: Tomas Elf <tomas.elf@intel.com>
Signed-off-by: Arun Siluvery <arun.siluvery@linux.intel.com>
Signed-off-by: Michel Thierry <michel.thierry@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/20170615201828.23144-4-michel.thierry@intel.com
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/20170620095751.13127-5-chris@chris-wilson.co.uk
2017-06-20 09:57:46 +00:00
|
|
|
/*
|
|
|
|
* Try engine reset when available. We fall back to full reset if
|
|
|
|
* single reset fails.
|
|
|
|
*/
|
|
|
|
if (intel_has_reset_engine(dev_priv)) {
|
|
|
|
for_each_engine_masked(engine, dev_priv, engine_mask, tmp) {
|
drm/i915: More surgically unbreak the modeset vs reset deadlock
There's no reason to entirely wedge the gpu, for the minimal deadlock
bugfix we only need to unbreak/decouple the atomic commit from the gpu
reset. The simplest way to fix that is by replacing the
unconditional fence wait a the top of commit_tail by a wait which
completes either when the fences are done (normal case, or when a
reset doesn't need to touch the display state). Or when the gpu reset
needs to force-unblock all pending modeset states.
The lesser source of deadlocks is when we try to pin a new framebuffer
and run into a stall. There's a bunch of places this can happen, like
eviction, changing the caching mode, acquiring a fence on older
platforms. And we can't just break the depency loop and keep going,
the only way would be to break out and restart. But the problem with
that approach is that we must stall for the reset to complete before
we grab any locks, and with the atomic infrastructure that's a bit
tricky. The only place is the ioctl code, and we don't want to insert
code into e.g. the BUSY ioctl. Hence for that problem just create a
critical section, and if any code is in there, wedge the GPU. For the
steady-state this should never be a problem.
Note that in both cases TDR itself keeps working, so from a userspace
pov this trickery isn't observable. Users themselvs might spot a short
glitch while the rendering is catching up again, but that's still
better than pre-TDR where we've thrown away all the rendering,
including innocent batches. Also, this fixes the regression TDR
introduced of making gpu resets deadlock-prone when we do need to
touch the display.
One thing I noticed is that gpu_error.flags seems to use both our own
wait-queue in gpu_error.wait_queue, and the generic wait_on_bit
facilities. Not entirely sure why this inconsistency exists, I just
picked one style.
A possible future avenue could be to insert the gpu reset in-between
ongoing modeset changes, which would avoid the momentary glitch. But
that's a lot more work to implement in the atomic commit machinery,
and given that we only need this for pre-g4x hw, of questionable
utility just for the sake of polishing gpu reset even more on those
old boxes. It might be useful for other features though.
v2: Rebase onto 4.13 with a s/wait_queue_t/struct wait_queue_entry/.
v3: Really emabarrassing fixup, I checked the wrong bit and broke the
unbreak/wakeup logic.
v4: Also handle deadlocks in pin_to_display.
v5: Review from Michel:
- Fixup the BUILD_BUG_ON
- Don't forget about the overlay
Cc: Michel Thierry <michel.thierry@intel.com>
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@intel.com>
Cc: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Cc: Tvrtko Ursulin <tvrtko.ursulin@intel.com> (v2)
Cc: Michel Thierry <michel.thierry@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20170808080828.23650-3-daniel.vetter@ffwll.ch
Reviewed-by: Michel Thierry <michel.thierry@intel.com>
2017-08-08 08:08:28 +00:00
|
|
|
BUILD_BUG_ON(I915_RESET_MODESET >= I915_RESET_ENGINE);
|
drm/i915: Modify error handler for per engine hang recovery
This is a preparatory patch which modifies error handler to do per engine
hang recovery. The actual patch which implements this sequence follows
later in the series. The aim is to prepare existing recovery function to
adapt to this new function where applicable (which fails at this point
because core implementation is lacking) and continue recovery using legacy
full gpu reset.
A helper function is also added to query the availability of engine
reset. A subsequent patch will add the capability to query which type
of reset is present (engine -> full -> no-reset) via the get-param
ioctl.
It has been decided that the error events that are used to notify user of
reset will only be sent in case if full chip reset. In case of just
single (or multiple) engine resets, userspace won't be notified by these
events.
Note that this implementation of engine reset is for i915 directly
submitting to the ELSP, where the driver manages the hang detection,
recovery and resubmission. With GuC submission these tasks are shared
between driver and firmware; i915 will still responsible for detecting a
hang, and when it does it will have to request GuC to reset that Engine and
remind the firmware about the outstanding submissions. This will be
added in different patch.
v2: rebase, advertise engine reset availability in platform definition,
add note about GuC submission.
v3: s/*engine_reset*/*reset_engine*/. (Chris)
Handle reset as 2 level resets, by first going to engine only and fall
backing to full/chip reset as needed, i.e. reset_engine will need the
struct_mutex.
v4: Pass the engine mask to i915_reset. (Chris)
v5: Rebase, update selftests.
v6: Rebase, prepare for mutex-less reset engine.
v7: Pass reset_engine mask as a function parameter, and iterate over the
engine mask for reset_engine. (Chris)
v8: Use i915.reset >=2 in has_reset_engine; remove redundant reset
logging; add a reset-engine-in-progress flag to prevent concurrent
resets, and avoid dual purposing of reset-backoff. (Chris)
v9: Support reset of different engines in parallel (Chris)
v10: Handle reset-engine flag locking better (Chris)
v11: Squash in reporting of per-engine-reset availability.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@linux.intel.com>
Signed-off-by: Ian Lister <ian.lister@intel.com>
Signed-off-by: Tomas Elf <tomas.elf@intel.com>
Signed-off-by: Arun Siluvery <arun.siluvery@linux.intel.com>
Signed-off-by: Michel Thierry <michel.thierry@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/20170615201828.23144-4-michel.thierry@intel.com
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/20170620095751.13127-5-chris@chris-wilson.co.uk
2017-06-20 09:57:46 +00:00
|
|
|
if (test_and_set_bit(I915_RESET_ENGINE + engine->id,
|
|
|
|
&dev_priv->gpu_error.flags))
|
|
|
|
continue;
|
|
|
|
|
2017-07-21 12:32:37 +00:00
|
|
|
if (i915_reset_engine(engine, 0) == 0)
|
drm/i915: Modify error handler for per engine hang recovery
This is a preparatory patch which modifies error handler to do per engine
hang recovery. The actual patch which implements this sequence follows
later in the series. The aim is to prepare existing recovery function to
adapt to this new function where applicable (which fails at this point
because core implementation is lacking) and continue recovery using legacy
full gpu reset.
A helper function is also added to query the availability of engine
reset. A subsequent patch will add the capability to query which type
of reset is present (engine -> full -> no-reset) via the get-param
ioctl.
It has been decided that the error events that are used to notify user of
reset will only be sent in case if full chip reset. In case of just
single (or multiple) engine resets, userspace won't be notified by these
events.
Note that this implementation of engine reset is for i915 directly
submitting to the ELSP, where the driver manages the hang detection,
recovery and resubmission. With GuC submission these tasks are shared
between driver and firmware; i915 will still responsible for detecting a
hang, and when it does it will have to request GuC to reset that Engine and
remind the firmware about the outstanding submissions. This will be
added in different patch.
v2: rebase, advertise engine reset availability in platform definition,
add note about GuC submission.
v3: s/*engine_reset*/*reset_engine*/. (Chris)
Handle reset as 2 level resets, by first going to engine only and fall
backing to full/chip reset as needed, i.e. reset_engine will need the
struct_mutex.
v4: Pass the engine mask to i915_reset. (Chris)
v5: Rebase, update selftests.
v6: Rebase, prepare for mutex-less reset engine.
v7: Pass reset_engine mask as a function parameter, and iterate over the
engine mask for reset_engine. (Chris)
v8: Use i915.reset >=2 in has_reset_engine; remove redundant reset
logging; add a reset-engine-in-progress flag to prevent concurrent
resets, and avoid dual purposing of reset-backoff. (Chris)
v9: Support reset of different engines in parallel (Chris)
v10: Handle reset-engine flag locking better (Chris)
v11: Squash in reporting of per-engine-reset availability.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@linux.intel.com>
Signed-off-by: Ian Lister <ian.lister@intel.com>
Signed-off-by: Tomas Elf <tomas.elf@intel.com>
Signed-off-by: Arun Siluvery <arun.siluvery@linux.intel.com>
Signed-off-by: Michel Thierry <michel.thierry@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/20170615201828.23144-4-michel.thierry@intel.com
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/20170620095751.13127-5-chris@chris-wilson.co.uk
2017-06-20 09:57:46 +00:00
|
|
|
engine_mask &= ~intel_engine_flag(engine);
|
|
|
|
|
|
|
|
clear_bit(I915_RESET_ENGINE + engine->id,
|
|
|
|
&dev_priv->gpu_error.flags);
|
|
|
|
wake_up_bit(&dev_priv->gpu_error.flags,
|
|
|
|
I915_RESET_ENGINE + engine->id);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-09-09 13:11:47 +00:00
|
|
|
if (!engine_mask)
|
2017-03-14 17:18:40 +00:00
|
|
|
goto out;
|
2009-09-14 21:48:47 +00:00
|
|
|
|
drm/i915: Modify error handler for per engine hang recovery
This is a preparatory patch which modifies error handler to do per engine
hang recovery. The actual patch which implements this sequence follows
later in the series. The aim is to prepare existing recovery function to
adapt to this new function where applicable (which fails at this point
because core implementation is lacking) and continue recovery using legacy
full gpu reset.
A helper function is also added to query the availability of engine
reset. A subsequent patch will add the capability to query which type
of reset is present (engine -> full -> no-reset) via the get-param
ioctl.
It has been decided that the error events that are used to notify user of
reset will only be sent in case if full chip reset. In case of just
single (or multiple) engine resets, userspace won't be notified by these
events.
Note that this implementation of engine reset is for i915 directly
submitting to the ELSP, where the driver manages the hang detection,
recovery and resubmission. With GuC submission these tasks are shared
between driver and firmware; i915 will still responsible for detecting a
hang, and when it does it will have to request GuC to reset that Engine and
remind the firmware about the outstanding submissions. This will be
added in different patch.
v2: rebase, advertise engine reset availability in platform definition,
add note about GuC submission.
v3: s/*engine_reset*/*reset_engine*/. (Chris)
Handle reset as 2 level resets, by first going to engine only and fall
backing to full/chip reset as needed, i.e. reset_engine will need the
struct_mutex.
v4: Pass the engine mask to i915_reset. (Chris)
v5: Rebase, update selftests.
v6: Rebase, prepare for mutex-less reset engine.
v7: Pass reset_engine mask as a function parameter, and iterate over the
engine mask for reset_engine. (Chris)
v8: Use i915.reset >=2 in has_reset_engine; remove redundant reset
logging; add a reset-engine-in-progress flag to prevent concurrent
resets, and avoid dual purposing of reset-backoff. (Chris)
v9: Support reset of different engines in parallel (Chris)
v10: Handle reset-engine flag locking better (Chris)
v11: Squash in reporting of per-engine-reset availability.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@linux.intel.com>
Signed-off-by: Ian Lister <ian.lister@intel.com>
Signed-off-by: Tomas Elf <tomas.elf@intel.com>
Signed-off-by: Arun Siluvery <arun.siluvery@linux.intel.com>
Signed-off-by: Michel Thierry <michel.thierry@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/20170615201828.23144-4-michel.thierry@intel.com
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/20170620095751.13127-5-chris@chris-wilson.co.uk
2017-06-20 09:57:46 +00:00
|
|
|
/* Full reset needs the mutex, stop any other user trying to do so. */
|
2017-06-20 09:57:43 +00:00
|
|
|
if (test_and_set_bit(I915_RESET_BACKOFF, &dev_priv->gpu_error.flags)) {
|
|
|
|
wait_event(dev_priv->gpu_error.reset_queue,
|
|
|
|
!test_bit(I915_RESET_BACKOFF,
|
|
|
|
&dev_priv->gpu_error.flags));
|
2017-03-14 17:18:40 +00:00
|
|
|
goto out;
|
2017-06-20 09:57:43 +00:00
|
|
|
}
|
|
|
|
|
drm/i915: Modify error handler for per engine hang recovery
This is a preparatory patch which modifies error handler to do per engine
hang recovery. The actual patch which implements this sequence follows
later in the series. The aim is to prepare existing recovery function to
adapt to this new function where applicable (which fails at this point
because core implementation is lacking) and continue recovery using legacy
full gpu reset.
A helper function is also added to query the availability of engine
reset. A subsequent patch will add the capability to query which type
of reset is present (engine -> full -> no-reset) via the get-param
ioctl.
It has been decided that the error events that are used to notify user of
reset will only be sent in case if full chip reset. In case of just
single (or multiple) engine resets, userspace won't be notified by these
events.
Note that this implementation of engine reset is for i915 directly
submitting to the ELSP, where the driver manages the hang detection,
recovery and resubmission. With GuC submission these tasks are shared
between driver and firmware; i915 will still responsible for detecting a
hang, and when it does it will have to request GuC to reset that Engine and
remind the firmware about the outstanding submissions. This will be
added in different patch.
v2: rebase, advertise engine reset availability in platform definition,
add note about GuC submission.
v3: s/*engine_reset*/*reset_engine*/. (Chris)
Handle reset as 2 level resets, by first going to engine only and fall
backing to full/chip reset as needed, i.e. reset_engine will need the
struct_mutex.
v4: Pass the engine mask to i915_reset. (Chris)
v5: Rebase, update selftests.
v6: Rebase, prepare for mutex-less reset engine.
v7: Pass reset_engine mask as a function parameter, and iterate over the
engine mask for reset_engine. (Chris)
v8: Use i915.reset >=2 in has_reset_engine; remove redundant reset
logging; add a reset-engine-in-progress flag to prevent concurrent
resets, and avoid dual purposing of reset-backoff. (Chris)
v9: Support reset of different engines in parallel (Chris)
v10: Handle reset-engine flag locking better (Chris)
v11: Squash in reporting of per-engine-reset availability.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@linux.intel.com>
Signed-off-by: Ian Lister <ian.lister@intel.com>
Signed-off-by: Tomas Elf <tomas.elf@intel.com>
Signed-off-by: Arun Siluvery <arun.siluvery@linux.intel.com>
Signed-off-by: Michel Thierry <michel.thierry@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/20170615201828.23144-4-michel.thierry@intel.com
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/20170620095751.13127-5-chris@chris-wilson.co.uk
2017-06-20 09:57:46 +00:00
|
|
|
/* Prevent any other reset-engine attempt. */
|
|
|
|
for_each_engine(engine, dev_priv, tmp) {
|
|
|
|
while (test_and_set_bit(I915_RESET_ENGINE + engine->id,
|
|
|
|
&dev_priv->gpu_error.flags))
|
|
|
|
wait_on_bit(&dev_priv->gpu_error.flags,
|
|
|
|
I915_RESET_ENGINE + engine->id,
|
|
|
|
TASK_UNINTERRUPTIBLE);
|
|
|
|
}
|
|
|
|
|
2017-06-20 09:57:43 +00:00
|
|
|
i915_reset_device(dev_priv);
|
2016-09-09 13:11:47 +00:00
|
|
|
|
drm/i915: Modify error handler for per engine hang recovery
This is a preparatory patch which modifies error handler to do per engine
hang recovery. The actual patch which implements this sequence follows
later in the series. The aim is to prepare existing recovery function to
adapt to this new function where applicable (which fails at this point
because core implementation is lacking) and continue recovery using legacy
full gpu reset.
A helper function is also added to query the availability of engine
reset. A subsequent patch will add the capability to query which type
of reset is present (engine -> full -> no-reset) via the get-param
ioctl.
It has been decided that the error events that are used to notify user of
reset will only be sent in case if full chip reset. In case of just
single (or multiple) engine resets, userspace won't be notified by these
events.
Note that this implementation of engine reset is for i915 directly
submitting to the ELSP, where the driver manages the hang detection,
recovery and resubmission. With GuC submission these tasks are shared
between driver and firmware; i915 will still responsible for detecting a
hang, and when it does it will have to request GuC to reset that Engine and
remind the firmware about the outstanding submissions. This will be
added in different patch.
v2: rebase, advertise engine reset availability in platform definition,
add note about GuC submission.
v3: s/*engine_reset*/*reset_engine*/. (Chris)
Handle reset as 2 level resets, by first going to engine only and fall
backing to full/chip reset as needed, i.e. reset_engine will need the
struct_mutex.
v4: Pass the engine mask to i915_reset. (Chris)
v5: Rebase, update selftests.
v6: Rebase, prepare for mutex-less reset engine.
v7: Pass reset_engine mask as a function parameter, and iterate over the
engine mask for reset_engine. (Chris)
v8: Use i915.reset >=2 in has_reset_engine; remove redundant reset
logging; add a reset-engine-in-progress flag to prevent concurrent
resets, and avoid dual purposing of reset-backoff. (Chris)
v9: Support reset of different engines in parallel (Chris)
v10: Handle reset-engine flag locking better (Chris)
v11: Squash in reporting of per-engine-reset availability.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@linux.intel.com>
Signed-off-by: Ian Lister <ian.lister@intel.com>
Signed-off-by: Tomas Elf <tomas.elf@intel.com>
Signed-off-by: Arun Siluvery <arun.siluvery@linux.intel.com>
Signed-off-by: Michel Thierry <michel.thierry@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/20170615201828.23144-4-michel.thierry@intel.com
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/20170620095751.13127-5-chris@chris-wilson.co.uk
2017-06-20 09:57:46 +00:00
|
|
|
for_each_engine(engine, dev_priv, tmp) {
|
|
|
|
clear_bit(I915_RESET_ENGINE + engine->id,
|
|
|
|
&dev_priv->gpu_error.flags);
|
|
|
|
}
|
|
|
|
|
2017-06-20 09:57:43 +00:00
|
|
|
clear_bit(I915_RESET_BACKOFF, &dev_priv->gpu_error.flags);
|
|
|
|
wake_up_all(&dev_priv->gpu_error.reset_queue);
|
2017-03-14 17:18:40 +00:00
|
|
|
|
|
|
|
out:
|
|
|
|
intel_runtime_pm_put(dev_priv);
|
2009-07-11 20:48:03 +00:00
|
|
|
}
|
|
|
|
|
2008-10-19 02:39:29 +00:00
|
|
|
/* Called from drm generic code, passed 'crtc' which
|
|
|
|
* we use as a pipe index
|
|
|
|
*/
|
2016-10-07 19:49:52 +00:00
|
|
|
static int i8xx_enable_vblank(struct drm_device *dev, unsigned int pipe)
|
2008-09-30 19:14:26 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2008-10-16 18:31:38 +00:00
|
|
|
unsigned long irqflags;
|
2009-01-08 18:42:15 +00:00
|
|
|
|
2010-12-04 11:30:53 +00:00
|
|
|
spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
|
2016-10-07 19:49:52 +00:00
|
|
|
i915_enable_pipestat(dev_priv, pipe, PIPE_VBLANK_INTERRUPT_STATUS);
|
2010-12-04 11:30:53 +00:00
|
|
|
spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
|
2011-02-05 10:08:21 +00:00
|
|
|
|
2008-09-30 19:14:26 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-10-07 19:49:52 +00:00
|
|
|
static int i965_enable_vblank(struct drm_device *dev, unsigned int pipe)
|
2011-04-07 20:58:17 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2011-04-07 20:58:17 +00:00
|
|
|
unsigned long irqflags;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
|
2016-10-07 19:49:52 +00:00
|
|
|
i915_enable_pipestat(dev_priv, pipe,
|
|
|
|
PIPE_START_VBLANK_INTERRUPT_STATUS);
|
2011-04-06 19:13:38 +00:00
|
|
|
spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-10-07 19:49:52 +00:00
|
|
|
static int ironlake_enable_vblank(struct drm_device *dev, unsigned int pipe)
|
2012-03-28 20:39:38 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2012-03-28 20:39:38 +00:00
|
|
|
unsigned long irqflags;
|
2016-10-14 08:17:22 +00:00
|
|
|
uint32_t bit = INTEL_GEN(dev_priv) >= 7 ?
|
2016-10-07 19:49:52 +00:00
|
|
|
DE_PIPE_VBLANK_IVB(pipe) : DE_PIPE_VBLANK(pipe);
|
2012-03-28 20:39:38 +00:00
|
|
|
|
|
|
|
spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
|
2016-10-07 19:49:52 +00:00
|
|
|
ilk_enable_display_irq(dev_priv, bit);
|
2012-03-28 20:39:38 +00:00
|
|
|
spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2015-09-24 16:35:31 +00:00
|
|
|
static int gen8_enable_vblank(struct drm_device *dev, unsigned int pipe)
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
unsigned long irqflags;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
|
2015-11-23 16:06:17 +00:00
|
|
|
bdw_enable_pipe_irq(dev_priv, pipe, GEN8_PIPE_VBLANK);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
|
2015-11-23 16:06:17 +00:00
|
|
|
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2008-10-19 02:39:29 +00:00
|
|
|
/* Called from drm generic code, passed 'crtc' which
|
|
|
|
* we use as a pipe index
|
|
|
|
*/
|
2016-10-07 19:49:52 +00:00
|
|
|
static void i8xx_disable_vblank(struct drm_device *dev, unsigned int pipe)
|
2008-09-30 19:14:26 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2008-10-16 18:31:38 +00:00
|
|
|
unsigned long irqflags;
|
2008-09-30 19:14:26 +00:00
|
|
|
|
2010-12-04 11:30:53 +00:00
|
|
|
spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
|
2016-10-07 19:49:52 +00:00
|
|
|
i915_disable_pipestat(dev_priv, pipe, PIPE_VBLANK_INTERRUPT_STATUS);
|
2011-04-07 20:58:17 +00:00
|
|
|
spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
|
|
|
|
}
|
|
|
|
|
2016-10-07 19:49:52 +00:00
|
|
|
static void i965_disable_vblank(struct drm_device *dev, unsigned int pipe)
|
2011-04-07 20:58:17 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2011-04-07 20:58:17 +00:00
|
|
|
unsigned long irqflags;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
|
2016-10-07 19:49:52 +00:00
|
|
|
i915_disable_pipestat(dev_priv, pipe,
|
|
|
|
PIPE_START_VBLANK_INTERRUPT_STATUS);
|
2011-04-06 19:13:38 +00:00
|
|
|
spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
|
|
|
|
}
|
|
|
|
|
2016-10-07 19:49:52 +00:00
|
|
|
static void ironlake_disable_vblank(struct drm_device *dev, unsigned int pipe)
|
2012-03-28 20:39:38 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2012-03-28 20:39:38 +00:00
|
|
|
unsigned long irqflags;
|
2016-10-14 08:17:22 +00:00
|
|
|
uint32_t bit = INTEL_GEN(dev_priv) >= 7 ?
|
2016-10-07 19:49:52 +00:00
|
|
|
DE_PIPE_VBLANK_IVB(pipe) : DE_PIPE_VBLANK(pipe);
|
2012-03-28 20:39:38 +00:00
|
|
|
|
|
|
|
spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
|
2016-10-07 19:49:52 +00:00
|
|
|
ilk_disable_display_irq(dev_priv, bit);
|
2012-03-28 20:39:38 +00:00
|
|
|
spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
|
|
|
|
}
|
|
|
|
|
2015-09-24 16:35:31 +00:00
|
|
|
static void gen8_disable_vblank(struct drm_device *dev, unsigned int pipe)
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
unsigned long irqflags;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
|
2015-11-23 16:06:17 +00:00
|
|
|
bdw_disable_pipe_irq(dev_priv, pipe, GEN8_PIPE_VBLANK);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
|
|
|
|
}
|
|
|
|
|
2016-11-16 08:55:38 +00:00
|
|
|
static void ibx_irq_reset(struct drm_i915_private *dev_priv)
|
2013-06-05 17:21:51 +00:00
|
|
|
{
|
2016-10-13 10:02:53 +00:00
|
|
|
if (HAS_PCH_NOP(dev_priv))
|
2013-06-05 17:21:51 +00:00
|
|
|
return;
|
|
|
|
|
2017-08-18 18:36:52 +00:00
|
|
|
GEN3_IRQ_RESET(SDE);
|
2014-04-01 18:37:17 +00:00
|
|
|
|
2016-10-13 10:02:53 +00:00
|
|
|
if (HAS_PCH_CPT(dev_priv) || HAS_PCH_LPT(dev_priv))
|
2014-04-01 18:37:17 +00:00
|
|
|
I915_WRITE(SERR_INT, 0xffffffff);
|
2014-04-01 18:37:22 +00:00
|
|
|
}
|
2014-04-01 18:37:17 +00:00
|
|
|
|
2014-04-01 18:37:22 +00:00
|
|
|
/*
|
|
|
|
* SDEIER is also touched by the interrupt handler to work around missed PCH
|
|
|
|
* interrupts. Hence we can't update it after the interrupt handler is enabled -
|
|
|
|
* instead we unconditionally enable all PCH interrupt sources here, but then
|
|
|
|
* only unmask them as needed with SDEIMR.
|
|
|
|
*
|
|
|
|
* This function needs to be called before interrupts are enabled.
|
|
|
|
*/
|
|
|
|
static void ibx_irq_pre_postinstall(struct drm_device *dev)
|
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2014-04-01 18:37:22 +00:00
|
|
|
|
2016-10-13 10:02:53 +00:00
|
|
|
if (HAS_PCH_NOP(dev_priv))
|
2014-04-01 18:37:22 +00:00
|
|
|
return;
|
|
|
|
|
|
|
|
WARN_ON(I915_READ(SDEIER) != 0);
|
2013-06-05 17:21:51 +00:00
|
|
|
I915_WRITE(SDEIER, 0xffffffff);
|
|
|
|
POSTING_READ(SDEIER);
|
|
|
|
}
|
|
|
|
|
2016-11-16 08:55:38 +00:00
|
|
|
static void gen5_gt_irq_reset(struct drm_i915_private *dev_priv)
|
2013-07-12 20:43:25 +00:00
|
|
|
{
|
2017-08-18 18:36:52 +00:00
|
|
|
GEN3_IRQ_RESET(GT);
|
2016-11-16 08:55:38 +00:00
|
|
|
if (INTEL_GEN(dev_priv) >= 6)
|
2017-08-18 18:36:52 +00:00
|
|
|
GEN3_IRQ_RESET(GEN6_PM);
|
2013-07-12 20:43:25 +00:00
|
|
|
}
|
|
|
|
|
2014-10-30 17:42:58 +00:00
|
|
|
static void vlv_display_irq_reset(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
2016-04-11 13:56:31 +00:00
|
|
|
if (IS_CHERRYVIEW(dev_priv))
|
|
|
|
I915_WRITE(DPINVGTT, DPINVGTT_STATUS_MASK_CHV);
|
|
|
|
else
|
|
|
|
I915_WRITE(DPINVGTT, DPINVGTT_STATUS_MASK);
|
|
|
|
|
2016-04-12 15:56:14 +00:00
|
|
|
i915_hotplug_interrupt_update_locked(dev_priv, 0xffffffff, 0);
|
2014-10-30 17:42:58 +00:00
|
|
|
I915_WRITE(PORT_HOTPLUG_STAT, I915_READ(PORT_HOTPLUG_STAT));
|
|
|
|
|
2017-08-18 18:36:51 +00:00
|
|
|
i9xx_pipestat_irq_reset(dev_priv);
|
2014-10-30 17:42:58 +00:00
|
|
|
|
2017-08-18 18:36:52 +00:00
|
|
|
GEN3_IRQ_RESET(VLV_);
|
2016-04-12 15:56:14 +00:00
|
|
|
dev_priv->irq_mask = ~0;
|
2014-10-30 17:42:58 +00:00
|
|
|
}
|
|
|
|
|
2016-04-12 15:56:44 +00:00
|
|
|
static void vlv_display_irq_postinstall(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
|
|
|
u32 pipestat_mask;
|
2016-04-11 13:56:28 +00:00
|
|
|
u32 enable_mask;
|
2016-04-12 15:56:44 +00:00
|
|
|
enum pipe pipe;
|
|
|
|
|
2017-08-18 18:36:50 +00:00
|
|
|
pipestat_mask = PIPE_CRC_DONE_INTERRUPT_STATUS;
|
2016-04-12 15:56:44 +00:00
|
|
|
|
|
|
|
i915_enable_pipestat(dev_priv, PIPE_A, PIPE_GMBUS_INTERRUPT_STATUS);
|
|
|
|
for_each_pipe(dev_priv, pipe)
|
|
|
|
i915_enable_pipestat(dev_priv, pipe, pipestat_mask);
|
|
|
|
|
2016-04-11 13:56:28 +00:00
|
|
|
enable_mask = I915_DISPLAY_PORT_INTERRUPT |
|
|
|
|
I915_DISPLAY_PIPE_A_EVENT_INTERRUPT |
|
2017-04-27 16:02:22 +00:00
|
|
|
I915_DISPLAY_PIPE_B_EVENT_INTERRUPT |
|
|
|
|
I915_LPE_PIPE_A_INTERRUPT |
|
|
|
|
I915_LPE_PIPE_B_INTERRUPT;
|
|
|
|
|
2016-04-12 15:56:44 +00:00
|
|
|
if (IS_CHERRYVIEW(dev_priv))
|
2017-04-27 16:02:22 +00:00
|
|
|
enable_mask |= I915_DISPLAY_PIPE_C_EVENT_INTERRUPT |
|
|
|
|
I915_LPE_PIPE_C_INTERRUPT;
|
2016-04-11 13:56:29 +00:00
|
|
|
|
|
|
|
WARN_ON(dev_priv->irq_mask != ~0);
|
|
|
|
|
2016-04-11 13:56:28 +00:00
|
|
|
dev_priv->irq_mask = ~enable_mask;
|
|
|
|
|
2017-08-18 18:36:52 +00:00
|
|
|
GEN3_IRQ_INIT(VLV_, dev_priv->irq_mask, enable_mask);
|
2016-04-12 15:56:44 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* drm_dma.h hooks
|
|
|
|
*/
|
|
|
|
static void ironlake_irq_reset(struct drm_device *dev)
|
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2016-04-12 15:56:44 +00:00
|
|
|
|
2017-08-18 18:37:03 +00:00
|
|
|
if (IS_GEN5(dev_priv))
|
|
|
|
I915_WRITE(HWSTAM, 0xffffffff);
|
2016-04-12 15:56:44 +00:00
|
|
|
|
2017-08-18 18:36:52 +00:00
|
|
|
GEN3_IRQ_RESET(DE);
|
2016-10-13 10:03:10 +00:00
|
|
|
if (IS_GEN7(dev_priv))
|
2016-04-12 15:56:44 +00:00
|
|
|
I915_WRITE(GEN7_ERR_INT, 0xffffffff);
|
|
|
|
|
2016-11-16 08:55:38 +00:00
|
|
|
gen5_gt_irq_reset(dev_priv);
|
2016-04-12 15:56:44 +00:00
|
|
|
|
2016-11-16 08:55:38 +00:00
|
|
|
ibx_irq_reset(dev_priv);
|
2016-04-12 15:56:44 +00:00
|
|
|
}
|
|
|
|
|
2017-08-18 18:37:04 +00:00
|
|
|
static void valleyview_irq_reset(struct drm_device *dev)
|
2012-03-28 20:39:38 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2012-03-28 20:39:38 +00:00
|
|
|
|
2016-04-13 18:19:48 +00:00
|
|
|
I915_WRITE(VLV_MASTER_IER, 0);
|
|
|
|
POSTING_READ(VLV_MASTER_IER);
|
|
|
|
|
2016-11-16 08:55:38 +00:00
|
|
|
gen5_gt_irq_reset(dev_priv);
|
2012-03-28 20:39:38 +00:00
|
|
|
|
2016-04-12 15:56:14 +00:00
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
2016-04-11 13:56:25 +00:00
|
|
|
if (dev_priv->display_irqs_enabled)
|
|
|
|
vlv_display_irq_reset(dev_priv);
|
2016-04-12 15:56:14 +00:00
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
2012-03-28 20:39:38 +00:00
|
|
|
}
|
|
|
|
|
2014-05-22 20:18:22 +00:00
|
|
|
static void gen8_gt_irq_reset(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
|
|
|
GEN8_IRQ_RESET_NDX(GT, 0);
|
|
|
|
GEN8_IRQ_RESET_NDX(GT, 1);
|
|
|
|
GEN8_IRQ_RESET_NDX(GT, 2);
|
|
|
|
GEN8_IRQ_RESET_NDX(GT, 3);
|
|
|
|
}
|
|
|
|
|
2014-04-01 18:37:26 +00:00
|
|
|
static void gen8_irq_reset(struct drm_device *dev)
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
int pipe;
|
|
|
|
|
|
|
|
I915_WRITE(GEN8_MASTER_IRQ, 0);
|
|
|
|
POSTING_READ(GEN8_MASTER_IRQ);
|
|
|
|
|
2014-05-22 20:18:22 +00:00
|
|
|
gen8_gt_irq_reset(dev_priv);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
|
2014-08-18 12:49:10 +00:00
|
|
|
for_each_pipe(dev_priv, pipe)
|
2014-09-30 08:56:39 +00:00
|
|
|
if (intel_display_power_is_enabled(dev_priv,
|
|
|
|
POWER_DOMAIN_PIPE(pipe)))
|
2014-07-04 14:50:29 +00:00
|
|
|
GEN8_IRQ_RESET_NDX(DE_PIPE, pipe);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
|
2017-08-18 18:36:52 +00:00
|
|
|
GEN3_IRQ_RESET(GEN8_DE_PORT_);
|
|
|
|
GEN3_IRQ_RESET(GEN8_DE_MISC_);
|
|
|
|
GEN3_IRQ_RESET(GEN8_PCU_);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
|
2016-10-13 10:02:53 +00:00
|
|
|
if (HAS_PCH_SPLIT(dev_priv))
|
2016-11-16 08:55:38 +00:00
|
|
|
ibx_irq_reset(dev_priv);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
}
|
2014-01-10 21:13:09 +00:00
|
|
|
|
2015-03-06 18:50:48 +00:00
|
|
|
void gen8_irq_power_well_post_enable(struct drm_i915_private *dev_priv,
|
2017-07-12 15:54:13 +00:00
|
|
|
u8 pipe_mask)
|
2014-07-04 14:50:31 +00:00
|
|
|
{
|
2014-10-07 21:02:52 +00:00
|
|
|
uint32_t extra_ier = GEN8_PIPE_VBLANK | GEN8_PIPE_FIFO_UNDERRUN;
|
2016-02-19 18:47:31 +00:00
|
|
|
enum pipe pipe;
|
2014-07-04 14:50:31 +00:00
|
|
|
|
2014-09-15 12:55:29 +00:00
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
2017-09-28 10:06:24 +00:00
|
|
|
|
|
|
|
if (!intel_irqs_enabled(dev_priv)) {
|
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2016-02-19 18:47:31 +00:00
|
|
|
for_each_pipe_masked(dev_priv, pipe, pipe_mask)
|
|
|
|
GEN8_IRQ_INIT_NDX(DE_PIPE, pipe,
|
|
|
|
dev_priv->de_irq_mask[pipe],
|
|
|
|
~dev_priv->de_irq_mask[pipe] | extra_ier);
|
2017-09-28 10:06:24 +00:00
|
|
|
|
2014-09-15 12:55:29 +00:00
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
2014-07-04 14:50:31 +00:00
|
|
|
}
|
|
|
|
|
2016-02-19 18:47:30 +00:00
|
|
|
void gen8_irq_power_well_pre_disable(struct drm_i915_private *dev_priv,
|
2017-07-12 15:54:13 +00:00
|
|
|
u8 pipe_mask)
|
2016-02-19 18:47:30 +00:00
|
|
|
{
|
2016-02-19 18:47:31 +00:00
|
|
|
enum pipe pipe;
|
|
|
|
|
2016-02-19 18:47:30 +00:00
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
2017-09-28 10:06:24 +00:00
|
|
|
|
|
|
|
if (!intel_irqs_enabled(dev_priv)) {
|
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2016-02-19 18:47:31 +00:00
|
|
|
for_each_pipe_masked(dev_priv, pipe, pipe_mask)
|
|
|
|
GEN8_IRQ_RESET_NDX(DE_PIPE, pipe);
|
2017-09-28 10:06:24 +00:00
|
|
|
|
2016-02-19 18:47:30 +00:00
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
|
|
|
|
|
|
|
/* make sure we're done processing display irqs */
|
2016-07-05 09:40:23 +00:00
|
|
|
synchronize_irq(dev_priv->drm.irq);
|
2016-02-19 18:47:30 +00:00
|
|
|
}
|
|
|
|
|
2017-08-18 18:37:04 +00:00
|
|
|
static void cherryview_irq_reset(struct drm_device *dev)
|
2014-04-09 17:40:52 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2014-04-09 17:40:52 +00:00
|
|
|
|
|
|
|
I915_WRITE(GEN8_MASTER_IRQ, 0);
|
|
|
|
POSTING_READ(GEN8_MASTER_IRQ);
|
|
|
|
|
2014-05-22 20:18:22 +00:00
|
|
|
gen8_gt_irq_reset(dev_priv);
|
2014-04-09 17:40:52 +00:00
|
|
|
|
2017-08-18 18:36:52 +00:00
|
|
|
GEN3_IRQ_RESET(GEN8_PCU_);
|
2014-04-09 17:40:52 +00:00
|
|
|
|
2016-04-12 15:56:14 +00:00
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
2016-04-11 13:56:25 +00:00
|
|
|
if (dev_priv->display_irqs_enabled)
|
|
|
|
vlv_display_irq_reset(dev_priv);
|
2016-04-12 15:56:14 +00:00
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
2014-04-09 17:40:52 +00:00
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static u32 intel_hpd_enabled_irqs(struct drm_i915_private *dev_priv,
|
2015-08-27 20:55:57 +00:00
|
|
|
const u32 hpd[HPD_NUM_PINS])
|
|
|
|
{
|
|
|
|
struct intel_encoder *encoder;
|
|
|
|
u32 enabled_irqs = 0;
|
|
|
|
|
2016-07-05 09:40:23 +00:00
|
|
|
for_each_intel_encoder(&dev_priv->drm, encoder)
|
2015-08-27 20:55:57 +00:00
|
|
|
if (dev_priv->hotplug.stats[encoder->hpd_pin].state == HPD_ENABLED)
|
|
|
|
enabled_irqs |= hpd[encoder->hpd_pin];
|
|
|
|
|
|
|
|
return enabled_irqs;
|
|
|
|
}
|
|
|
|
|
2017-01-27 09:39:21 +00:00
|
|
|
static void ibx_hpd_detection_setup(struct drm_i915_private *dev_priv)
|
2011-09-19 20:31:02 +00:00
|
|
|
{
|
2017-01-27 09:39:21 +00:00
|
|
|
u32 hotplug;
|
2013-03-27 14:55:01 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Enable digital hotplug on the PCH, and configure the DP short pulse
|
2015-08-27 20:56:02 +00:00
|
|
|
* duration to 2ms (which is the minimum in the Display Port spec).
|
|
|
|
* The pulse duration bits are reserved on LPT+.
|
2013-03-27 14:55:01 +00:00
|
|
|
*/
|
2011-09-19 20:31:02 +00:00
|
|
|
hotplug = I915_READ(PCH_PORT_HOTPLUG);
|
2017-01-27 09:39:21 +00:00
|
|
|
hotplug &= ~(PORTB_PULSE_DURATION_MASK |
|
|
|
|
PORTC_PULSE_DURATION_MASK |
|
|
|
|
PORTD_PULSE_DURATION_MASK);
|
2011-09-19 20:31:02 +00:00
|
|
|
hotplug |= PORTB_HOTPLUG_ENABLE | PORTB_PULSE_DURATION_2ms;
|
2017-01-27 09:39:21 +00:00
|
|
|
hotplug |= PORTC_HOTPLUG_ENABLE | PORTC_PULSE_DURATION_2ms;
|
|
|
|
hotplug |= PORTD_HOTPLUG_ENABLE | PORTD_PULSE_DURATION_2ms;
|
2015-08-27 20:56:05 +00:00
|
|
|
/*
|
|
|
|
* When CPU and PCH are on the same package, port A
|
|
|
|
* HPD must be enabled in both north and south.
|
|
|
|
*/
|
2016-05-06 13:48:28 +00:00
|
|
|
if (HAS_PCH_LPT_LP(dev_priv))
|
2015-08-27 20:56:05 +00:00
|
|
|
hotplug |= PORTA_HOTPLUG_ENABLE;
|
2011-09-19 20:31:02 +00:00
|
|
|
I915_WRITE(PCH_PORT_HOTPLUG, hotplug);
|
2015-08-27 20:56:02 +00:00
|
|
|
}
|
2015-08-17 07:55:50 +00:00
|
|
|
|
2017-01-27 09:39:21 +00:00
|
|
|
static void ibx_hpd_irq_setup(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
|
|
|
u32 hotplug_irqs, enabled_irqs;
|
|
|
|
|
|
|
|
if (HAS_PCH_IBX(dev_priv)) {
|
|
|
|
hotplug_irqs = SDE_HOTPLUG_MASK;
|
|
|
|
enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_ibx);
|
|
|
|
} else {
|
|
|
|
hotplug_irqs = SDE_HOTPLUG_MASK_CPT;
|
|
|
|
enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_cpt);
|
|
|
|
}
|
|
|
|
|
|
|
|
ibx_display_interrupt_update(dev_priv, hotplug_irqs, enabled_irqs);
|
|
|
|
|
|
|
|
ibx_hpd_detection_setup(dev_priv);
|
|
|
|
}
|
|
|
|
|
2017-01-27 09:39:18 +00:00
|
|
|
static void spt_hpd_detection_setup(struct drm_i915_private *dev_priv)
|
2015-08-27 20:56:02 +00:00
|
|
|
{
|
2017-09-19 21:57:03 +00:00
|
|
|
u32 val, hotplug;
|
|
|
|
|
|
|
|
/* Display WA #1179 WaHardHangonHotPlug: cnp */
|
|
|
|
if (HAS_PCH_CNP(dev_priv)) {
|
|
|
|
val = I915_READ(SOUTH_CHICKEN1);
|
|
|
|
val &= ~CHASSIS_CLK_REQ_DURATION_MASK;
|
|
|
|
val |= CHASSIS_CLK_REQ_DURATION(0xf);
|
|
|
|
I915_WRITE(SOUTH_CHICKEN1, val);
|
|
|
|
}
|
2015-08-27 20:56:02 +00:00
|
|
|
|
|
|
|
/* Enable digital hotplug on the PCH */
|
|
|
|
hotplug = I915_READ(PCH_PORT_HOTPLUG);
|
2017-01-27 09:39:18 +00:00
|
|
|
hotplug |= PORTA_HOTPLUG_ENABLE |
|
|
|
|
PORTB_HOTPLUG_ENABLE |
|
|
|
|
PORTC_HOTPLUG_ENABLE |
|
|
|
|
PORTD_HOTPLUG_ENABLE;
|
2015-08-27 20:56:02 +00:00
|
|
|
I915_WRITE(PCH_PORT_HOTPLUG, hotplug);
|
|
|
|
|
|
|
|
hotplug = I915_READ(PCH_PORT_HOTPLUG2);
|
|
|
|
hotplug |= PORTE_HOTPLUG_ENABLE;
|
|
|
|
I915_WRITE(PCH_PORT_HOTPLUG2, hotplug);
|
2011-09-19 20:31:02 +00:00
|
|
|
}
|
|
|
|
|
2017-01-27 09:39:18 +00:00
|
|
|
static void spt_hpd_irq_setup(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
|
|
|
u32 hotplug_irqs, enabled_irqs;
|
|
|
|
|
|
|
|
hotplug_irqs = SDE_HOTPLUG_MASK_SPT;
|
|
|
|
enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_spt);
|
|
|
|
|
|
|
|
ibx_display_interrupt_update(dev_priv, hotplug_irqs, enabled_irqs);
|
|
|
|
|
|
|
|
spt_hpd_detection_setup(dev_priv);
|
|
|
|
}
|
|
|
|
|
2017-01-27 09:39:21 +00:00
|
|
|
static void ilk_hpd_detection_setup(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
|
|
|
u32 hotplug;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Enable digital hotplug on the CPU, and configure the DP short pulse
|
|
|
|
* duration to 2ms (which is the minimum in the Display Port spec)
|
|
|
|
* The pulse duration bits are reserved on HSW+.
|
|
|
|
*/
|
|
|
|
hotplug = I915_READ(DIGITAL_PORT_HOTPLUG_CNTRL);
|
|
|
|
hotplug &= ~DIGITAL_PORTA_PULSE_DURATION_MASK;
|
|
|
|
hotplug |= DIGITAL_PORTA_HOTPLUG_ENABLE |
|
|
|
|
DIGITAL_PORTA_PULSE_DURATION_2ms;
|
|
|
|
I915_WRITE(DIGITAL_PORT_HOTPLUG_CNTRL, hotplug);
|
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void ilk_hpd_irq_setup(struct drm_i915_private *dev_priv)
|
2015-08-27 20:56:03 +00:00
|
|
|
{
|
2017-01-27 09:39:21 +00:00
|
|
|
u32 hotplug_irqs, enabled_irqs;
|
2015-08-27 20:56:03 +00:00
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
if (INTEL_GEN(dev_priv) >= 8) {
|
2015-08-27 20:56:06 +00:00
|
|
|
hotplug_irqs = GEN8_PORT_DP_A_HOTPLUG;
|
2016-05-06 13:48:28 +00:00
|
|
|
enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_bdw);
|
2015-08-27 20:56:06 +00:00
|
|
|
|
|
|
|
bdw_update_port_irq(dev_priv, hotplug_irqs, enabled_irqs);
|
2016-05-06 13:48:28 +00:00
|
|
|
} else if (INTEL_GEN(dev_priv) >= 7) {
|
2015-08-27 20:56:04 +00:00
|
|
|
hotplug_irqs = DE_DP_A_HOTPLUG_IVB;
|
2016-05-06 13:48:28 +00:00
|
|
|
enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_ivb);
|
2015-08-27 20:56:06 +00:00
|
|
|
|
|
|
|
ilk_update_display_irq(dev_priv, hotplug_irqs, enabled_irqs);
|
2015-08-27 20:56:04 +00:00
|
|
|
} else {
|
|
|
|
hotplug_irqs = DE_DP_A_HOTPLUG;
|
2016-05-06 13:48:28 +00:00
|
|
|
enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_ilk);
|
2015-08-27 20:56:03 +00:00
|
|
|
|
2015-08-27 20:56:06 +00:00
|
|
|
ilk_update_display_irq(dev_priv, hotplug_irqs, enabled_irqs);
|
|
|
|
}
|
2015-08-27 20:56:03 +00:00
|
|
|
|
2017-01-27 09:39:21 +00:00
|
|
|
ilk_hpd_detection_setup(dev_priv);
|
2015-08-27 20:56:03 +00:00
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
ibx_hpd_irq_setup(dev_priv);
|
2015-08-27 20:56:03 +00:00
|
|
|
}
|
|
|
|
|
2017-01-27 09:39:18 +00:00
|
|
|
static void __bxt_hpd_detection_setup(struct drm_i915_private *dev_priv,
|
|
|
|
u32 enabled_irqs)
|
2015-03-27 12:54:14 +00:00
|
|
|
{
|
2017-01-27 09:39:18 +00:00
|
|
|
u32 hotplug;
|
2015-03-27 12:54:14 +00:00
|
|
|
|
2015-08-27 20:56:11 +00:00
|
|
|
hotplug = I915_READ(PCH_PORT_HOTPLUG);
|
2017-01-27 09:39:18 +00:00
|
|
|
hotplug |= PORTA_HOTPLUG_ENABLE |
|
|
|
|
PORTB_HOTPLUG_ENABLE |
|
|
|
|
PORTC_HOTPLUG_ENABLE;
|
2016-03-31 10:41:47 +00:00
|
|
|
|
|
|
|
DRM_DEBUG_KMS("Invert bit setting: hp_ctl:%x hp_port:%x\n",
|
|
|
|
hotplug, enabled_irqs);
|
|
|
|
hotplug &= ~BXT_DDI_HPD_INVERT_MASK;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* For BXT invert bit has to be set based on AOB design
|
|
|
|
* for HPD detection logic, update it based on VBT fields.
|
|
|
|
*/
|
|
|
|
if ((enabled_irqs & BXT_DE_PORT_HP_DDIA) &&
|
|
|
|
intel_bios_is_port_hpd_inverted(dev_priv, PORT_A))
|
|
|
|
hotplug |= BXT_DDIA_HPD_INVERT;
|
|
|
|
if ((enabled_irqs & BXT_DE_PORT_HP_DDIB) &&
|
|
|
|
intel_bios_is_port_hpd_inverted(dev_priv, PORT_B))
|
|
|
|
hotplug |= BXT_DDIB_HPD_INVERT;
|
|
|
|
if ((enabled_irqs & BXT_DE_PORT_HP_DDIC) &&
|
|
|
|
intel_bios_is_port_hpd_inverted(dev_priv, PORT_C))
|
|
|
|
hotplug |= BXT_DDIC_HPD_INVERT;
|
|
|
|
|
2015-08-27 20:56:11 +00:00
|
|
|
I915_WRITE(PCH_PORT_HOTPLUG, hotplug);
|
2015-03-27 12:54:14 +00:00
|
|
|
}
|
|
|
|
|
2017-01-27 09:39:18 +00:00
|
|
|
static void bxt_hpd_detection_setup(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
|
|
|
__bxt_hpd_detection_setup(dev_priv, BXT_DE_PORT_HOTPLUG_MASK);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void bxt_hpd_irq_setup(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
|
|
|
u32 hotplug_irqs, enabled_irqs;
|
|
|
|
|
|
|
|
enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_bxt);
|
|
|
|
hotplug_irqs = BXT_DE_PORT_HOTPLUG_MASK;
|
|
|
|
|
|
|
|
bdw_update_port_irq(dev_priv, hotplug_irqs, enabled_irqs);
|
|
|
|
|
|
|
|
__bxt_hpd_detection_setup(dev_priv, enabled_irqs);
|
|
|
|
}
|
|
|
|
|
2013-02-08 19:35:15 +00:00
|
|
|
static void ibx_irq_postinstall(struct drm_device *dev)
|
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2013-03-27 14:55:01 +00:00
|
|
|
u32 mask;
|
2013-02-28 09:17:12 +00:00
|
|
|
|
2016-10-13 10:02:53 +00:00
|
|
|
if (HAS_PCH_NOP(dev_priv))
|
2013-05-29 19:43:05 +00:00
|
|
|
return;
|
|
|
|
|
2016-10-13 10:02:53 +00:00
|
|
|
if (HAS_PCH_IBX(dev_priv))
|
2014-03-07 19:34:46 +00:00
|
|
|
mask = SDE_GMBUS | SDE_AUX_MASK | SDE_POISON;
|
2017-09-09 00:42:55 +00:00
|
|
|
else if (HAS_PCH_CPT(dev_priv) || HAS_PCH_LPT(dev_priv))
|
2014-03-07 19:34:46 +00:00
|
|
|
mask = SDE_GMBUS_CPT | SDE_AUX_MASK_CPT;
|
2017-09-09 00:42:55 +00:00
|
|
|
else
|
|
|
|
mask = SDE_GMBUS_CPT;
|
drm/i915: report Gen5+ CPU and PCH FIFO underruns
In this commit we enable both CPU and PCH FIFO underrun reporting and
start reporting them. We follow a few rules:
- after we receive one of these errors, we mask the interrupt, so
we won't get an "interrupt storm" and we also won't flood dmesg;
- at each mode set we enable the interrupts again, so we'll see each
message at most once per mode set;
- in the specific places where we need to ignore the errors, we
completely mask the interrupts.
The downside of this patch is that since we're completely disabling
(masking) the interrupts instead of just not printing error messages,
we will mask more than just what we want on IVB/HSW CPU interrupts
(due to GEN7_ERR_INT) and on CPT/PPT/LPT PCHs (due to SERR_INT). So
when we decide to mask PCH FIFO underruns for pipe A on CPT, we'll
also be masking PCH FIFO underruns for pipe B, because both are
reported by SERR_INT, which has to be either completely enabled or
completely disabled (in othe words, there's no way to disable/enable
specific bits of GEN7_ERR_INT and SERR_INT).
V2: Rename some functions and variables, downgrade messages to
DRM_DEBUG_DRIVER and rebase.
Signed-off-by: Paulo Zanoni <paulo.r.zanoni@intel.com>
Reviewed-by: Imre Deak <imre.deak@intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2013-04-12 20:57:57 +00:00
|
|
|
|
2017-08-18 18:36:52 +00:00
|
|
|
gen3_assert_iir_is_zero(dev_priv, SDEIIR);
|
2013-02-08 19:35:15 +00:00
|
|
|
I915_WRITE(SDEIMR, ~mask);
|
2017-01-27 09:39:18 +00:00
|
|
|
|
|
|
|
if (HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv) ||
|
|
|
|
HAS_PCH_LPT(dev_priv))
|
2017-01-27 09:39:21 +00:00
|
|
|
ibx_hpd_detection_setup(dev_priv);
|
2017-01-27 09:39:18 +00:00
|
|
|
else
|
|
|
|
spt_hpd_detection_setup(dev_priv);
|
2013-02-08 19:35:15 +00:00
|
|
|
}
|
|
|
|
|
2013-07-12 20:43:26 +00:00
|
|
|
static void gen5_gt_irq_postinstall(struct drm_device *dev)
|
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2013-07-12 20:43:26 +00:00
|
|
|
u32 pm_irqs, gt_irqs;
|
|
|
|
|
|
|
|
pm_irqs = gt_irqs = 0;
|
|
|
|
|
|
|
|
dev_priv->gt_irq_mask = ~0;
|
2016-10-13 10:03:05 +00:00
|
|
|
if (HAS_L3_DPF(dev_priv)) {
|
2013-07-12 20:43:26 +00:00
|
|
|
/* L3 parity interrupt is always unmasked. */
|
2016-10-13 10:03:01 +00:00
|
|
|
dev_priv->gt_irq_mask = ~GT_PARITY_ERROR(dev_priv);
|
|
|
|
gt_irqs |= GT_PARITY_ERROR(dev_priv);
|
2013-07-12 20:43:26 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
gt_irqs |= GT_RENDER_USER_INTERRUPT;
|
2016-10-13 10:03:10 +00:00
|
|
|
if (IS_GEN5(dev_priv)) {
|
2016-07-01 16:23:21 +00:00
|
|
|
gt_irqs |= ILK_BSD_USER_INTERRUPT;
|
2013-07-12 20:43:26 +00:00
|
|
|
} else {
|
|
|
|
gt_irqs |= GT_BLT_USER_INTERRUPT | GT_BSD_USER_INTERRUPT;
|
|
|
|
}
|
|
|
|
|
2017-08-18 18:36:52 +00:00
|
|
|
GEN3_IRQ_INIT(GT, dev_priv->gt_irq_mask, gt_irqs);
|
2013-07-12 20:43:26 +00:00
|
|
|
|
2016-11-16 08:55:38 +00:00
|
|
|
if (INTEL_GEN(dev_priv) >= 6) {
|
2014-12-15 16:59:27 +00:00
|
|
|
/*
|
|
|
|
* RPS interrupts will get enabled/disabled on demand when RPS
|
|
|
|
* itself is enabled/disabled.
|
|
|
|
*/
|
2016-10-12 16:24:30 +00:00
|
|
|
if (HAS_VEBOX(dev_priv)) {
|
2013-07-12 20:43:26 +00:00
|
|
|
pm_irqs |= PM_VEBOX_USER_INTERRUPT;
|
2016-10-12 16:24:30 +00:00
|
|
|
dev_priv->pm_ier |= PM_VEBOX_USER_INTERRUPT;
|
|
|
|
}
|
2013-07-12 20:43:26 +00:00
|
|
|
|
2016-10-12 16:24:30 +00:00
|
|
|
dev_priv->pm_imr = 0xffffffff;
|
2017-08-18 18:36:52 +00:00
|
|
|
GEN3_IRQ_INIT(GEN6_PM, dev_priv->pm_imr, pm_irqs);
|
2013-07-12 20:43:26 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-06-28 20:00:41 +00:00
|
|
|
static int ironlake_irq_postinstall(struct drm_device *dev)
|
2009-06-08 06:40:19 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2013-07-12 23:01:56 +00:00
|
|
|
u32 display_mask, extra_mask;
|
|
|
|
|
2016-11-16 08:55:38 +00:00
|
|
|
if (INTEL_GEN(dev_priv) >= 7) {
|
2013-07-12 23:01:56 +00:00
|
|
|
display_mask = (DE_MASTER_IRQ_CONTROL | DE_GSE_IVB |
|
2017-08-18 18:36:50 +00:00
|
|
|
DE_PCH_EVENT_IVB | DE_AUX_CHANNEL_A_IVB);
|
2013-07-12 23:01:56 +00:00
|
|
|
extra_mask = (DE_PIPEC_VBLANK_IVB | DE_PIPEB_VBLANK_IVB |
|
2015-08-27 20:56:04 +00:00
|
|
|
DE_PIPEA_VBLANK_IVB | DE_ERR_INT_IVB |
|
|
|
|
DE_DP_A_HOTPLUG_IVB);
|
2013-07-12 23:01:56 +00:00
|
|
|
} else {
|
|
|
|
display_mask = (DE_MASTER_IRQ_CONTROL | DE_GSE | DE_PCH_EVENT |
|
2017-08-18 18:36:50 +00:00
|
|
|
DE_AUX_CHANNEL_A | DE_PIPEB_CRC_DONE |
|
|
|
|
DE_PIPEA_CRC_DONE | DE_POISON);
|
2015-08-27 20:56:03 +00:00
|
|
|
extra_mask = (DE_PIPEA_VBLANK | DE_PIPEB_VBLANK | DE_PCU_EVENT |
|
|
|
|
DE_PIPEB_FIFO_UNDERRUN | DE_PIPEA_FIFO_UNDERRUN |
|
|
|
|
DE_DP_A_HOTPLUG);
|
2013-07-12 23:01:56 +00:00
|
|
|
}
|
2009-06-08 06:40:19 +00:00
|
|
|
|
2010-12-04 11:30:53 +00:00
|
|
|
dev_priv->irq_mask = ~display_mask;
|
2009-06-08 06:40:19 +00:00
|
|
|
|
2014-04-01 18:37:22 +00:00
|
|
|
ibx_irq_pre_postinstall(dev);
|
|
|
|
|
2017-08-18 18:36:52 +00:00
|
|
|
GEN3_IRQ_INIT(DE, dev_priv->irq_mask, display_mask | extra_mask);
|
2009-06-08 06:40:19 +00:00
|
|
|
|
2013-07-12 20:43:26 +00:00
|
|
|
gen5_gt_irq_postinstall(dev);
|
2009-06-08 06:40:19 +00:00
|
|
|
|
2017-01-27 09:39:21 +00:00
|
|
|
ilk_hpd_detection_setup(dev_priv);
|
|
|
|
|
2013-02-08 19:35:15 +00:00
|
|
|
ibx_irq_postinstall(dev);
|
2011-09-19 20:31:02 +00:00
|
|
|
|
2016-10-13 10:02:58 +00:00
|
|
|
if (IS_IRONLAKE_M(dev_priv)) {
|
2013-06-27 11:44:59 +00:00
|
|
|
/* Enable PCU event interrupts
|
|
|
|
*
|
|
|
|
* spinlocking not required here for correctness since interrupt
|
2013-06-27 11:44:58 +00:00
|
|
|
* setup is guaranteed to run in single-threaded context. But we
|
|
|
|
* need it to make the assert_spin_locked happy. */
|
2014-09-15 12:55:27 +00:00
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
2015-11-23 16:06:16 +00:00
|
|
|
ilk_enable_display_irq(dev_priv, DE_PCU_EVENT);
|
2014-09-15 12:55:27 +00:00
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
2010-01-29 19:27:07 +00:00
|
|
|
}
|
|
|
|
|
2009-06-08 06:40:19 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2014-03-04 17:23:07 +00:00
|
|
|
void valleyview_enable_display_irqs(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
2017-03-02 13:28:01 +00:00
|
|
|
lockdep_assert_held(&dev_priv->irq_lock);
|
2014-03-04 17:23:07 +00:00
|
|
|
|
|
|
|
if (dev_priv->display_irqs_enabled)
|
|
|
|
return;
|
|
|
|
|
|
|
|
dev_priv->display_irqs_enabled = true;
|
|
|
|
|
2016-04-11 13:56:27 +00:00
|
|
|
if (intel_irqs_enabled(dev_priv)) {
|
|
|
|
vlv_display_irq_reset(dev_priv);
|
2016-04-12 15:56:14 +00:00
|
|
|
vlv_display_irq_postinstall(dev_priv);
|
2016-04-11 13:56:27 +00:00
|
|
|
}
|
2014-03-04 17:23:07 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
void valleyview_disable_display_irqs(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
2017-03-02 13:28:01 +00:00
|
|
|
lockdep_assert_held(&dev_priv->irq_lock);
|
2014-03-04 17:23:07 +00:00
|
|
|
|
|
|
|
if (!dev_priv->display_irqs_enabled)
|
|
|
|
return;
|
|
|
|
|
|
|
|
dev_priv->display_irqs_enabled = false;
|
|
|
|
|
2014-09-08 12:21:09 +00:00
|
|
|
if (intel_irqs_enabled(dev_priv))
|
2016-04-12 15:56:14 +00:00
|
|
|
vlv_display_irq_reset(dev_priv);
|
2014-03-04 17:23:07 +00:00
|
|
|
}
|
|
|
|
|
2014-10-30 17:43:00 +00:00
|
|
|
|
|
|
|
static int valleyview_irq_postinstall(struct drm_device *dev)
|
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2014-10-30 17:43:00 +00:00
|
|
|
|
2013-07-12 20:43:26 +00:00
|
|
|
gen5_gt_irq_postinstall(dev);
|
2012-03-28 20:39:38 +00:00
|
|
|
|
2016-04-12 15:56:14 +00:00
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
2016-04-11 13:56:25 +00:00
|
|
|
if (dev_priv->display_irqs_enabled)
|
|
|
|
vlv_display_irq_postinstall(dev_priv);
|
2016-04-12 15:56:14 +00:00
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
|
|
|
|
2012-03-28 20:39:38 +00:00
|
|
|
I915_WRITE(VLV_MASTER_IER, MASTER_INTERRUPT_ENABLE);
|
2016-04-13 18:19:48 +00:00
|
|
|
POSTING_READ(VLV_MASTER_IER);
|
2012-12-11 13:05:07 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
static void gen8_gt_irq_postinstall(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
|
|
|
/* These are interrupts we'll toggle with the ring mask register */
|
|
|
|
uint32_t gt_interrupts[] = {
|
|
|
|
GT_RENDER_USER_INTERRUPT << GEN8_RCS_IRQ_SHIFT |
|
2014-07-24 16:04:31 +00:00
|
|
|
GT_CONTEXT_SWITCH_INTERRUPT << GEN8_RCS_IRQ_SHIFT |
|
|
|
|
GT_RENDER_USER_INTERRUPT << GEN8_BCS_IRQ_SHIFT |
|
|
|
|
GT_CONTEXT_SWITCH_INTERRUPT << GEN8_BCS_IRQ_SHIFT,
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
GT_RENDER_USER_INTERRUPT << GEN8_VCS1_IRQ_SHIFT |
|
2014-07-24 16:04:31 +00:00
|
|
|
GT_CONTEXT_SWITCH_INTERRUPT << GEN8_VCS1_IRQ_SHIFT |
|
|
|
|
GT_RENDER_USER_INTERRUPT << GEN8_VCS2_IRQ_SHIFT |
|
|
|
|
GT_CONTEXT_SWITCH_INTERRUPT << GEN8_VCS2_IRQ_SHIFT,
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
0,
|
2014-07-24 16:04:31 +00:00
|
|
|
GT_RENDER_USER_INTERRUPT << GEN8_VECS_IRQ_SHIFT |
|
|
|
|
GT_CONTEXT_SWITCH_INTERRUPT << GEN8_VECS_IRQ_SHIFT
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
};
|
|
|
|
|
2016-04-19 15:46:08 +00:00
|
|
|
if (HAS_L3_DPF(dev_priv))
|
|
|
|
gt_interrupts[0] |= GT_RENDER_L3_PARITY_ERROR_INTERRUPT;
|
|
|
|
|
2016-10-12 16:24:30 +00:00
|
|
|
dev_priv->pm_ier = 0x0;
|
|
|
|
dev_priv->pm_imr = ~dev_priv->pm_ier;
|
2014-08-22 03:02:40 +00:00
|
|
|
GEN8_IRQ_INIT_NDX(GT, 0, ~gt_interrupts[0], gt_interrupts[0]);
|
|
|
|
GEN8_IRQ_INIT_NDX(GT, 1, ~gt_interrupts[1], gt_interrupts[1]);
|
2014-12-15 16:59:27 +00:00
|
|
|
/*
|
|
|
|
* RPS interrupts will get enabled/disabled on demand when RPS itself
|
2016-10-12 16:24:31 +00:00
|
|
|
* is enabled/disabled. Same wil be the case for GuC interrupts.
|
2014-12-15 16:59:27 +00:00
|
|
|
*/
|
2016-10-12 16:24:30 +00:00
|
|
|
GEN8_IRQ_INIT_NDX(GT, 2, dev_priv->pm_imr, dev_priv->pm_ier);
|
2014-08-22 03:02:40 +00:00
|
|
|
GEN8_IRQ_INIT_NDX(GT, 3, ~gt_interrupts[3], gt_interrupts[3]);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void gen8_de_irq_postinstall(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
2014-03-20 20:45:01 +00:00
|
|
|
uint32_t de_pipe_masked = GEN8_PIPE_CDCLK_CRC_DONE;
|
|
|
|
uint32_t de_pipe_enables;
|
2015-08-27 20:56:06 +00:00
|
|
|
u32 de_port_masked = GEN8_AUX_CHANNEL_A;
|
|
|
|
u32 de_port_enables;
|
2016-05-19 09:14:43 +00:00
|
|
|
u32 de_misc_masked = GEN8_DE_MISC_GSE;
|
2015-08-27 20:56:06 +00:00
|
|
|
enum pipe pipe;
|
2014-03-20 20:45:01 +00:00
|
|
|
|
2017-07-18 18:28:00 +00:00
|
|
|
if (INTEL_GEN(dev_priv) >= 9) {
|
2017-08-18 18:36:50 +00:00
|
|
|
de_pipe_masked |= GEN9_DE_PIPE_IRQ_FAULT_ERRORS;
|
2015-08-27 20:56:06 +00:00
|
|
|
de_port_masked |= GEN9_AUX_CHANNEL_B | GEN9_AUX_CHANNEL_C |
|
|
|
|
GEN9_AUX_CHANNEL_D;
|
2016-12-02 08:23:49 +00:00
|
|
|
if (IS_GEN9_LP(dev_priv))
|
2015-08-27 20:56:06 +00:00
|
|
|
de_port_masked |= BXT_DE_PORT_GMBUS;
|
|
|
|
} else {
|
2017-08-18 18:36:50 +00:00
|
|
|
de_pipe_masked |= GEN8_DE_PIPE_IRQ_FAULT_ERRORS;
|
2015-08-27 20:56:06 +00:00
|
|
|
}
|
2014-03-20 20:45:01 +00:00
|
|
|
|
|
|
|
de_pipe_enables = de_pipe_masked | GEN8_PIPE_VBLANK |
|
|
|
|
GEN8_PIPE_FIFO_UNDERRUN;
|
|
|
|
|
2015-08-27 20:56:06 +00:00
|
|
|
de_port_enables = de_port_masked;
|
2016-12-02 08:23:49 +00:00
|
|
|
if (IS_GEN9_LP(dev_priv))
|
2015-08-27 20:56:11 +00:00
|
|
|
de_port_enables |= BXT_DE_PORT_HOTPLUG_MASK;
|
|
|
|
else if (IS_BROADWELL(dev_priv))
|
2015-08-27 20:56:06 +00:00
|
|
|
de_port_enables |= GEN8_PORT_DP_A_HOTPLUG;
|
|
|
|
|
2013-11-07 14:31:52 +00:00
|
|
|
dev_priv->de_irq_mask[PIPE_A] = ~de_pipe_masked;
|
|
|
|
dev_priv->de_irq_mask[PIPE_B] = ~de_pipe_masked;
|
|
|
|
dev_priv->de_irq_mask[PIPE_C] = ~de_pipe_masked;
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
|
2014-08-18 12:49:10 +00:00
|
|
|
for_each_pipe(dev_priv, pipe)
|
2014-09-30 08:56:39 +00:00
|
|
|
if (intel_display_power_is_enabled(dev_priv,
|
2014-07-04 14:50:29 +00:00
|
|
|
POWER_DOMAIN_PIPE(pipe)))
|
|
|
|
GEN8_IRQ_INIT_NDX(DE_PIPE, pipe,
|
|
|
|
dev_priv->de_irq_mask[pipe],
|
|
|
|
de_pipe_enables);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
|
2017-08-18 18:36:52 +00:00
|
|
|
GEN3_IRQ_INIT(GEN8_DE_PORT_, ~de_port_masked, de_port_enables);
|
|
|
|
GEN3_IRQ_INIT(GEN8_DE_MISC_, ~de_misc_masked, de_misc_masked);
|
2017-01-27 09:39:18 +00:00
|
|
|
|
|
|
|
if (IS_GEN9_LP(dev_priv))
|
|
|
|
bxt_hpd_detection_setup(dev_priv);
|
2017-01-27 09:39:21 +00:00
|
|
|
else if (IS_BROADWELL(dev_priv))
|
|
|
|
ilk_hpd_detection_setup(dev_priv);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static int gen8_irq_postinstall(struct drm_device *dev)
|
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
|
2016-10-13 10:02:53 +00:00
|
|
|
if (HAS_PCH_SPLIT(dev_priv))
|
2014-08-22 12:10:42 +00:00
|
|
|
ibx_irq_pre_postinstall(dev);
|
2014-04-01 18:37:22 +00:00
|
|
|
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
gen8_gt_irq_postinstall(dev_priv);
|
|
|
|
gen8_de_irq_postinstall(dev_priv);
|
|
|
|
|
2016-10-13 10:02:53 +00:00
|
|
|
if (HAS_PCH_SPLIT(dev_priv))
|
2014-08-22 12:10:42 +00:00
|
|
|
ibx_irq_postinstall(dev);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
|
2016-04-13 18:19:47 +00:00
|
|
|
I915_WRITE(GEN8_MASTER_IRQ, GEN8_MASTER_IRQ_CONTROL);
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
POSTING_READ(GEN8_MASTER_IRQ);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2014-04-09 17:40:52 +00:00
|
|
|
static int cherryview_irq_postinstall(struct drm_device *dev)
|
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2014-04-09 17:40:52 +00:00
|
|
|
|
|
|
|
gen8_gt_irq_postinstall(dev_priv);
|
|
|
|
|
2016-04-12 15:56:14 +00:00
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
2016-04-11 13:56:25 +00:00
|
|
|
if (dev_priv->display_irqs_enabled)
|
|
|
|
vlv_display_irq_postinstall(dev_priv);
|
2016-04-12 15:56:14 +00:00
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
|
|
|
|
2016-04-13 18:19:47 +00:00
|
|
|
I915_WRITE(GEN8_MASTER_IRQ, GEN8_MASTER_IRQ_CONTROL);
|
2014-04-09 17:40:52 +00:00
|
|
|
POSTING_READ(GEN8_MASTER_IRQ);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2017-08-18 18:37:04 +00:00
|
|
|
static void i8xx_irq_reset(struct drm_device *dev)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2006-02-18 04:17:04 +00:00
|
|
|
|
2017-08-18 18:36:51 +00:00
|
|
|
i9xx_pipestat_irq_reset(dev_priv);
|
|
|
|
|
2017-08-18 18:37:03 +00:00
|
|
|
I915_WRITE16(HWSTAM, 0xffff);
|
|
|
|
|
2017-08-18 18:36:54 +00:00
|
|
|
GEN2_IRQ_RESET();
|
2012-04-22 20:13:57 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static int i8xx_irq_postinstall(struct drm_device *dev)
|
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2017-08-18 18:36:54 +00:00
|
|
|
u16 enable_mask;
|
2012-04-22 20:13:57 +00:00
|
|
|
|
2017-08-18 18:36:55 +00:00
|
|
|
I915_WRITE16(EMR, ~(I915_ERROR_PAGE_TABLE |
|
|
|
|
I915_ERROR_MEMORY_REFRESH));
|
2012-04-22 20:13:57 +00:00
|
|
|
|
|
|
|
/* Unmask the interrupts that we always want on. */
|
|
|
|
dev_priv->irq_mask =
|
|
|
|
~(I915_DISPLAY_PIPE_A_EVENT_INTERRUPT |
|
2017-08-18 18:36:50 +00:00
|
|
|
I915_DISPLAY_PIPE_B_EVENT_INTERRUPT);
|
2012-04-22 20:13:57 +00:00
|
|
|
|
2017-08-18 18:36:54 +00:00
|
|
|
enable_mask =
|
|
|
|
I915_DISPLAY_PIPE_A_EVENT_INTERRUPT |
|
|
|
|
I915_DISPLAY_PIPE_B_EVENT_INTERRUPT |
|
|
|
|
I915_USER_INTERRUPT;
|
|
|
|
|
|
|
|
GEN2_IRQ_INIT(, dev_priv->irq_mask, enable_mask);
|
2012-04-22 20:13:57 +00:00
|
|
|
|
2013-10-16 20:55:56 +00:00
|
|
|
/* Interrupt setup is already guaranteed to be single-threaded, this is
|
|
|
|
* just to make the assert_spin_locked check happy. */
|
2014-09-15 12:55:27 +00:00
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
2014-02-10 16:42:47 +00:00
|
|
|
i915_enable_pipestat(dev_priv, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS);
|
|
|
|
i915_enable_pipestat(dev_priv, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS);
|
2014-09-15 12:55:27 +00:00
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
2013-10-16 20:55:56 +00:00
|
|
|
|
2012-04-22 20:13:57 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2012-10-02 13:10:55 +00:00
|
|
|
static irqreturn_t i8xx_irq_handler(int irq, void *arg)
|
2012-04-22 20:13:57 +00:00
|
|
|
{
|
2014-05-12 17:17:55 +00:00
|
|
|
struct drm_device *dev = arg;
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2017-08-18 18:37:00 +00:00
|
|
|
irqreturn_t ret = IRQ_NONE;
|
2012-04-22 20:13:57 +00:00
|
|
|
|
2015-02-24 09:14:30 +00:00
|
|
|
if (!intel_irqs_enabled(dev_priv))
|
|
|
|
return IRQ_NONE;
|
|
|
|
|
2015-12-16 00:52:19 +00:00
|
|
|
/* IRQs are synced during runtime_suspend, we don't require a wakeref */
|
|
|
|
disable_rpm_wakeref_asserts(dev_priv);
|
|
|
|
|
2017-08-18 18:37:00 +00:00
|
|
|
do {
|
2017-08-18 18:36:59 +00:00
|
|
|
u32 pipe_stats[I915_MAX_PIPES] = {};
|
2017-08-18 18:37:00 +00:00
|
|
|
u16 iir;
|
2017-08-18 18:36:59 +00:00
|
|
|
|
2017-08-18 18:37:00 +00:00
|
|
|
iir = I915_READ16(IIR);
|
|
|
|
if (iir == 0)
|
|
|
|
break;
|
|
|
|
|
|
|
|
ret = IRQ_HANDLED;
|
2012-04-22 20:13:57 +00:00
|
|
|
|
2017-08-18 18:36:59 +00:00
|
|
|
/* Call regardless, as some status bits might not be
|
|
|
|
* signalled in iir */
|
|
|
|
i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats);
|
2012-04-22 20:13:57 +00:00
|
|
|
|
2017-07-20 17:57:51 +00:00
|
|
|
I915_WRITE16(IIR, iir);
|
2012-04-22 20:13:57 +00:00
|
|
|
|
|
|
|
if (iir & I915_USER_INTERRUPT)
|
drm/i915: Allocate intel_engine_cs structure only for the enabled engines
With the possibility of addition of many more number of rings in future,
the drm_i915_private structure could bloat as an array, of type
intel_engine_cs, is embedded inside it.
struct intel_engine_cs engine[I915_NUM_ENGINES];
Though this is still fine as generally there is only a single instance of
drm_i915_private structure used, but not all of the possible rings would be
enabled or active on most of the platforms. Some memory can be saved by
allocating intel_engine_cs structure only for the enabled/active engines.
Currently the engine/ring ID is kept static and dev_priv->engine[] is simply
indexed using the enums defined in intel_engine_id.
To save memory and continue using the static engine/ring IDs, 'engine' is
defined as an array of pointers.
struct intel_engine_cs *engine[I915_NUM_ENGINES];
dev_priv->engine[engine_ID] will be NULL for disabled engine instances.
There is a text size reduction of 928 bytes, from 1028200 to 1027272, for
i915.o file (but for i915.ko file text size remain same as 1193131 bytes).
v2:
- Remove the engine iterator field added in drm_i915_private structure,
instead pass a local iterator variable to the for_each_engine**
macros. (Chris)
- Do away with intel_engine_initialized() and instead directly use the
NULL pointer check on engine pointer. (Chris)
v3:
- Remove for_each_engine_id() macro, as the updated macro for_each_engine()
can be used in place of it. (Chris)
- Protect the access to Render engine Fault register with a NULL check, as
engine specific init is done later in Driver load sequence.
v4:
- Use !!dev_priv->engine[VCS] style for the engine check in getparam. (Chris)
- Kill the superfluous init_engine_lists().
v5:
- Cleanup the intel_engines_init() & intel_engines_setup(), with respect to
allocation of intel_engine_cs structure. (Chris)
v6:
- Rebase.
v7:
- Optimize the for_each_engine_masked() macro. (Chris)
- Change the type of 'iter' local variable to enum intel_engine_id. (Chris)
- Rebase.
v8: Rebase.
v9: Rebase.
v10:
- For index calculation use engine ID instead of pointer based arithmetic in
intel_engine_sync_index() as engine pointers are not contiguous now (Chris)
- For appropriateness, rename local enum variable 'iter' to 'id'. (Joonas)
- Use for_each_engine macro for cleanup in intel_engines_init() and remove
check for NULL engine pointer in cleanup() routines. (Joonas)
v11: Rebase.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/1476378888-7372-1-git-send-email-akash.goel@intel.com
2016-10-13 17:14:48 +00:00
|
|
|
notify_ring(dev_priv->engine[RCS]);
|
2012-04-22 20:13:57 +00:00
|
|
|
|
2017-08-18 18:37:00 +00:00
|
|
|
if (iir & I915_RENDER_COMMAND_PARSER_ERROR_INTERRUPT)
|
|
|
|
DRM_DEBUG("Command parser error, iir 0x%08x\n", iir);
|
2012-04-22 20:13:57 +00:00
|
|
|
|
2017-08-18 18:37:00 +00:00
|
|
|
i8xx_pipestat_irq_handler(dev_priv, iir, pipe_stats);
|
|
|
|
} while (0);
|
2015-12-16 00:52:19 +00:00
|
|
|
|
|
|
|
enable_rpm_wakeref_asserts(dev_priv);
|
2012-04-22 20:13:57 +00:00
|
|
|
|
2015-12-16 00:52:19 +00:00
|
|
|
return ret;
|
2012-04-22 20:13:57 +00:00
|
|
|
}
|
|
|
|
|
2017-08-18 18:37:04 +00:00
|
|
|
static void i915_irq_reset(struct drm_device *dev)
|
2012-04-24 21:59:44 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2016-11-07 09:29:20 +00:00
|
|
|
if (I915_HAS_HOTPLUG(dev_priv)) {
|
2015-09-23 14:15:27 +00:00
|
|
|
i915_hotplug_interrupt_update(dev_priv, 0xffffffff, 0);
|
2012-04-24 21:59:44 +00:00
|
|
|
I915_WRITE(PORT_HOTPLUG_STAT, I915_READ(PORT_HOTPLUG_STAT));
|
|
|
|
}
|
|
|
|
|
2017-08-18 18:36:51 +00:00
|
|
|
i9xx_pipestat_irq_reset(dev_priv);
|
|
|
|
|
2017-08-18 18:37:03 +00:00
|
|
|
I915_WRITE(HWSTAM, 0xffffffff);
|
2017-08-18 18:36:51 +00:00
|
|
|
|
2017-08-18 18:36:53 +00:00
|
|
|
GEN3_IRQ_RESET();
|
2012-04-24 21:59:44 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static int i915_irq_postinstall(struct drm_device *dev)
|
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2012-04-24 21:59:50 +00:00
|
|
|
u32 enable_mask;
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2017-08-18 18:36:55 +00:00
|
|
|
I915_WRITE(EMR, ~(I915_ERROR_PAGE_TABLE |
|
|
|
|
I915_ERROR_MEMORY_REFRESH));
|
2012-04-24 21:59:50 +00:00
|
|
|
|
|
|
|
/* Unmask the interrupts that we always want on. */
|
|
|
|
dev_priv->irq_mask =
|
|
|
|
~(I915_ASLE_INTERRUPT |
|
|
|
|
I915_DISPLAY_PIPE_A_EVENT_INTERRUPT |
|
2017-08-18 18:36:50 +00:00
|
|
|
I915_DISPLAY_PIPE_B_EVENT_INTERRUPT);
|
2012-04-24 21:59:50 +00:00
|
|
|
|
|
|
|
enable_mask =
|
|
|
|
I915_ASLE_INTERRUPT |
|
|
|
|
I915_DISPLAY_PIPE_A_EVENT_INTERRUPT |
|
|
|
|
I915_DISPLAY_PIPE_B_EVENT_INTERRUPT |
|
|
|
|
I915_USER_INTERRUPT;
|
|
|
|
|
2016-11-07 09:29:20 +00:00
|
|
|
if (I915_HAS_HOTPLUG(dev_priv)) {
|
2012-04-24 21:59:44 +00:00
|
|
|
/* Enable in IER... */
|
|
|
|
enable_mask |= I915_DISPLAY_PORT_INTERRUPT;
|
|
|
|
/* and unmask in IMR */
|
|
|
|
dev_priv->irq_mask &= ~I915_DISPLAY_PORT_INTERRUPT;
|
|
|
|
}
|
|
|
|
|
2017-08-18 18:36:53 +00:00
|
|
|
GEN3_IRQ_INIT(, dev_priv->irq_mask, enable_mask);
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2013-10-16 20:55:56 +00:00
|
|
|
/* Interrupt setup is already guaranteed to be single-threaded, this is
|
|
|
|
* just to make the assert_spin_locked check happy. */
|
2014-09-15 12:55:27 +00:00
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
2014-02-10 16:42:47 +00:00
|
|
|
i915_enable_pipestat(dev_priv, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS);
|
|
|
|
i915_enable_pipestat(dev_priv, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS);
|
2014-09-15 12:55:27 +00:00
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
2013-10-16 20:55:56 +00:00
|
|
|
|
2017-08-18 18:36:57 +00:00
|
|
|
i915_enable_asle_pipestat(dev_priv);
|
|
|
|
|
2012-12-11 13:05:07 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2012-10-02 13:10:55 +00:00
|
|
|
static irqreturn_t i915_irq_handler(int irq, void *arg)
|
2012-04-24 21:59:44 +00:00
|
|
|
{
|
2014-05-12 17:17:55 +00:00
|
|
|
struct drm_device *dev = arg;
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2017-08-18 18:37:00 +00:00
|
|
|
irqreturn_t ret = IRQ_NONE;
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2015-02-24 09:14:30 +00:00
|
|
|
if (!intel_irqs_enabled(dev_priv))
|
|
|
|
return IRQ_NONE;
|
|
|
|
|
2015-12-16 00:52:19 +00:00
|
|
|
/* IRQs are synced during runtime_suspend, we don't require a wakeref */
|
|
|
|
disable_rpm_wakeref_asserts(dev_priv);
|
|
|
|
|
2012-04-24 21:59:50 +00:00
|
|
|
do {
|
2017-08-18 18:36:59 +00:00
|
|
|
u32 pipe_stats[I915_MAX_PIPES] = {};
|
2017-08-18 18:37:00 +00:00
|
|
|
u32 hotplug_status = 0;
|
|
|
|
u32 iir;
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2017-08-18 18:37:00 +00:00
|
|
|
iir = I915_READ(IIR);
|
|
|
|
if (iir == 0)
|
|
|
|
break;
|
|
|
|
|
|
|
|
ret = IRQ_HANDLED;
|
|
|
|
|
|
|
|
if (I915_HAS_HOTPLUG(dev_priv) &&
|
|
|
|
iir & I915_DISPLAY_PORT_INTERRUPT)
|
|
|
|
hotplug_status = i9xx_hpd_irq_ack(dev_priv);
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2017-08-18 18:36:59 +00:00
|
|
|
/* Call regardless, as some status bits might not be
|
|
|
|
* signalled in iir */
|
|
|
|
i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats);
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2017-07-20 17:57:51 +00:00
|
|
|
I915_WRITE(IIR, iir);
|
2012-04-24 21:59:44 +00:00
|
|
|
|
|
|
|
if (iir & I915_USER_INTERRUPT)
|
drm/i915: Allocate intel_engine_cs structure only for the enabled engines
With the possibility of addition of many more number of rings in future,
the drm_i915_private structure could bloat as an array, of type
intel_engine_cs, is embedded inside it.
struct intel_engine_cs engine[I915_NUM_ENGINES];
Though this is still fine as generally there is only a single instance of
drm_i915_private structure used, but not all of the possible rings would be
enabled or active on most of the platforms. Some memory can be saved by
allocating intel_engine_cs structure only for the enabled/active engines.
Currently the engine/ring ID is kept static and dev_priv->engine[] is simply
indexed using the enums defined in intel_engine_id.
To save memory and continue using the static engine/ring IDs, 'engine' is
defined as an array of pointers.
struct intel_engine_cs *engine[I915_NUM_ENGINES];
dev_priv->engine[engine_ID] will be NULL for disabled engine instances.
There is a text size reduction of 928 bytes, from 1028200 to 1027272, for
i915.o file (but for i915.ko file text size remain same as 1193131 bytes).
v2:
- Remove the engine iterator field added in drm_i915_private structure,
instead pass a local iterator variable to the for_each_engine**
macros. (Chris)
- Do away with intel_engine_initialized() and instead directly use the
NULL pointer check on engine pointer. (Chris)
v3:
- Remove for_each_engine_id() macro, as the updated macro for_each_engine()
can be used in place of it. (Chris)
- Protect the access to Render engine Fault register with a NULL check, as
engine specific init is done later in Driver load sequence.
v4:
- Use !!dev_priv->engine[VCS] style for the engine check in getparam. (Chris)
- Kill the superfluous init_engine_lists().
v5:
- Cleanup the intel_engines_init() & intel_engines_setup(), with respect to
allocation of intel_engine_cs structure. (Chris)
v6:
- Rebase.
v7:
- Optimize the for_each_engine_masked() macro. (Chris)
- Change the type of 'iter' local variable to enum intel_engine_id. (Chris)
- Rebase.
v8: Rebase.
v9: Rebase.
v10:
- For index calculation use engine ID instead of pointer based arithmetic in
intel_engine_sync_index() as engine pointers are not contiguous now (Chris)
- For appropriateness, rename local enum variable 'iter' to 'id'. (Joonas)
- Use for_each_engine macro for cleanup in intel_engines_init() and remove
check for NULL engine pointer in cleanup() routines. (Joonas)
v11: Rebase.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/1476378888-7372-1-git-send-email-akash.goel@intel.com
2016-10-13 17:14:48 +00:00
|
|
|
notify_ring(dev_priv->engine[RCS]);
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2017-08-18 18:37:00 +00:00
|
|
|
if (iir & I915_RENDER_COMMAND_PARSER_ERROR_INTERRUPT)
|
|
|
|
DRM_DEBUG("Command parser error, iir 0x%08x\n", iir);
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2017-08-18 18:37:00 +00:00
|
|
|
if (hotplug_status)
|
|
|
|
i9xx_hpd_irq_handler(dev_priv, hotplug_status);
|
|
|
|
|
|
|
|
i915_pipestat_irq_handler(dev_priv, iir, pipe_stats);
|
|
|
|
} while (0);
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2015-12-16 00:52:19 +00:00
|
|
|
enable_rpm_wakeref_asserts(dev_priv);
|
|
|
|
|
2012-04-24 21:59:44 +00:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2017-08-18 18:37:04 +00:00
|
|
|
static void i965_irq_reset(struct drm_device *dev)
|
2012-04-24 21:59:44 +00:00
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2015-09-23 14:15:27 +00:00
|
|
|
i915_hotplug_interrupt_update(dev_priv, 0xffffffff, 0);
|
2012-05-11 17:01:31 +00:00
|
|
|
I915_WRITE(PORT_HOTPLUG_STAT, I915_READ(PORT_HOTPLUG_STAT));
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2017-08-18 18:36:51 +00:00
|
|
|
i9xx_pipestat_irq_reset(dev_priv);
|
|
|
|
|
2017-08-18 18:37:03 +00:00
|
|
|
I915_WRITE(HWSTAM, 0xffffffff);
|
2017-08-18 18:36:51 +00:00
|
|
|
|
2017-08-18 18:36:53 +00:00
|
|
|
GEN3_IRQ_RESET();
|
2012-04-24 21:59:44 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static int i965_irq_postinstall(struct drm_device *dev)
|
|
|
|
{
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2012-04-24 21:59:51 +00:00
|
|
|
u32 enable_mask;
|
2012-04-24 21:59:44 +00:00
|
|
|
u32 error_mask;
|
|
|
|
|
2017-08-18 18:36:55 +00:00
|
|
|
/*
|
|
|
|
* Enable some error detection, note the instruction error mask
|
|
|
|
* bit is reserved, so we leave it masked.
|
|
|
|
*/
|
|
|
|
if (IS_G4X(dev_priv)) {
|
|
|
|
error_mask = ~(GM45_ERROR_PAGE_TABLE |
|
|
|
|
GM45_ERROR_MEM_PRIV |
|
|
|
|
GM45_ERROR_CP_PRIV |
|
|
|
|
I915_ERROR_MEMORY_REFRESH);
|
|
|
|
} else {
|
|
|
|
error_mask = ~(I915_ERROR_PAGE_TABLE |
|
|
|
|
I915_ERROR_MEMORY_REFRESH);
|
|
|
|
}
|
|
|
|
I915_WRITE(EMR, error_mask);
|
|
|
|
|
2012-04-24 21:59:44 +00:00
|
|
|
/* Unmask the interrupts that we always want on. */
|
2017-08-18 18:36:57 +00:00
|
|
|
dev_priv->irq_mask =
|
|
|
|
~(I915_ASLE_INTERRUPT |
|
|
|
|
I915_DISPLAY_PORT_INTERRUPT |
|
|
|
|
I915_DISPLAY_PIPE_A_EVENT_INTERRUPT |
|
|
|
|
I915_DISPLAY_PIPE_B_EVENT_INTERRUPT |
|
|
|
|
I915_RENDER_COMMAND_PARSER_ERROR_INTERRUPT);
|
2012-04-24 21:59:51 +00:00
|
|
|
|
2017-08-18 18:36:57 +00:00
|
|
|
enable_mask =
|
|
|
|
I915_ASLE_INTERRUPT |
|
|
|
|
I915_DISPLAY_PORT_INTERRUPT |
|
|
|
|
I915_DISPLAY_PIPE_A_EVENT_INTERRUPT |
|
|
|
|
I915_DISPLAY_PIPE_B_EVENT_INTERRUPT |
|
|
|
|
I915_RENDER_COMMAND_PARSER_ERROR_INTERRUPT |
|
|
|
|
I915_USER_INTERRUPT;
|
2012-04-24 21:59:51 +00:00
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
if (IS_G4X(dev_priv))
|
2012-04-24 21:59:51 +00:00
|
|
|
enable_mask |= I915_BSD_USER_INTERRUPT;
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2017-08-18 18:36:57 +00:00
|
|
|
GEN3_IRQ_INIT(, dev_priv->irq_mask, enable_mask);
|
|
|
|
|
2013-06-27 15:52:10 +00:00
|
|
|
/* Interrupt setup is already guaranteed to be single-threaded, this is
|
|
|
|
* just to make the assert_spin_locked check happy. */
|
2014-09-15 12:55:27 +00:00
|
|
|
spin_lock_irq(&dev_priv->irq_lock);
|
2014-02-10 16:42:47 +00:00
|
|
|
i915_enable_pipestat(dev_priv, PIPE_A, PIPE_GMBUS_INTERRUPT_STATUS);
|
|
|
|
i915_enable_pipestat(dev_priv, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS);
|
|
|
|
i915_enable_pipestat(dev_priv, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS);
|
2014-09-15 12:55:27 +00:00
|
|
|
spin_unlock_irq(&dev_priv->irq_lock);
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
i915_enable_asle_pipestat(dev_priv);
|
2012-12-11 13:05:07 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-05-06 13:48:28 +00:00
|
|
|
static void i915_hpd_irq_setup(struct drm_i915_private *dev_priv)
|
2012-12-11 13:05:07 +00:00
|
|
|
{
|
|
|
|
u32 hotplug_en;
|
|
|
|
|
2017-03-02 13:28:01 +00:00
|
|
|
lockdep_assert_held(&dev_priv->irq_lock);
|
2013-06-27 15:52:15 +00:00
|
|
|
|
2015-01-09 12:21:13 +00:00
|
|
|
/* Note HDMI and DP share hotplug bits */
|
|
|
|
/* enable bits are the same for all generations */
|
2016-05-06 13:48:28 +00:00
|
|
|
hotplug_en = intel_hpd_enabled_irqs(dev_priv, hpd_mask_i915);
|
2015-01-09 12:21:13 +00:00
|
|
|
/* Programming the CRT detection parameters tends
|
|
|
|
to generate a spurious hotplug event about three
|
|
|
|
seconds later. So just do it once.
|
|
|
|
*/
|
2016-05-06 13:48:28 +00:00
|
|
|
if (IS_G4X(dev_priv))
|
2015-01-09 12:21:13 +00:00
|
|
|
hotplug_en |= CRT_HOTPLUG_ACTIVATION_PERIOD_64;
|
|
|
|
hotplug_en |= CRT_HOTPLUG_VOLTAGE_COMPARE_50;
|
|
|
|
|
|
|
|
/* Ignore TV since it's buggy */
|
2015-09-23 14:15:27 +00:00
|
|
|
i915_hotplug_interrupt_update_locked(dev_priv,
|
2015-10-21 14:22:43 +00:00
|
|
|
HOTPLUG_INT_EN_MASK |
|
|
|
|
CRT_HOTPLUG_VOLTAGE_COMPARE_MASK |
|
|
|
|
CRT_HOTPLUG_ACTIVATION_PERIOD_64,
|
|
|
|
hotplug_en);
|
2012-04-24 21:59:44 +00:00
|
|
|
}
|
|
|
|
|
2012-10-02 13:10:55 +00:00
|
|
|
static irqreturn_t i965_irq_handler(int irq, void *arg)
|
2012-04-24 21:59:44 +00:00
|
|
|
{
|
2014-05-12 17:17:55 +00:00
|
|
|
struct drm_device *dev = arg;
|
2016-07-04 10:34:36 +00:00
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2017-08-18 18:37:00 +00:00
|
|
|
irqreturn_t ret = IRQ_NONE;
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2015-02-24 09:14:30 +00:00
|
|
|
if (!intel_irqs_enabled(dev_priv))
|
|
|
|
return IRQ_NONE;
|
|
|
|
|
2015-12-16 00:52:19 +00:00
|
|
|
/* IRQs are synced during runtime_suspend, we don't require a wakeref */
|
|
|
|
disable_rpm_wakeref_asserts(dev_priv);
|
|
|
|
|
2017-08-18 18:37:00 +00:00
|
|
|
do {
|
2017-08-18 18:36:59 +00:00
|
|
|
u32 pipe_stats[I915_MAX_PIPES] = {};
|
2017-08-18 18:37:00 +00:00
|
|
|
u32 hotplug_status = 0;
|
|
|
|
u32 iir;
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2017-08-18 18:37:00 +00:00
|
|
|
iir = I915_READ(IIR);
|
|
|
|
if (iir == 0)
|
2012-04-24 21:59:44 +00:00
|
|
|
break;
|
|
|
|
|
|
|
|
ret = IRQ_HANDLED;
|
|
|
|
|
2017-08-18 18:37:00 +00:00
|
|
|
if (iir & I915_DISPLAY_PORT_INTERRUPT)
|
|
|
|
hotplug_status = i9xx_hpd_irq_ack(dev_priv);
|
|
|
|
|
|
|
|
/* Call regardless, as some status bits might not be
|
|
|
|
* signalled in iir */
|
|
|
|
i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats);
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2017-07-20 17:57:51 +00:00
|
|
|
I915_WRITE(IIR, iir);
|
2012-04-24 21:59:44 +00:00
|
|
|
|
|
|
|
if (iir & I915_USER_INTERRUPT)
|
drm/i915: Allocate intel_engine_cs structure only for the enabled engines
With the possibility of addition of many more number of rings in future,
the drm_i915_private structure could bloat as an array, of type
intel_engine_cs, is embedded inside it.
struct intel_engine_cs engine[I915_NUM_ENGINES];
Though this is still fine as generally there is only a single instance of
drm_i915_private structure used, but not all of the possible rings would be
enabled or active on most of the platforms. Some memory can be saved by
allocating intel_engine_cs structure only for the enabled/active engines.
Currently the engine/ring ID is kept static and dev_priv->engine[] is simply
indexed using the enums defined in intel_engine_id.
To save memory and continue using the static engine/ring IDs, 'engine' is
defined as an array of pointers.
struct intel_engine_cs *engine[I915_NUM_ENGINES];
dev_priv->engine[engine_ID] will be NULL for disabled engine instances.
There is a text size reduction of 928 bytes, from 1028200 to 1027272, for
i915.o file (but for i915.ko file text size remain same as 1193131 bytes).
v2:
- Remove the engine iterator field added in drm_i915_private structure,
instead pass a local iterator variable to the for_each_engine**
macros. (Chris)
- Do away with intel_engine_initialized() and instead directly use the
NULL pointer check on engine pointer. (Chris)
v3:
- Remove for_each_engine_id() macro, as the updated macro for_each_engine()
can be used in place of it. (Chris)
- Protect the access to Render engine Fault register with a NULL check, as
engine specific init is done later in Driver load sequence.
v4:
- Use !!dev_priv->engine[VCS] style for the engine check in getparam. (Chris)
- Kill the superfluous init_engine_lists().
v5:
- Cleanup the intel_engines_init() & intel_engines_setup(), with respect to
allocation of intel_engine_cs structure. (Chris)
v6:
- Rebase.
v7:
- Optimize the for_each_engine_masked() macro. (Chris)
- Change the type of 'iter' local variable to enum intel_engine_id. (Chris)
- Rebase.
v8: Rebase.
v9: Rebase.
v10:
- For index calculation use engine ID instead of pointer based arithmetic in
intel_engine_sync_index() as engine pointers are not contiguous now (Chris)
- For appropriateness, rename local enum variable 'iter' to 'id'. (Joonas)
- Use for_each_engine macro for cleanup in intel_engines_init() and remove
check for NULL engine pointer in cleanup() routines. (Joonas)
v11: Rebase.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/1476378888-7372-1-git-send-email-akash.goel@intel.com
2016-10-13 17:14:48 +00:00
|
|
|
notify_ring(dev_priv->engine[RCS]);
|
2017-08-18 18:37:00 +00:00
|
|
|
|
2012-04-24 21:59:44 +00:00
|
|
|
if (iir & I915_BSD_USER_INTERRUPT)
|
drm/i915: Allocate intel_engine_cs structure only for the enabled engines
With the possibility of addition of many more number of rings in future,
the drm_i915_private structure could bloat as an array, of type
intel_engine_cs, is embedded inside it.
struct intel_engine_cs engine[I915_NUM_ENGINES];
Though this is still fine as generally there is only a single instance of
drm_i915_private structure used, but not all of the possible rings would be
enabled or active on most of the platforms. Some memory can be saved by
allocating intel_engine_cs structure only for the enabled/active engines.
Currently the engine/ring ID is kept static and dev_priv->engine[] is simply
indexed using the enums defined in intel_engine_id.
To save memory and continue using the static engine/ring IDs, 'engine' is
defined as an array of pointers.
struct intel_engine_cs *engine[I915_NUM_ENGINES];
dev_priv->engine[engine_ID] will be NULL for disabled engine instances.
There is a text size reduction of 928 bytes, from 1028200 to 1027272, for
i915.o file (but for i915.ko file text size remain same as 1193131 bytes).
v2:
- Remove the engine iterator field added in drm_i915_private structure,
instead pass a local iterator variable to the for_each_engine**
macros. (Chris)
- Do away with intel_engine_initialized() and instead directly use the
NULL pointer check on engine pointer. (Chris)
v3:
- Remove for_each_engine_id() macro, as the updated macro for_each_engine()
can be used in place of it. (Chris)
- Protect the access to Render engine Fault register with a NULL check, as
engine specific init is done later in Driver load sequence.
v4:
- Use !!dev_priv->engine[VCS] style for the engine check in getparam. (Chris)
- Kill the superfluous init_engine_lists().
v5:
- Cleanup the intel_engines_init() & intel_engines_setup(), with respect to
allocation of intel_engine_cs structure. (Chris)
v6:
- Rebase.
v7:
- Optimize the for_each_engine_masked() macro. (Chris)
- Change the type of 'iter' local variable to enum intel_engine_id. (Chris)
- Rebase.
v8: Rebase.
v9: Rebase.
v10:
- For index calculation use engine ID instead of pointer based arithmetic in
intel_engine_sync_index() as engine pointers are not contiguous now (Chris)
- For appropriateness, rename local enum variable 'iter' to 'id'. (Joonas)
- Use for_each_engine macro for cleanup in intel_engines_init() and remove
check for NULL engine pointer in cleanup() routines. (Joonas)
v11: Rebase.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/1476378888-7372-1-git-send-email-akash.goel@intel.com
2016-10-13 17:14:48 +00:00
|
|
|
notify_ring(dev_priv->engine[VCS]);
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2017-08-18 18:37:00 +00:00
|
|
|
if (iir & I915_RENDER_COMMAND_PARSER_ERROR_INTERRUPT)
|
|
|
|
DRM_DEBUG("Command parser error, iir 0x%08x\n", iir);
|
2012-12-01 12:53:44 +00:00
|
|
|
|
2017-08-18 18:37:00 +00:00
|
|
|
if (hotplug_status)
|
|
|
|
i9xx_hpd_irq_handler(dev_priv, hotplug_status);
|
|
|
|
|
|
|
|
i965_pipestat_irq_handler(dev_priv, iir, pipe_stats);
|
|
|
|
} while (0);
|
2012-04-24 21:59:44 +00:00
|
|
|
|
2015-12-16 00:52:19 +00:00
|
|
|
enable_rpm_wakeref_asserts(dev_priv);
|
|
|
|
|
2012-04-24 21:59:44 +00:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2014-09-30 08:56:45 +00:00
|
|
|
/**
|
|
|
|
* intel_irq_init - initializes irq support
|
|
|
|
* @dev_priv: i915 device instance
|
|
|
|
*
|
|
|
|
* This function initializes all the irq support including work items, timers
|
|
|
|
* and all the vtables. It does not setup the interrupt itself though.
|
|
|
|
*/
|
2014-09-30 08:56:44 +00:00
|
|
|
void intel_irq_init(struct drm_i915_private *dev_priv)
|
2011-06-28 20:00:41 +00:00
|
|
|
{
|
2016-07-05 09:40:23 +00:00
|
|
|
struct drm_device *dev = &dev_priv->drm;
|
2017-04-28 07:58:39 +00:00
|
|
|
int i;
|
2012-04-24 21:59:41 +00:00
|
|
|
|
2015-06-18 10:06:16 +00:00
|
|
|
intel_hpd_init_work(dev_priv);
|
|
|
|
|
2012-08-08 21:35:35 +00:00
|
|
|
INIT_WORK(&dev_priv->rps.work, gen6_pm_rps_work);
|
2017-04-28 07:58:39 +00:00
|
|
|
|
2012-11-02 18:55:07 +00:00
|
|
|
INIT_WORK(&dev_priv->l3_parity.error_work, ivybridge_parity_work);
|
2017-04-28 07:58:39 +00:00
|
|
|
for (i = 0; i < MAX_L3_SLICES; ++i)
|
|
|
|
dev_priv->l3_parity.remap_info[i] = NULL;
|
2012-04-24 21:59:41 +00:00
|
|
|
|
2016-11-04 14:42:46 +00:00
|
|
|
if (HAS_GUC_SCHED(dev_priv))
|
2016-10-12 16:24:31 +00:00
|
|
|
dev_priv->pm_guc_events = GEN9_GUC_TO_HOST_INT_EVENT;
|
|
|
|
|
2014-03-15 14:53:22 +00:00
|
|
|
/* Let's track the enabled rps events */
|
2015-12-09 20:29:35 +00:00
|
|
|
if (IS_VALLEYVIEW(dev_priv))
|
2014-08-29 11:14:07 +00:00
|
|
|
/* WaGsvRC0ResidencyMethod:vlv */
|
2017-03-09 21:12:30 +00:00
|
|
|
dev_priv->pm_rps_events = GEN6_PM_RP_UP_EI_EXPIRED;
|
2014-07-03 21:33:01 +00:00
|
|
|
else
|
|
|
|
dev_priv->pm_rps_events = GEN6_PM_RPS_EVENTS;
|
2014-03-15 14:53:22 +00:00
|
|
|
|
2017-03-11 02:37:00 +00:00
|
|
|
dev_priv->rps.pm_intrmsk_mbz = 0;
|
2016-05-31 08:28:27 +00:00
|
|
|
|
|
|
|
/*
|
2017-04-13 11:15:27 +00:00
|
|
|
* SNB,IVB,HSW can while VLV,CHV may hard hang on looping batchbuffer
|
2016-05-31 08:28:27 +00:00
|
|
|
* if GEN6_PM_UP_EI_EXPIRED is masked.
|
|
|
|
*
|
|
|
|
* TODO: verify if this can be reproduced on VLV,CHV.
|
|
|
|
*/
|
2017-07-18 18:28:00 +00:00
|
|
|
if (INTEL_GEN(dev_priv) <= 7)
|
2017-03-11 02:37:00 +00:00
|
|
|
dev_priv->rps.pm_intrmsk_mbz |= GEN6_PM_RP_UP_EI_EXPIRED;
|
2016-05-31 08:28:27 +00:00
|
|
|
|
2017-07-18 18:28:00 +00:00
|
|
|
if (INTEL_GEN(dev_priv) >= 8)
|
2017-03-12 13:27:45 +00:00
|
|
|
dev_priv->rps.pm_intrmsk_mbz |= GEN8_PMINTR_DISABLE_REDIRECT_TO_GUC;
|
2016-05-31 08:28:27 +00:00
|
|
|
|
2014-09-30 08:56:44 +00:00
|
|
|
if (IS_GEN2(dev_priv)) {
|
2016-08-03 17:00:56 +00:00
|
|
|
/* Gen2 doesn't have a hardware frame counter */
|
2013-10-11 18:52:44 +00:00
|
|
|
dev->max_vblank_count = 0;
|
2017-07-18 18:28:00 +00:00
|
|
|
} else if (IS_G4X(dev_priv) || INTEL_GEN(dev_priv) >= 5) {
|
2011-06-28 20:00:41 +00:00
|
|
|
dev->max_vblank_count = 0xffffffff; /* full 32 bit counter */
|
2015-09-18 17:03:42 +00:00
|
|
|
dev->driver->get_vblank_counter = g4x_get_vblank_counter;
|
2013-09-25 16:55:26 +00:00
|
|
|
} else {
|
|
|
|
dev->driver->get_vblank_counter = i915_get_vblank_counter;
|
|
|
|
dev->max_vblank_count = 0xffffff; /* only 24 bits of frame count */
|
2011-06-28 20:00:41 +00:00
|
|
|
}
|
|
|
|
|
2014-08-06 11:49:55 +00:00
|
|
|
/*
|
|
|
|
* Opt out of the vblank disable timer on everything except gen2.
|
|
|
|
* Gen2 doesn't have a hardware frame counter and so depends on
|
|
|
|
* vblank interrupts to produce sane vblank seuquence numbers.
|
|
|
|
*/
|
2014-09-30 08:56:44 +00:00
|
|
|
if (!IS_GEN2(dev_priv))
|
2014-08-06 11:49:55 +00:00
|
|
|
dev->vblank_disable_immediate = true;
|
|
|
|
|
2017-02-15 13:15:47 +00:00
|
|
|
/* Most platforms treat the display irq block as an always-on
|
|
|
|
* power domain. vlv/chv can disable it at runtime and need
|
|
|
|
* special care to avoid writing any of the display block registers
|
|
|
|
* outside of the power domain. We defer setting up the display irqs
|
|
|
|
* in this case to the runtime pm.
|
|
|
|
*/
|
|
|
|
dev_priv->display_irqs_enabled = true;
|
|
|
|
if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
|
|
|
|
dev_priv->display_irqs_enabled = false;
|
|
|
|
|
2017-02-04 02:18:25 +00:00
|
|
|
dev_priv->hotplug.hpd_storm_threshold = HPD_STORM_DEFAULT_THRESHOLD;
|
|
|
|
|
drm/vblank: drop the mode argument from drm_calc_vbltimestamp_from_scanoutpos
If we restrict this helper to only kms drivers (which is the case) we
can look up the correct mode easily ourselves. But it's a bit tricky:
- All legacy drivers look at crtc->hwmode. But that is updated already
at the beginning of the modeset helper, which means when we disable
a pipe. Hence the final timestamps might be a bit off. But since
this is an existing bug I'm not going to change it, but just try to
be bug-for-bug compatible with the current code. This only applies
to radeon&amdgpu.
- i915 tries to get it perfect by updating crtc->hwmode when the pipe
is off (i.e. vblank->enabled = false).
- All other atomic drivers look at crtc->state->adjusted_mode. Those
that look at state->requested_mode simply don't adjust their mode,
so it's the same. That has two problems: Accessing crtc->state from
interrupt handling code is unsafe, and it's updated before we shut
down the pipe. For nonblocking modesets it's even worse.
For atomic drivers try to implement what i915 does. To do that we add
a new hwmode field to the vblank structure, and update it from
drm_calc_timestamping_constants(). For atomic drivers that's called
from the right spot by the helper library already, so all fine. But
for safety let's enforce that.
For legacy driver this function is only called at the end (oh the
fun), which is broken, so again let's not bother and just stay
bug-for-bug compatible.
The benefit is that we can use drm_calc_vbltimestamp_from_scanoutpos
directly to implement ->get_vblank_timestamp in every driver, deleting
a lot of code.
v2: Completely new approach, trying to mimick the i915 solution.
v3: Fixup kerneldoc.
v4: Drop the WARN_ON to check that the vblank is off, atomic helpers
currently unconditionally call this. Recomputing the same stuff should
be harmless.
v5: Fix typos and move misplaced hunks to the right patches (Neil).
v6: Undo hunk movement (kbuild).
Cc: Mario Kleiner <mario.kleiner@tuebingen.mpg.de>
Cc: Eric Anholt <eric@anholt.net>
Cc: Rob Clark <robdclark@gmail.com>
Cc: linux-arm-msm@vger.kernel.org
Cc: freedreno@lists.freedesktop.org
Cc: Alex Deucher <alexander.deucher@amd.com>
Cc: Christian König <christian.koenig@amd.com>
Cc: Ben Skeggs <bskeggs@redhat.com>
Reviewed-by: Neil Armstrong <narmstrong@baylibre.com>
Acked-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/20170509140329.24114-4-daniel.vetter@ffwll.ch
2017-05-09 14:03:28 +00:00
|
|
|
dev->driver->get_vblank_timestamp = drm_calc_vbltimestamp_from_scanoutpos;
|
2015-02-13 20:03:44 +00:00
|
|
|
dev->driver->get_scanout_position = i915_get_crtc_scanoutpos;
|
2011-06-28 20:00:41 +00:00
|
|
|
|
2014-09-30 08:56:44 +00:00
|
|
|
if (IS_CHERRYVIEW(dev_priv)) {
|
2014-04-09 17:40:52 +00:00
|
|
|
dev->driver->irq_handler = cherryview_irq_handler;
|
2017-08-18 18:37:04 +00:00
|
|
|
dev->driver->irq_preinstall = cherryview_irq_reset;
|
2014-04-09 17:40:52 +00:00
|
|
|
dev->driver->irq_postinstall = cherryview_irq_postinstall;
|
2017-08-18 18:37:04 +00:00
|
|
|
dev->driver->irq_uninstall = cherryview_irq_reset;
|
2016-10-07 19:49:52 +00:00
|
|
|
dev->driver->enable_vblank = i965_enable_vblank;
|
|
|
|
dev->driver->disable_vblank = i965_disable_vblank;
|
2014-04-09 17:40:52 +00:00
|
|
|
dev_priv->display.hpd_irq_setup = i915_hpd_irq_setup;
|
2014-09-30 08:56:44 +00:00
|
|
|
} else if (IS_VALLEYVIEW(dev_priv)) {
|
2012-03-28 20:39:38 +00:00
|
|
|
dev->driver->irq_handler = valleyview_irq_handler;
|
2017-08-18 18:37:04 +00:00
|
|
|
dev->driver->irq_preinstall = valleyview_irq_reset;
|
2012-03-28 20:39:38 +00:00
|
|
|
dev->driver->irq_postinstall = valleyview_irq_postinstall;
|
2017-08-18 18:37:04 +00:00
|
|
|
dev->driver->irq_uninstall = valleyview_irq_reset;
|
2016-10-07 19:49:52 +00:00
|
|
|
dev->driver->enable_vblank = i965_enable_vblank;
|
|
|
|
dev->driver->disable_vblank = i965_disable_vblank;
|
2013-02-25 17:06:48 +00:00
|
|
|
dev_priv->display.hpd_irq_setup = i915_hpd_irq_setup;
|
2017-07-18 18:28:00 +00:00
|
|
|
} else if (INTEL_GEN(dev_priv) >= 8) {
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
dev->driver->irq_handler = gen8_irq_handler;
|
2014-05-22 15:56:34 +00:00
|
|
|
dev->driver->irq_preinstall = gen8_irq_reset;
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
dev->driver->irq_postinstall = gen8_irq_postinstall;
|
2017-08-18 18:37:04 +00:00
|
|
|
dev->driver->irq_uninstall = gen8_irq_reset;
|
drm/i915/bdw: Implement interrupt changes
The interrupt handling implementation remains the same as previous
generations with the 4 types of registers, status, identity, mask, and
enable. However the layout of where the bits go have changed entirely.
To address these changes, all of the interrupt vfuncs needed special
gen8 code.
The way it works is there is a top level status register now which
informs the interrupt service routine which unit caused the interrupt,
and therefore which interrupt registers to read to process the
interrupt. For display the division is quite logical, a set of interrupt
registers for each pipe, and in addition to those, a set each for "misc"
and port.
For GT the things get a bit hairy, as seen by the code. Each of the GT
units has it's own bits defined. They all look *very similar* and
resides in 16 bits of a GT register. As an example, RCS and BCS share
register 0. To compact the code a bit, at a slight expense to
complexity, this is exactly how the code works as well. 2 structures are
added to the ring buffer so that our ring buffer interrupt handling code
knows which ring shares the interrupt registers, and a shift value (ie.
the top or bottom 16 bits of the register).
The above allows us to kept the interrupt register caching scheme, the
per interrupt enables, and the code to mask and unmask interrupts
relatively clean (again at the cost of some more complexity).
Most of the GT units mentioned above are command streamers, and so the
symmetry should work quite well for even the yet to be implemented rings
which Broadwell adds.
v2: Fixes up a couple of bugs, and is more verbose about errors in the
Broadwell interrupt handler.
v3: fix DE_MISC IER offset
v4: Simplify interrupts:
I totally misread the docs the first time I implemented interrupts, and
so this should greatly simplify the mess. Unlike GEN6, we never touch
the regular mask registers in irq_get/put.
v5: Rebased on to of recent pch hotplug setup changes.
v6: Fixup on top of moving num_pipes to intel_info.
v7: Rebased on top of Egbert Eich's hpd irq handling rework. Also
wired up ibx_hpd_irq_setup for gen8.
v8: Rebase on top of Jani's asle handling rework.
v9: Rebase on top of Ben's VECS enabling for Haswell, where he
unfortunately went OCD on the gt irq #defines. Not that they're still
not yet fully consistent:
- Used the GT_RENDER_ #defines + bdw shifts.
- Dropped the shift from the L3_PARITY stuff, seemed clearer.
- s/irq_refcount/irq_refcount.gt/
v10: Squash in VECS enabling patches and the gen8_gt_irq_handler
refactoring from Zhao Yakui <yakui.zhao@intel.com>
v11: Rebase on top of the interrupt cleanups in upstream.
v12: Rebase on top of Ben's DPF changes in upstream.
v13: Drop bdw from the HAS_L3_DPF feature flag for now, it's unclear what
exactly needs to be done. Requested by Ben.
v14: Fix the patch.
- Drop the mask of reserved bits and assorted logic, it doesn't match
the spec.
- Do the posting read inconditionally instead of commenting it out.
- Add a GEN8_MASTER_IRQ_CONTROL definition and use it.
- Fix up the GEN8_PIPE interrupt defines and give the GEN8_ prefixes -
we actually will need to use them.
- Enclose macros in do {} while (0) (checkpatch).
- Clear DE_MISC interrupt bits only after having processed them.
- Fix whitespace fail (checkpatch).
- Fix overtly long lines where appropriate (checkpatch).
- Don't use typedef'ed private_t (maintainer-scripts).
- Align the function parameter list correctly.
Signed-off-by: Ben Widawsky <ben@bwidawsk.net> (v4)
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
bikeshed
2013-11-03 04:07:09 +00:00
|
|
|
dev->driver->enable_vblank = gen8_enable_vblank;
|
|
|
|
dev->driver->disable_vblank = gen8_disable_vblank;
|
2016-12-02 08:23:49 +00:00
|
|
|
if (IS_GEN9_LP(dev_priv))
|
2015-03-27 12:54:14 +00:00
|
|
|
dev_priv->display.hpd_irq_setup = bxt_hpd_irq_setup;
|
2017-06-02 20:06:39 +00:00
|
|
|
else if (HAS_PCH_SPT(dev_priv) || HAS_PCH_KBP(dev_priv) ||
|
|
|
|
HAS_PCH_CNP(dev_priv))
|
2015-08-27 20:56:02 +00:00
|
|
|
dev_priv->display.hpd_irq_setup = spt_hpd_irq_setup;
|
|
|
|
else
|
2015-08-27 20:56:06 +00:00
|
|
|
dev_priv->display.hpd_irq_setup = ilk_hpd_irq_setup;
|
2016-10-13 10:02:53 +00:00
|
|
|
} else if (HAS_PCH_SPLIT(dev_priv)) {
|
2011-06-28 20:00:41 +00:00
|
|
|
dev->driver->irq_handler = ironlake_irq_handler;
|
2014-05-22 15:56:34 +00:00
|
|
|
dev->driver->irq_preinstall = ironlake_irq_reset;
|
2011-06-28 20:00:41 +00:00
|
|
|
dev->driver->irq_postinstall = ironlake_irq_postinstall;
|
2017-08-18 18:37:04 +00:00
|
|
|
dev->driver->irq_uninstall = ironlake_irq_reset;
|
2011-06-28 20:00:41 +00:00
|
|
|
dev->driver->enable_vblank = ironlake_enable_vblank;
|
|
|
|
dev->driver->disable_vblank = ironlake_disable_vblank;
|
2015-08-27 20:56:04 +00:00
|
|
|
dev_priv->display.hpd_irq_setup = ilk_hpd_irq_setup;
|
2011-06-28 20:00:41 +00:00
|
|
|
} else {
|
2016-05-10 09:57:06 +00:00
|
|
|
if (IS_GEN2(dev_priv)) {
|
2017-08-18 18:37:04 +00:00
|
|
|
dev->driver->irq_preinstall = i8xx_irq_reset;
|
2012-04-22 20:13:57 +00:00
|
|
|
dev->driver->irq_postinstall = i8xx_irq_postinstall;
|
|
|
|
dev->driver->irq_handler = i8xx_irq_handler;
|
2017-08-18 18:37:04 +00:00
|
|
|
dev->driver->irq_uninstall = i8xx_irq_reset;
|
2016-10-07 19:49:52 +00:00
|
|
|
dev->driver->enable_vblank = i8xx_enable_vblank;
|
|
|
|
dev->driver->disable_vblank = i8xx_disable_vblank;
|
2016-05-10 09:57:06 +00:00
|
|
|
} else if (IS_GEN3(dev_priv)) {
|
2017-08-18 18:37:04 +00:00
|
|
|
dev->driver->irq_preinstall = i915_irq_reset;
|
2012-04-24 21:59:44 +00:00
|
|
|
dev->driver->irq_postinstall = i915_irq_postinstall;
|
2017-08-18 18:37:04 +00:00
|
|
|
dev->driver->irq_uninstall = i915_irq_reset;
|
2012-04-24 21:59:44 +00:00
|
|
|
dev->driver->irq_handler = i915_irq_handler;
|
2016-10-07 19:49:52 +00:00
|
|
|
dev->driver->enable_vblank = i8xx_enable_vblank;
|
|
|
|
dev->driver->disable_vblank = i8xx_disable_vblank;
|
2012-04-22 20:13:57 +00:00
|
|
|
} else {
|
2017-08-18 18:37:04 +00:00
|
|
|
dev->driver->irq_preinstall = i965_irq_reset;
|
2012-04-24 21:59:44 +00:00
|
|
|
dev->driver->irq_postinstall = i965_irq_postinstall;
|
2017-08-18 18:37:04 +00:00
|
|
|
dev->driver->irq_uninstall = i965_irq_reset;
|
2012-04-24 21:59:44 +00:00
|
|
|
dev->driver->irq_handler = i965_irq_handler;
|
2016-10-07 19:49:52 +00:00
|
|
|
dev->driver->enable_vblank = i965_enable_vblank;
|
|
|
|
dev->driver->disable_vblank = i965_disable_vblank;
|
2012-04-22 20:13:57 +00:00
|
|
|
}
|
2015-01-09 12:21:13 +00:00
|
|
|
if (I915_HAS_HOTPLUG(dev_priv))
|
|
|
|
dev_priv->display.hpd_irq_setup = i915_hpd_irq_setup;
|
2011-06-28 20:00:41 +00:00
|
|
|
}
|
|
|
|
}
|
2012-12-11 13:05:07 +00:00
|
|
|
|
2017-04-28 07:58:39 +00:00
|
|
|
/**
|
|
|
|
* intel_irq_fini - deinitializes IRQ support
|
|
|
|
* @i915: i915 device instance
|
|
|
|
*
|
|
|
|
* This function deinitializes all the IRQ support.
|
|
|
|
*/
|
|
|
|
void intel_irq_fini(struct drm_i915_private *i915)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < MAX_L3_SLICES; ++i)
|
|
|
|
kfree(i915->l3_parity.remap_info[i]);
|
|
|
|
}
|
|
|
|
|
2014-09-30 08:56:45 +00:00
|
|
|
/**
|
|
|
|
* intel_irq_install - enables the hardware interrupt
|
|
|
|
* @dev_priv: i915 device instance
|
|
|
|
*
|
|
|
|
* This function enables the hardware interrupt handling, but leaves the hotplug
|
|
|
|
* handling still disabled. It is called after intel_irq_init().
|
|
|
|
*
|
|
|
|
* In the driver load and resume code we need working interrupts in a few places
|
|
|
|
* but don't want to deal with the hassle of concurrent probe and hotplug
|
|
|
|
* workers. Hence the split into this two-stage approach.
|
|
|
|
*/
|
2014-09-30 08:56:43 +00:00
|
|
|
int intel_irq_install(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* We enable some interrupt sources in our postinstall hooks, so mark
|
|
|
|
* interrupts as enabled _before_ actually enabling them to avoid
|
|
|
|
* special cases in our ordering checks.
|
|
|
|
*/
|
|
|
|
dev_priv->pm.irqs_enabled = true;
|
|
|
|
|
2016-07-05 09:40:23 +00:00
|
|
|
return drm_irq_install(&dev_priv->drm, dev_priv->drm.pdev->irq);
|
2014-09-30 08:56:43 +00:00
|
|
|
}
|
|
|
|
|
2014-09-30 08:56:45 +00:00
|
|
|
/**
|
|
|
|
* intel_irq_uninstall - finilizes all irq handling
|
|
|
|
* @dev_priv: i915 device instance
|
|
|
|
*
|
|
|
|
* This stops interrupt and hotplug handling and unregisters and frees all
|
|
|
|
* resources acquired in the init functions.
|
|
|
|
*/
|
2014-09-30 08:56:43 +00:00
|
|
|
void intel_irq_uninstall(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
2016-07-05 09:40:23 +00:00
|
|
|
drm_irq_uninstall(&dev_priv->drm);
|
2014-09-30 08:56:43 +00:00
|
|
|
intel_hpd_cancel_work(dev_priv);
|
|
|
|
dev_priv->pm.irqs_enabled = false;
|
|
|
|
}
|
|
|
|
|
2014-09-30 08:56:45 +00:00
|
|
|
/**
|
|
|
|
* intel_runtime_pm_disable_interrupts - runtime interrupt disabling
|
|
|
|
* @dev_priv: i915 device instance
|
|
|
|
*
|
|
|
|
* This function is used to disable interrupts at runtime, both in the runtime
|
|
|
|
* pm and the system suspend/resume code.
|
|
|
|
*/
|
2014-09-30 08:56:44 +00:00
|
|
|
void intel_runtime_pm_disable_interrupts(struct drm_i915_private *dev_priv)
|
2013-08-19 16:18:09 +00:00
|
|
|
{
|
2016-07-05 09:40:23 +00:00
|
|
|
dev_priv->drm.driver->irq_uninstall(&dev_priv->drm);
|
2014-09-30 08:56:43 +00:00
|
|
|
dev_priv->pm.irqs_enabled = false;
|
2016-07-05 09:40:23 +00:00
|
|
|
synchronize_irq(dev_priv->drm.irq);
|
2013-08-19 16:18:09 +00:00
|
|
|
}
|
|
|
|
|
2014-09-30 08:56:45 +00:00
|
|
|
/**
|
|
|
|
* intel_runtime_pm_enable_interrupts - runtime interrupt enabling
|
|
|
|
* @dev_priv: i915 device instance
|
|
|
|
*
|
|
|
|
* This function is used to enable interrupts at runtime, both in the runtime
|
|
|
|
* pm and the system suspend/resume code.
|
|
|
|
*/
|
2014-09-30 08:56:44 +00:00
|
|
|
void intel_runtime_pm_enable_interrupts(struct drm_i915_private *dev_priv)
|
2013-08-19 16:18:09 +00:00
|
|
|
{
|
2014-09-30 08:56:43 +00:00
|
|
|
dev_priv->pm.irqs_enabled = true;
|
2016-07-05 09:40:23 +00:00
|
|
|
dev_priv->drm.driver->irq_preinstall(&dev_priv->drm);
|
|
|
|
dev_priv->drm.driver->irq_postinstall(&dev_priv->drm);
|
2013-08-19 16:18:09 +00:00
|
|
|
}
|