|
|
|
@ -7,6 +7,7 @@
|
|
|
|
|
|
|
|
|
|
#include "gt/intel_gt.h"
|
|
|
|
|
#include "gt/intel_lrc.h"
|
|
|
|
|
#include "gt/shmem_utils.h"
|
|
|
|
|
#include "intel_guc_ads.h"
|
|
|
|
|
#include "intel_guc_fwif.h"
|
|
|
|
|
#include "intel_uc.h"
|
|
|
|
@ -33,6 +34,10 @@
|
|
|
|
|
* +---------------------------------------+ <== dynamic
|
|
|
|
|
* | padding |
|
|
|
|
|
* +---------------------------------------+ <== 4K aligned
|
|
|
|
|
* | golden contexts |
|
|
|
|
|
* +---------------------------------------+
|
|
|
|
|
* | padding |
|
|
|
|
|
* +---------------------------------------+ <== 4K aligned
|
|
|
|
|
* | private data |
|
|
|
|
|
* +---------------------------------------+
|
|
|
|
|
* | padding |
|
|
|
|
@ -52,6 +57,11 @@ static u32 guc_ads_regset_size(struct intel_guc *guc)
|
|
|
|
|
return guc->ads_regset_size;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static u32 guc_ads_golden_ctxt_size(struct intel_guc *guc)
|
|
|
|
|
{
|
|
|
|
|
return PAGE_ALIGN(guc->ads_golden_ctxt_size);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static u32 guc_ads_private_data_size(struct intel_guc *guc)
|
|
|
|
|
{
|
|
|
|
|
return PAGE_ALIGN(guc->fw.private_data_size);
|
|
|
|
@ -62,12 +72,23 @@ static u32 guc_ads_regset_offset(struct intel_guc *guc)
|
|
|
|
|
return offsetof(struct __guc_ads_blob, regset);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static u32 guc_ads_private_data_offset(struct intel_guc *guc)
|
|
|
|
|
static u32 guc_ads_golden_ctxt_offset(struct intel_guc *guc)
|
|
|
|
|
{
|
|
|
|
|
u32 offset;
|
|
|
|
|
|
|
|
|
|
offset = guc_ads_regset_offset(guc) +
|
|
|
|
|
guc_ads_regset_size(guc);
|
|
|
|
|
|
|
|
|
|
return PAGE_ALIGN(offset);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static u32 guc_ads_private_data_offset(struct intel_guc *guc)
|
|
|
|
|
{
|
|
|
|
|
u32 offset;
|
|
|
|
|
|
|
|
|
|
offset = guc_ads_golden_ctxt_offset(guc) +
|
|
|
|
|
guc_ads_golden_ctxt_size(guc);
|
|
|
|
|
|
|
|
|
|
return PAGE_ALIGN(offset);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
@ -319,53 +340,166 @@ static void guc_mmio_reg_state_init(struct intel_guc *guc,
|
|
|
|
|
GEM_BUG_ON(temp_set.size);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* The first 80 dwords of the register state context, containing the
|
|
|
|
|
* execlists and ppgtt registers.
|
|
|
|
|
*/
|
|
|
|
|
#define LR_HW_CONTEXT_SIZE (80 * sizeof(u32))
|
|
|
|
|
static void fill_engine_enable_masks(struct intel_gt *gt,
|
|
|
|
|
struct guc_gt_system_info *info)
|
|
|
|
|
{
|
|
|
|
|
info->engine_enabled_masks[GUC_RENDER_CLASS] = 1;
|
|
|
|
|
info->engine_enabled_masks[GUC_BLITTER_CLASS] = 1;
|
|
|
|
|
info->engine_enabled_masks[GUC_VIDEO_CLASS] = VDBOX_MASK(gt);
|
|
|
|
|
info->engine_enabled_masks[GUC_VIDEOENHANCE_CLASS] = VEBOX_MASK(gt);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void __guc_ads_init(struct intel_guc *guc)
|
|
|
|
|
static int guc_prep_golden_context(struct intel_guc *guc,
|
|
|
|
|
struct __guc_ads_blob *blob)
|
|
|
|
|
{
|
|
|
|
|
struct intel_gt *gt = guc_to_gt(guc);
|
|
|
|
|
struct drm_i915_private *i915 = gt->i915;
|
|
|
|
|
struct __guc_ads_blob *blob = guc->ads_blob;
|
|
|
|
|
const u32 skipped_size = LRC_PPHWSP_SZ * PAGE_SIZE + LR_HW_CONTEXT_SIZE;
|
|
|
|
|
u32 base;
|
|
|
|
|
u32 addr_ggtt, offset;
|
|
|
|
|
u32 total_size = 0, alloc_size, real_size;
|
|
|
|
|
u8 engine_class, guc_class;
|
|
|
|
|
|
|
|
|
|
/* GuC scheduling policies */
|
|
|
|
|
guc_policies_init(guc, &blob->policies);
|
|
|
|
|
struct guc_gt_system_info *info, local_info;
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* GuC expects a per-engine-class context image and size
|
|
|
|
|
* (minus hwsp and ring context). The context image will be
|
|
|
|
|
* used to reinitialize engines after a reset. It must exist
|
|
|
|
|
* and be pinned in the GGTT, so that the address won't change after
|
|
|
|
|
* we have told GuC where to find it. The context size will be used
|
|
|
|
|
* to validate that the LRC base + size fall within allowed GGTT.
|
|
|
|
|
* Reserve the memory for the golden contexts and point GuC at it but
|
|
|
|
|
* leave it empty for now. The context data will be filled in later
|
|
|
|
|
* once there is something available to put there.
|
|
|
|
|
*
|
|
|
|
|
* Note that the HWSP and ring context are not included.
|
|
|
|
|
*
|
|
|
|
|
* Note also that the storage must be pinned in the GGTT, so that the
|
|
|
|
|
* address won't change after GuC has been told where to find it. The
|
|
|
|
|
* GuC will also validate that the LRC base + size fall within the
|
|
|
|
|
* allowed GGTT range.
|
|
|
|
|
*/
|
|
|
|
|
if (blob) {
|
|
|
|
|
offset = guc_ads_golden_ctxt_offset(guc);
|
|
|
|
|
addr_ggtt = intel_guc_ggtt_offset(guc, guc->ads_vma) + offset;
|
|
|
|
|
info = &blob->system_info;
|
|
|
|
|
} else {
|
|
|
|
|
memset(&local_info, 0, sizeof(local_info));
|
|
|
|
|
info = &local_info;
|
|
|
|
|
fill_engine_enable_masks(gt, info);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
for (engine_class = 0; engine_class <= MAX_ENGINE_CLASS; ++engine_class) {
|
|
|
|
|
if (engine_class == OTHER_CLASS)
|
|
|
|
|
continue;
|
|
|
|
|
|
|
|
|
|
guc_class = engine_class_to_guc_class(engine_class);
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* TODO: Set context pointer to default state to allow
|
|
|
|
|
* GuC to re-init guilty contexts after internal reset.
|
|
|
|
|
*/
|
|
|
|
|
blob->ads.golden_context_lrca[guc_class] = 0;
|
|
|
|
|
blob->ads.eng_state_size[guc_class] =
|
|
|
|
|
intel_engine_context_size(gt, engine_class) -
|
|
|
|
|
skipped_size;
|
|
|
|
|
if (!info->engine_enabled_masks[guc_class])
|
|
|
|
|
continue;
|
|
|
|
|
|
|
|
|
|
real_size = intel_engine_context_size(gt, engine_class);
|
|
|
|
|
alloc_size = PAGE_ALIGN(real_size);
|
|
|
|
|
total_size += alloc_size;
|
|
|
|
|
|
|
|
|
|
if (!blob)
|
|
|
|
|
continue;
|
|
|
|
|
|
|
|
|
|
blob->ads.eng_state_size[guc_class] = real_size;
|
|
|
|
|
blob->ads.golden_context_lrca[guc_class] = addr_ggtt;
|
|
|
|
|
addr_ggtt += alloc_size;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (!blob)
|
|
|
|
|
return total_size;
|
|
|
|
|
|
|
|
|
|
GEM_BUG_ON(guc->ads_golden_ctxt_size != total_size);
|
|
|
|
|
return total_size;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static struct intel_engine_cs *find_engine_state(struct intel_gt *gt, u8 engine_class)
|
|
|
|
|
{
|
|
|
|
|
struct intel_engine_cs *engine;
|
|
|
|
|
enum intel_engine_id id;
|
|
|
|
|
|
|
|
|
|
for_each_engine(engine, gt, id) {
|
|
|
|
|
if (engine->class != engine_class)
|
|
|
|
|
continue;
|
|
|
|
|
|
|
|
|
|
if (!engine->default_state)
|
|
|
|
|
continue;
|
|
|
|
|
|
|
|
|
|
return engine;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return NULL;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void guc_init_golden_context(struct intel_guc *guc)
|
|
|
|
|
{
|
|
|
|
|
struct __guc_ads_blob *blob = guc->ads_blob;
|
|
|
|
|
struct intel_engine_cs *engine;
|
|
|
|
|
struct intel_gt *gt = guc_to_gt(guc);
|
|
|
|
|
u32 addr_ggtt, offset;
|
|
|
|
|
u32 total_size = 0, alloc_size, real_size;
|
|
|
|
|
u8 engine_class, guc_class;
|
|
|
|
|
u8 *ptr;
|
|
|
|
|
|
|
|
|
|
/* Skip execlist and PPGTT registers + HWSP */
|
|
|
|
|
const u32 lr_hw_context_size = 80 * sizeof(u32);
|
|
|
|
|
const u32 skip_size = LRC_PPHWSP_SZ * PAGE_SIZE +
|
|
|
|
|
lr_hw_context_size;
|
|
|
|
|
|
|
|
|
|
if (!intel_uc_uses_guc_submission(>->uc))
|
|
|
|
|
return;
|
|
|
|
|
|
|
|
|
|
GEM_BUG_ON(!blob);
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Go back and fill in the golden context data now that it is
|
|
|
|
|
* available.
|
|
|
|
|
*/
|
|
|
|
|
offset = guc_ads_golden_ctxt_offset(guc);
|
|
|
|
|
addr_ggtt = intel_guc_ggtt_offset(guc, guc->ads_vma) + offset;
|
|
|
|
|
ptr = ((u8 *)blob) + offset;
|
|
|
|
|
|
|
|
|
|
for (engine_class = 0; engine_class <= MAX_ENGINE_CLASS; ++engine_class) {
|
|
|
|
|
if (engine_class == OTHER_CLASS)
|
|
|
|
|
continue;
|
|
|
|
|
|
|
|
|
|
guc_class = engine_class_to_guc_class(engine_class);
|
|
|
|
|
|
|
|
|
|
if (!blob->system_info.engine_enabled_masks[guc_class])
|
|
|
|
|
continue;
|
|
|
|
|
|
|
|
|
|
real_size = intel_engine_context_size(gt, engine_class);
|
|
|
|
|
alloc_size = PAGE_ALIGN(real_size);
|
|
|
|
|
total_size += alloc_size;
|
|
|
|
|
|
|
|
|
|
engine = find_engine_state(gt, engine_class);
|
|
|
|
|
if (!engine) {
|
|
|
|
|
drm_err(>->i915->drm, "No engine state recorded for class %d!\n",
|
|
|
|
|
engine_class);
|
|
|
|
|
blob->ads.eng_state_size[guc_class] = 0;
|
|
|
|
|
blob->ads.golden_context_lrca[guc_class] = 0;
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
GEM_BUG_ON(blob->ads.eng_state_size[guc_class] != real_size);
|
|
|
|
|
GEM_BUG_ON(blob->ads.golden_context_lrca[guc_class] != addr_ggtt);
|
|
|
|
|
addr_ggtt += alloc_size;
|
|
|
|
|
|
|
|
|
|
shmem_read(engine->default_state, skip_size, ptr + skip_size,
|
|
|
|
|
real_size - skip_size);
|
|
|
|
|
ptr += alloc_size;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
GEM_BUG_ON(guc->ads_golden_ctxt_size != total_size);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void __guc_ads_init(struct intel_guc *guc)
|
|
|
|
|
{
|
|
|
|
|
struct intel_gt *gt = guc_to_gt(guc);
|
|
|
|
|
struct drm_i915_private *i915 = gt->i915;
|
|
|
|
|
struct __guc_ads_blob *blob = guc->ads_blob;
|
|
|
|
|
u32 base;
|
|
|
|
|
|
|
|
|
|
/* GuC scheduling policies */
|
|
|
|
|
guc_policies_init(guc, &blob->policies);
|
|
|
|
|
|
|
|
|
|
/* System info */
|
|
|
|
|
blob->system_info.engine_enabled_masks[GUC_RENDER_CLASS] = 1;
|
|
|
|
|
blob->system_info.engine_enabled_masks[GUC_BLITTER_CLASS] = 1;
|
|
|
|
|
blob->system_info.engine_enabled_masks[GUC_VIDEO_CLASS] = VDBOX_MASK(gt);
|
|
|
|
|
blob->system_info.engine_enabled_masks[GUC_VIDEOENHANCE_CLASS] = VEBOX_MASK(gt);
|
|
|
|
|
fill_engine_enable_masks(gt, &blob->system_info);
|
|
|
|
|
|
|
|
|
|
blob->system_info.generic_gt_sysinfo[GUC_GENERIC_GT_SYSINFO_SLICE_ENABLED] =
|
|
|
|
|
hweight8(gt->info.sseu.slice_mask);
|
|
|
|
@ -380,6 +514,9 @@ static void __guc_ads_init(struct intel_guc *guc)
|
|
|
|
|
GEN12_DOORBELLS_PER_SQIDI) + 1;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Golden contexts for re-initialising after a watchdog reset */
|
|
|
|
|
guc_prep_golden_context(guc, blob);
|
|
|
|
|
|
|
|
|
|
guc_mapping_table_init(guc_to_gt(guc), &blob->system_info);
|
|
|
|
|
|
|
|
|
|
base = intel_guc_ggtt_offset(guc, guc->ads_vma);
|
|
|
|
@ -417,6 +554,13 @@ int intel_guc_ads_create(struct intel_guc *guc)
|
|
|
|
|
return ret;
|
|
|
|
|
guc->ads_regset_size = ret;
|
|
|
|
|
|
|
|
|
|
/* Likewise the golden contexts: */
|
|
|
|
|
ret = guc_prep_golden_context(guc, NULL);
|
|
|
|
|
if (ret < 0)
|
|
|
|
|
return ret;
|
|
|
|
|
guc->ads_golden_ctxt_size = ret;
|
|
|
|
|
|
|
|
|
|
/* Now the total size can be determined: */
|
|
|
|
|
size = guc_ads_blob_size(guc);
|
|
|
|
|
|
|
|
|
|
ret = intel_guc_allocate_and_map_vma(guc, size, &guc->ads_vma,
|
|
|
|
@ -429,6 +573,18 @@ int intel_guc_ads_create(struct intel_guc *guc)
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
void intel_guc_ads_init_late(struct intel_guc *guc)
|
|
|
|
|
{
|
|
|
|
|
/*
|
|
|
|
|
* The golden context setup requires the saved engine state from
|
|
|
|
|
* __engines_record_defaults(). However, that requires engines to be
|
|
|
|
|
* operational which means the ADS must already have been configured.
|
|
|
|
|
* Fortunately, the golden context state is not needed until a hang
|
|
|
|
|
* occurs, so it can be filled in during this late init phase.
|
|
|
|
|
*/
|
|
|
|
|
guc_init_golden_context(guc);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
void intel_guc_ads_destroy(struct intel_guc *guc)
|
|
|
|
|
{
|
|
|
|
|
i915_vma_unpin_and_release(&guc->ads_vma, I915_VMA_RELEASE_MAP);
|
|
|
|
|