forked from Minki/linux
Merge branch 'drm-fixes' of git://people.freedesktop.org/~airlied/linux
Pull drm fixes from Dave Airlie: "Radeon and amdkfd fixes. Radeon ones mostly for oops in some test/benchmark functions since fencing changes, and one regression fix for old GPUs, There is one cirrus regression fix, the 32bpp broke userspace, so this hides it behind a module option for the few users who care. I'm off for a few days, so this is probably the final pull I have, if I see fixes from Intel I'll forward the pull as I should have email" * 'drm-fixes' of git://people.freedesktop.org/~airlied/linux: drm/cirrus: Limit modes depending on bpp option drm/radeon: fix the crash in test functions drm/radeon: fix the crash in benchmark functions drm/radeon: properly set vm fragment size for TN/RL drm/radeon: don't init gpuvm if accel is disabled (v3) drm/radeon: fix PLLs on RS880 and older v2 drm/amdkfd: Don't create BUG due to incorrect user parameter drm/amdkfd: max num of queues can't be 0 drm/amdkfd: Fix bug in accounting of queues
This commit is contained in:
commit
42345d63fd
@ -822,7 +822,7 @@ static void destroy_kernel_queue_cpsch(struct device_queue_manager *dqm,
|
|||||||
* Unconditionally decrement this counter, regardless of the queue's
|
* Unconditionally decrement this counter, regardless of the queue's
|
||||||
* type.
|
* type.
|
||||||
*/
|
*/
|
||||||
dqm->total_queue_count++;
|
dqm->total_queue_count--;
|
||||||
pr_debug("Total of %d queues are accountable so far\n",
|
pr_debug("Total of %d queues are accountable so far\n",
|
||||||
dqm->total_queue_count);
|
dqm->total_queue_count);
|
||||||
mutex_unlock(&dqm->lock);
|
mutex_unlock(&dqm->lock);
|
||||||
|
@ -95,10 +95,10 @@ static int __init kfd_module_init(void)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/* Verify module parameters */
|
/* Verify module parameters */
|
||||||
if ((max_num_of_queues_per_device < 0) ||
|
if ((max_num_of_queues_per_device < 1) ||
|
||||||
(max_num_of_queues_per_device >
|
(max_num_of_queues_per_device >
|
||||||
KFD_MAX_NUM_OF_QUEUES_PER_DEVICE)) {
|
KFD_MAX_NUM_OF_QUEUES_PER_DEVICE)) {
|
||||||
pr_err("kfd: max_num_of_queues_per_device must be between 0 to KFD_MAX_NUM_OF_QUEUES_PER_DEVICE\n");
|
pr_err("kfd: max_num_of_queues_per_device must be between 1 to KFD_MAX_NUM_OF_QUEUES_PER_DEVICE\n");
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -315,7 +315,11 @@ int pqm_update_queue(struct process_queue_manager *pqm, unsigned int qid,
|
|||||||
BUG_ON(!pqm);
|
BUG_ON(!pqm);
|
||||||
|
|
||||||
pqn = get_queue_by_qid(pqm, qid);
|
pqn = get_queue_by_qid(pqm, qid);
|
||||||
BUG_ON(!pqn);
|
if (!pqn) {
|
||||||
|
pr_debug("amdkfd: No queue %d exists for update operation\n",
|
||||||
|
qid);
|
||||||
|
return -EFAULT;
|
||||||
|
}
|
||||||
|
|
||||||
pqn->q->properties.queue_address = p->queue_address;
|
pqn->q->properties.queue_address = p->queue_address;
|
||||||
pqn->q->properties.queue_size = p->queue_size;
|
pqn->q->properties.queue_size = p->queue_size;
|
||||||
|
@ -16,9 +16,12 @@
|
|||||||
#include "cirrus_drv.h"
|
#include "cirrus_drv.h"
|
||||||
|
|
||||||
int cirrus_modeset = -1;
|
int cirrus_modeset = -1;
|
||||||
|
int cirrus_bpp = 24;
|
||||||
|
|
||||||
MODULE_PARM_DESC(modeset, "Disable/Enable modesetting");
|
MODULE_PARM_DESC(modeset, "Disable/Enable modesetting");
|
||||||
module_param_named(modeset, cirrus_modeset, int, 0400);
|
module_param_named(modeset, cirrus_modeset, int, 0400);
|
||||||
|
MODULE_PARM_DESC(bpp, "Max bits-per-pixel (default:24)");
|
||||||
|
module_param_named(bpp, cirrus_bpp, int, 0400);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* This is the generic driver code. This binds the driver to the drm core,
|
* This is the generic driver code. This binds the driver to the drm core,
|
||||||
|
@ -262,4 +262,7 @@ static inline void cirrus_bo_unreserve(struct cirrus_bo *bo)
|
|||||||
|
|
||||||
int cirrus_bo_push_sysram(struct cirrus_bo *bo);
|
int cirrus_bo_push_sysram(struct cirrus_bo *bo);
|
||||||
int cirrus_bo_pin(struct cirrus_bo *bo, u32 pl_flag, u64 *gpu_addr);
|
int cirrus_bo_pin(struct cirrus_bo *bo, u32 pl_flag, u64 *gpu_addr);
|
||||||
|
|
||||||
|
extern int cirrus_bpp;
|
||||||
|
|
||||||
#endif /* __CIRRUS_DRV_H__ */
|
#endif /* __CIRRUS_DRV_H__ */
|
||||||
|
@ -320,6 +320,8 @@ bool cirrus_check_framebuffer(struct cirrus_device *cdev, int width, int height,
|
|||||||
const int max_pitch = 0x1FF << 3; /* (4096 - 1) & ~111b bytes */
|
const int max_pitch = 0x1FF << 3; /* (4096 - 1) & ~111b bytes */
|
||||||
const int max_size = cdev->mc.vram_size;
|
const int max_size = cdev->mc.vram_size;
|
||||||
|
|
||||||
|
if (bpp > cirrus_bpp)
|
||||||
|
return false;
|
||||||
if (bpp > 32)
|
if (bpp > 32)
|
||||||
return false;
|
return false;
|
||||||
|
|
||||||
|
@ -501,8 +501,13 @@ static int cirrus_vga_get_modes(struct drm_connector *connector)
|
|||||||
int count;
|
int count;
|
||||||
|
|
||||||
/* Just add a static list of modes */
|
/* Just add a static list of modes */
|
||||||
count = drm_add_modes_noedid(connector, 1280, 1024);
|
if (cirrus_bpp <= 24) {
|
||||||
drm_set_preferred_mode(connector, 1024, 768);
|
count = drm_add_modes_noedid(connector, 1280, 1024);
|
||||||
|
drm_set_preferred_mode(connector, 1024, 768);
|
||||||
|
} else {
|
||||||
|
count = drm_add_modes_noedid(connector, 800, 600);
|
||||||
|
drm_set_preferred_mode(connector, 800, 600);
|
||||||
|
}
|
||||||
return count;
|
return count;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -34,7 +34,8 @@
|
|||||||
|
|
||||||
static int radeon_benchmark_do_move(struct radeon_device *rdev, unsigned size,
|
static int radeon_benchmark_do_move(struct radeon_device *rdev, unsigned size,
|
||||||
uint64_t saddr, uint64_t daddr,
|
uint64_t saddr, uint64_t daddr,
|
||||||
int flag, int n)
|
int flag, int n,
|
||||||
|
struct reservation_object *resv)
|
||||||
{
|
{
|
||||||
unsigned long start_jiffies;
|
unsigned long start_jiffies;
|
||||||
unsigned long end_jiffies;
|
unsigned long end_jiffies;
|
||||||
@ -47,12 +48,12 @@ static int radeon_benchmark_do_move(struct radeon_device *rdev, unsigned size,
|
|||||||
case RADEON_BENCHMARK_COPY_DMA:
|
case RADEON_BENCHMARK_COPY_DMA:
|
||||||
fence = radeon_copy_dma(rdev, saddr, daddr,
|
fence = radeon_copy_dma(rdev, saddr, daddr,
|
||||||
size / RADEON_GPU_PAGE_SIZE,
|
size / RADEON_GPU_PAGE_SIZE,
|
||||||
NULL);
|
resv);
|
||||||
break;
|
break;
|
||||||
case RADEON_BENCHMARK_COPY_BLIT:
|
case RADEON_BENCHMARK_COPY_BLIT:
|
||||||
fence = radeon_copy_blit(rdev, saddr, daddr,
|
fence = radeon_copy_blit(rdev, saddr, daddr,
|
||||||
size / RADEON_GPU_PAGE_SIZE,
|
size / RADEON_GPU_PAGE_SIZE,
|
||||||
NULL);
|
resv);
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
DRM_ERROR("Unknown copy method\n");
|
DRM_ERROR("Unknown copy method\n");
|
||||||
@ -120,7 +121,8 @@ static void radeon_benchmark_move(struct radeon_device *rdev, unsigned size,
|
|||||||
|
|
||||||
if (rdev->asic->copy.dma) {
|
if (rdev->asic->copy.dma) {
|
||||||
time = radeon_benchmark_do_move(rdev, size, saddr, daddr,
|
time = radeon_benchmark_do_move(rdev, size, saddr, daddr,
|
||||||
RADEON_BENCHMARK_COPY_DMA, n);
|
RADEON_BENCHMARK_COPY_DMA, n,
|
||||||
|
dobj->tbo.resv);
|
||||||
if (time < 0)
|
if (time < 0)
|
||||||
goto out_cleanup;
|
goto out_cleanup;
|
||||||
if (time > 0)
|
if (time > 0)
|
||||||
@ -130,7 +132,8 @@ static void radeon_benchmark_move(struct radeon_device *rdev, unsigned size,
|
|||||||
|
|
||||||
if (rdev->asic->copy.blit) {
|
if (rdev->asic->copy.blit) {
|
||||||
time = radeon_benchmark_do_move(rdev, size, saddr, daddr,
|
time = radeon_benchmark_do_move(rdev, size, saddr, daddr,
|
||||||
RADEON_BENCHMARK_COPY_BLIT, n);
|
RADEON_BENCHMARK_COPY_BLIT, n,
|
||||||
|
dobj->tbo.resv);
|
||||||
if (time < 0)
|
if (time < 0)
|
||||||
goto out_cleanup;
|
goto out_cleanup;
|
||||||
if (time > 0)
|
if (time > 0)
|
||||||
|
@ -960,6 +960,9 @@ void radeon_compute_pll_avivo(struct radeon_pll *pll,
|
|||||||
if (pll->flags & RADEON_PLL_USE_FRAC_FB_DIV &&
|
if (pll->flags & RADEON_PLL_USE_FRAC_FB_DIV &&
|
||||||
pll->flags & RADEON_PLL_USE_REF_DIV)
|
pll->flags & RADEON_PLL_USE_REF_DIV)
|
||||||
ref_div_max = pll->reference_div;
|
ref_div_max = pll->reference_div;
|
||||||
|
else if (pll->flags & RADEON_PLL_PREFER_MINM_OVER_MAXP)
|
||||||
|
/* fix for problems on RS880 */
|
||||||
|
ref_div_max = min(pll->max_ref_div, 7u);
|
||||||
else
|
else
|
||||||
ref_div_max = pll->max_ref_div;
|
ref_div_max = pll->max_ref_div;
|
||||||
|
|
||||||
|
@ -146,7 +146,8 @@ int radeon_gem_object_open(struct drm_gem_object *obj, struct drm_file *file_pri
|
|||||||
struct radeon_bo_va *bo_va;
|
struct radeon_bo_va *bo_va;
|
||||||
int r;
|
int r;
|
||||||
|
|
||||||
if (rdev->family < CHIP_CAYMAN) {
|
if ((rdev->family < CHIP_CAYMAN) ||
|
||||||
|
(!rdev->accel_working)) {
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -176,7 +177,8 @@ void radeon_gem_object_close(struct drm_gem_object *obj,
|
|||||||
struct radeon_bo_va *bo_va;
|
struct radeon_bo_va *bo_va;
|
||||||
int r;
|
int r;
|
||||||
|
|
||||||
if (rdev->family < CHIP_CAYMAN) {
|
if ((rdev->family < CHIP_CAYMAN) ||
|
||||||
|
(!rdev->accel_working)) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -605,14 +605,14 @@ int radeon_driver_open_kms(struct drm_device *dev, struct drm_file *file_priv)
|
|||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
}
|
}
|
||||||
|
|
||||||
vm = &fpriv->vm;
|
|
||||||
r = radeon_vm_init(rdev, vm);
|
|
||||||
if (r) {
|
|
||||||
kfree(fpriv);
|
|
||||||
return r;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (rdev->accel_working) {
|
if (rdev->accel_working) {
|
||||||
|
vm = &fpriv->vm;
|
||||||
|
r = radeon_vm_init(rdev, vm);
|
||||||
|
if (r) {
|
||||||
|
kfree(fpriv);
|
||||||
|
return r;
|
||||||
|
}
|
||||||
|
|
||||||
r = radeon_bo_reserve(rdev->ring_tmp_bo.bo, false);
|
r = radeon_bo_reserve(rdev->ring_tmp_bo.bo, false);
|
||||||
if (r) {
|
if (r) {
|
||||||
radeon_vm_fini(rdev, vm);
|
radeon_vm_fini(rdev, vm);
|
||||||
@ -668,9 +668,9 @@ void radeon_driver_postclose_kms(struct drm_device *dev,
|
|||||||
radeon_vm_bo_rmv(rdev, vm->ib_bo_va);
|
radeon_vm_bo_rmv(rdev, vm->ib_bo_va);
|
||||||
radeon_bo_unreserve(rdev->ring_tmp_bo.bo);
|
radeon_bo_unreserve(rdev->ring_tmp_bo.bo);
|
||||||
}
|
}
|
||||||
|
radeon_vm_fini(rdev, vm);
|
||||||
}
|
}
|
||||||
|
|
||||||
radeon_vm_fini(rdev, vm);
|
|
||||||
kfree(fpriv);
|
kfree(fpriv);
|
||||||
file_priv->driver_priv = NULL;
|
file_priv->driver_priv = NULL;
|
||||||
}
|
}
|
||||||
|
@ -119,11 +119,11 @@ static void radeon_do_test_moves(struct radeon_device *rdev, int flag)
|
|||||||
if (ring == R600_RING_TYPE_DMA_INDEX)
|
if (ring == R600_RING_TYPE_DMA_INDEX)
|
||||||
fence = radeon_copy_dma(rdev, gtt_addr, vram_addr,
|
fence = radeon_copy_dma(rdev, gtt_addr, vram_addr,
|
||||||
size / RADEON_GPU_PAGE_SIZE,
|
size / RADEON_GPU_PAGE_SIZE,
|
||||||
NULL);
|
vram_obj->tbo.resv);
|
||||||
else
|
else
|
||||||
fence = radeon_copy_blit(rdev, gtt_addr, vram_addr,
|
fence = radeon_copy_blit(rdev, gtt_addr, vram_addr,
|
||||||
size / RADEON_GPU_PAGE_SIZE,
|
size / RADEON_GPU_PAGE_SIZE,
|
||||||
NULL);
|
vram_obj->tbo.resv);
|
||||||
if (IS_ERR(fence)) {
|
if (IS_ERR(fence)) {
|
||||||
DRM_ERROR("Failed GTT->VRAM copy %d\n", i);
|
DRM_ERROR("Failed GTT->VRAM copy %d\n", i);
|
||||||
r = PTR_ERR(fence);
|
r = PTR_ERR(fence);
|
||||||
@ -170,11 +170,11 @@ static void radeon_do_test_moves(struct radeon_device *rdev, int flag)
|
|||||||
if (ring == R600_RING_TYPE_DMA_INDEX)
|
if (ring == R600_RING_TYPE_DMA_INDEX)
|
||||||
fence = radeon_copy_dma(rdev, vram_addr, gtt_addr,
|
fence = radeon_copy_dma(rdev, vram_addr, gtt_addr,
|
||||||
size / RADEON_GPU_PAGE_SIZE,
|
size / RADEON_GPU_PAGE_SIZE,
|
||||||
NULL);
|
vram_obj->tbo.resv);
|
||||||
else
|
else
|
||||||
fence = radeon_copy_blit(rdev, vram_addr, gtt_addr,
|
fence = radeon_copy_blit(rdev, vram_addr, gtt_addr,
|
||||||
size / RADEON_GPU_PAGE_SIZE,
|
size / RADEON_GPU_PAGE_SIZE,
|
||||||
NULL);
|
vram_obj->tbo.resv);
|
||||||
if (IS_ERR(fence)) {
|
if (IS_ERR(fence)) {
|
||||||
DRM_ERROR("Failed VRAM->GTT copy %d\n", i);
|
DRM_ERROR("Failed VRAM->GTT copy %d\n", i);
|
||||||
r = PTR_ERR(fence);
|
r = PTR_ERR(fence);
|
||||||
|
@ -743,9 +743,11 @@ static void radeon_vm_frag_ptes(struct radeon_device *rdev,
|
|||||||
*/
|
*/
|
||||||
|
|
||||||
/* NI is optimized for 256KB fragments, SI and newer for 64KB */
|
/* NI is optimized for 256KB fragments, SI and newer for 64KB */
|
||||||
uint64_t frag_flags = rdev->family == CHIP_CAYMAN ?
|
uint64_t frag_flags = ((rdev->family == CHIP_CAYMAN) ||
|
||||||
|
(rdev->family == CHIP_ARUBA)) ?
|
||||||
R600_PTE_FRAG_256KB : R600_PTE_FRAG_64KB;
|
R600_PTE_FRAG_256KB : R600_PTE_FRAG_64KB;
|
||||||
uint64_t frag_align = rdev->family == CHIP_CAYMAN ? 0x200 : 0x80;
|
uint64_t frag_align = ((rdev->family == CHIP_CAYMAN) ||
|
||||||
|
(rdev->family == CHIP_ARUBA)) ? 0x200 : 0x80;
|
||||||
|
|
||||||
uint64_t frag_start = ALIGN(pe_start, frag_align);
|
uint64_t frag_start = ALIGN(pe_start, frag_align);
|
||||||
uint64_t frag_end = pe_end & ~(frag_align - 1);
|
uint64_t frag_end = pe_end & ~(frag_align - 1);
|
||||||
|
Loading…
Reference in New Issue
Block a user