summaryrefslogtreecommitdiff
path: root/drivers/gpu/drm/amd/scheduler/gpu_scheduler.c
diff options
context:
space:
mode:
authorMonk Liu <Monk.Liu@amd.com>2016-03-10 12:14:44 +0800
committerAlex Deucher <alexander.deucher@amd.com>2016-05-02 15:20:07 -0400
commitb6723c8da55af5309cf06e71a5228f3c02846c5a (patch)
tree268b16b48fca9556569ce72c02b9c1a0985b3df3 /drivers/gpu/drm/amd/scheduler/gpu_scheduler.c
parent0de2479c953ae07fd11e7b1bc8d4fc831e6842bb (diff)
downloadlinux-b6723c8da55af5309cf06e71a5228f3c02846c5a.tar.gz
linux-b6723c8da55af5309cf06e71a5228f3c02846c5a.tar.bz2
linux-b6723c8da55af5309cf06e71a5228f3c02846c5a.zip
drm/amdgpu: use ref to keep job alive
this is to fix fatal page fault error that occured if: job is signaled/released after its timeout work is already put to the global queue (in this case the cancel_delayed_work will return false), which will lead to NX-protection error page fault during job_timeout_func. Signed-off-by: Monk Liu <Monk.Liu@amd.com> Reviewed-by: Chunming Zhou <david1.zhou@amd.com> Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
Diffstat (limited to 'drivers/gpu/drm/amd/scheduler/gpu_scheduler.c')
-rw-r--r--drivers/gpu/drm/amd/scheduler/gpu_scheduler.c8
1 files changed, 7 insertions, 1 deletions
diff --git a/drivers/gpu/drm/amd/scheduler/gpu_scheduler.c b/drivers/gpu/drm/amd/scheduler/gpu_scheduler.c
index b7e8071448c6..639c70de217c 100644
--- a/drivers/gpu/drm/amd/scheduler/gpu_scheduler.c
+++ b/drivers/gpu/drm/amd/scheduler/gpu_scheduler.c
@@ -333,7 +333,8 @@ void amd_sched_job_finish(struct amd_sched_job *s_job)
struct amd_gpu_scheduler *sched = s_job->sched;
if (sched->timeout != MAX_SCHEDULE_TIMEOUT) {
- cancel_delayed_work(&s_job->work_tdr); /*TODO: how to deal the case that tdr is running */
+ if (cancel_delayed_work(&s_job->work_tdr))
+ amd_sched_job_put(s_job);
/* queue TDR for next job */
next = list_first_entry_or_null(&sched->ring_mirror_list,
@@ -341,6 +342,7 @@ void amd_sched_job_finish(struct amd_sched_job *s_job)
if (next) {
INIT_DELAYED_WORK(&next->work_tdr, s_job->timeout_callback);
+ amd_sched_job_get(next);
schedule_delayed_work(&next->work_tdr, sched->timeout);
}
}
@@ -354,6 +356,7 @@ void amd_sched_job_begin(struct amd_sched_job *s_job)
list_first_entry_or_null(&sched->ring_mirror_list, struct amd_sched_job, node) == s_job)
{
INIT_DELAYED_WORK(&s_job->work_tdr, s_job->timeout_callback);
+ amd_sched_job_get(s_job);
schedule_delayed_work(&s_job->work_tdr, sched->timeout);
}
}
@@ -382,9 +385,11 @@ int amd_sched_job_init(struct amd_sched_job *job,
struct amd_gpu_scheduler *sched,
struct amd_sched_entity *entity,
void (*timeout_cb)(struct work_struct *work),
+ void (*free_cb)(struct kref *refcount),
void *owner, struct fence **fence)
{
INIT_LIST_HEAD(&job->node);
+ kref_init(&job->refcount);
job->sched = sched;
job->s_entity = entity;
job->s_fence = amd_sched_fence_create(entity, owner);
@@ -393,6 +398,7 @@ int amd_sched_job_init(struct amd_sched_job *job,
job->s_fence->s_job = job;
job->timeout_callback = timeout_cb;
+ job->free_callback = free_cb;
if (fence)
*fence = &job->s_fence->base;