summaryrefslogtreecommitdiff
path: root/drivers
diff options
context:
space:
mode:
authorChunming Zhou <david1.zhou@amd.com>2015-08-05 19:52:14 +0800
committerAlex Deucher <alexander.deucher@amd.com>2015-08-17 16:51:03 -0400
commit4cef92670bc908aaa48771fc9c72f4bcfb7d6a35 (patch)
tree4fbd3338f432f461966bc38a1ade2deb90f0a73f /drivers
parent80de5913cf31c86d64547af0715de4822c9b1abe (diff)
drm/amdgpu: process sched job exactly triggered by fence signal
Signed-off-by: Chunming Zhou <david1.zhou@amd.com> Reviewed-by: Christian K?nig <christian.koenig@amd.com>
Diffstat (limited to 'drivers')
-rw-r--r--drivers/gpu/drm/amd/amdgpu/amdgpu.h2
-rw-r--r--drivers/gpu/drm/amd/amdgpu/amdgpu_sched.c19
-rw-r--r--drivers/gpu/drm/amd/scheduler/gpu_scheduler.c77
-rw-r--r--drivers/gpu/drm/amd/scheduler/gpu_scheduler.h18
4 files changed, 68 insertions, 48 deletions
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu.h b/drivers/gpu/drm/amd/amdgpu/amdgpu.h
index 53d70f766afe..423cf91ef652 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu.h
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu.h
@@ -404,7 +404,7 @@ struct amdgpu_fence_driver {
struct amdgpu_fence {
struct fence base;
- struct fence_cb cb;
+
/* RB, DMA, etc. */
struct amdgpu_ring *ring;
uint64_t seq;
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_sched.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_sched.c
index 0fcf020917d0..71a4a7e4b1ae 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_sched.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_sched.c
@@ -45,19 +45,24 @@ static int amdgpu_sched_prepare_job(struct amd_gpu_scheduler *sched,
static void amdgpu_fence_sched_cb(struct fence *f, struct fence_cb *cb)
{
- struct amdgpu_fence *fence =
- container_of(cb, struct amdgpu_fence, cb);
- amd_sched_isr(fence->ring->scheduler);
+ struct amd_sched_job *sched_job =
+ container_of(cb, struct amd_sched_job, cb);
+ amd_sched_process_job(sched_job);
}
static void amdgpu_sched_run_job(struct amd_gpu_scheduler *sched,
struct amd_context_entity *c_entity,
- void *job)
+ struct amd_sched_job *job)
{
int r = 0;
- struct amdgpu_cs_parser *sched_job = (struct amdgpu_cs_parser *)job;
+ struct amdgpu_cs_parser *sched_job;
struct amdgpu_fence *fence;
+ if (!job || !job->job) {
+ DRM_ERROR("job is null\n");
+ return;
+ }
+ sched_job = (struct amdgpu_cs_parser *)job->job;
mutex_lock(&sched_job->job_lock);
r = amdgpu_ib_schedule(sched_job->adev,
sched_job->num_ibs,
@@ -67,8 +72,10 @@ static void amdgpu_sched_run_job(struct amd_gpu_scheduler *sched,
goto err;
fence = sched_job->ibs[sched_job->num_ibs - 1].fence;
if (fence_add_callback(&fence->base,
- &fence->cb, amdgpu_fence_sched_cb))
+ &job->cb, amdgpu_fence_sched_cb)) {
+ DRM_ERROR("fence add callback failed\n");
goto err;
+ }
if (sched_job->run_job) {
r = sched_job->run_job(sched_job);
diff --git a/drivers/gpu/drm/amd/scheduler/gpu_scheduler.c b/drivers/gpu/drm/amd/scheduler/gpu_scheduler.c
index 1204b7386b39..4c2c5adbc537 100644
--- a/drivers/gpu/drm/amd/scheduler/gpu_scheduler.c
+++ b/drivers/gpu/drm/amd/scheduler/gpu_scheduler.c
@@ -107,7 +107,14 @@ static int gpu_entity_check_status(struct amd_sched_entity *entity)
*/
static bool is_scheduler_ready(struct amd_gpu_scheduler *sched)
{
- return !kfifo_is_full(&sched->active_hw_rq);
+ unsigned long flags;
+ bool full;
+ spin_lock_irqsave(&sched->queue_lock, flags);
+ full = atomic64_read(&sched->hw_rq_count) <
+ sched->hw_submission_limit ? true : false;
+ spin_unlock_irqrestore(&sched->queue_lock, flags);
+
+ return full;
}
/**
@@ -369,6 +376,7 @@ static int amd_sched_main(void *param)
sched_setscheduler(current, SCHED_FIFO, &sparam);
while (!kthread_should_stop()) {
+ struct amd_sched_job *sched_job = NULL;
wait_event_interruptible(sched->wait_queue,
is_scheduler_ready(sched) &&
(c_entity = select_context(sched)));
@@ -376,43 +384,48 @@ static int amd_sched_main(void *param)
if (r != sizeof(void *))
continue;
r = sched->ops->prepare_job(sched, c_entity, job);
- if (!r)
- WARN_ON(kfifo_in_spinlocked(
- &sched->active_hw_rq,
- &job,
- sizeof(void *),
- &sched->queue_lock) != sizeof(void *));
+ if (!r) {
+ unsigned long flags;
+ sched_job = kzalloc(sizeof(struct amd_sched_job),
+ GFP_KERNEL);
+ if (!sched_job) {
+ WARN(true, "No memory to allocate\n");
+ continue;
+ }
+ sched_job->job = job;
+ sched_job->sched = sched;
+ spin_lock_irqsave(&sched->queue_lock, flags);
+ list_add_tail(&sched_job->list, &sched->active_hw_rq);
+ atomic64_inc(&sched->hw_rq_count);
+ spin_unlock_irqrestore(&sched->queue_lock, flags);
+ }
mutex_lock(&sched->sched_lock);
- sched->ops->run_job(sched, c_entity, job);
+ sched->ops->run_job(sched, c_entity, sched_job);
mutex_unlock(&sched->sched_lock);
}
return 0;
}
-uint64_t amd_sched_get_handled_seq(struct amd_gpu_scheduler *sched)
-{
- return atomic64_read(&sched->last_handled_seq);
-}
-
/**
* ISR to handle EOP inetrrupts
*
* @sched: gpu scheduler
*
*/
-void amd_sched_isr(struct amd_gpu_scheduler *sched)
+void amd_sched_process_job(struct amd_sched_job *sched_job)
{
- int r;
- void *job;
- r = kfifo_out_spinlocked(&sched->active_hw_rq,
- &job, sizeof(void *),
- &sched->queue_lock);
-
- if (r != sizeof(void *))
- job = NULL;
-
- sched->ops->process_job(sched, job);
- atomic64_inc(&sched->last_handled_seq);
+ unsigned long flags;
+ struct amd_gpu_scheduler *sched;
+ if (!sched_job)
+ return;
+ sched = sched_job->sched;
+ spin_lock_irqsave(&sched->queue_lock, flags);
+ list_del(&sched_job->list);
+ atomic64_dec(&sched->hw_rq_count);
+ spin_unlock_irqrestore(&sched->queue_lock, flags);
+
+ sched->ops->process_job(sched, sched_job->job);
+ kfree(sched_job);
wake_up_interruptible(&sched->wait_queue);
}
@@ -446,8 +459,7 @@ struct amd_gpu_scheduler *amd_sched_create(void *device,
sched->granularity = granularity;
sched->ring_id = ring;
sched->preemption = preemption;
- atomic64_set(&sched->last_handled_seq, 0);
-
+ sched->hw_submission_limit = hw_submission;
snprintf(name, sizeof(name), "gpu_sched[%d]", ring);
mutex_init(&sched->sched_lock);
spin_lock_init(&sched->queue_lock);
@@ -458,13 +470,8 @@ struct amd_gpu_scheduler *amd_sched_create(void *device,
sched->kernel_rq.check_entity_status = gpu_entity_check_status;
init_waitqueue_head(&sched->wait_queue);
- if(kfifo_alloc(&sched->active_hw_rq,
- hw_submission * sizeof(void *),
- GFP_KERNEL)) {
- kfree(sched);
- return NULL;
- }
-
+ INIT_LIST_HEAD(&sched->active_hw_rq);
+ atomic64_set(&sched->hw_rq_count, 0);
/* Each scheduler will run on a seperate kernel thread */
sched->thread = kthread_create(amd_sched_main, sched, name);
if (sched->thread) {
@@ -473,7 +480,6 @@ struct amd_gpu_scheduler *amd_sched_create(void *device,
}
DRM_ERROR("Failed to create scheduler for id %d.\n", ring);
- kfifo_free(&sched->active_hw_rq);
kfree(sched);
return NULL;
}
@@ -488,7 +494,6 @@ struct amd_gpu_scheduler *amd_sched_create(void *device,
int amd_sched_destroy(struct amd_gpu_scheduler *sched)
{
kthread_stop(sched->thread);
- kfifo_free(&sched->active_hw_rq);
kfree(sched);
return 0;
}
diff --git a/drivers/gpu/drm/amd/scheduler/gpu_scheduler.h b/drivers/gpu/drm/amd/scheduler/gpu_scheduler.h
index 1a01ac45cd4c..8a756a565583 100644
--- a/drivers/gpu/drm/amd/scheduler/gpu_scheduler.h
+++ b/drivers/gpu/drm/amd/scheduler/gpu_scheduler.h
@@ -25,6 +25,7 @@
#define _GPU_SCHEDULER_H_
#include <linux/kfifo.h>
+#include <linux/fence.h>
#define AMD_GPU_WAIT_IDLE_TIMEOUT_IN_MS 3000
@@ -80,6 +81,13 @@ struct amd_context_entity {
bool is_pending;
};
+struct amd_sched_job {
+ struct list_head list;
+ struct fence_cb cb;
+ struct amd_gpu_scheduler *sched;
+ void *job;
+};
+
/**
* Define the backend operations called by the scheduler,
* these functions should be implemented in driver side
@@ -90,7 +98,7 @@ struct amd_sched_backend_ops {
void *job);
void (*run_job)(struct amd_gpu_scheduler *sched,
struct amd_context_entity *c_entity,
- void *job);
+ struct amd_sched_job *job);
void (*process_job)(struct amd_gpu_scheduler *sched, void *job);
};
@@ -102,19 +110,19 @@ struct amd_gpu_scheduler {
struct task_struct *thread;
struct amd_run_queue sched_rq;
struct amd_run_queue kernel_rq;
- struct kfifo active_hw_rq;
+ struct list_head active_hw_rq;
+ atomic64_t hw_rq_count;
struct amd_sched_backend_ops *ops;
uint32_t ring_id;
uint32_t granularity; /* in ms unit */
uint32_t preemption;
- atomic64_t last_handled_seq;
wait_queue_head_t wait_queue;
struct amd_context_entity *current_entity;
struct mutex sched_lock;
spinlock_t queue_lock;
+ uint32_t hw_submission_limit;
};
-
struct amd_gpu_scheduler *amd_sched_create(void *device,
struct amd_sched_backend_ops *ops,
uint32_t ring,
@@ -133,7 +141,7 @@ int amd_sched_wait_emit(struct amd_context_entity *c_entity,
bool intr,
long timeout);
-void amd_sched_isr(struct amd_gpu_scheduler *sched);
+void amd_sched_process_job(struct amd_sched_job *sched_job);
uint64_t amd_sched_get_handled_seq(struct amd_gpu_scheduler *sched);
int amd_context_entity_fini(struct amd_gpu_scheduler *sched,