struct amdgpu_user_fence uf;
        int (*free_job)(struct amdgpu_job *job);
 };
+#define to_amdgpu_job(sched_job)               \
+               container_of((sched_job), struct amdgpu_job, base)
 
 static inline u32 amdgpu_get_ib_value(struct amdgpu_cs_parser *p, uint32_t ib_idx, int idx)
 {
 
 
                job->free_job = amdgpu_cs_free_job;
                mutex_lock(&job->job_lock);
-               r = amd_sched_entity_push_job((struct amd_sched_job *)job);
+               r = amd_sched_entity_push_job(&job->base);
                if (r) {
                        mutex_unlock(&job->job_lock);
                        amdgpu_cs_free_job(job);
 
 
 static struct fence *amdgpu_sched_dependency(struct amd_sched_job *sched_job)
 {
-       struct amdgpu_job *job = (struct amdgpu_job *)sched_job;
+       struct amdgpu_job *job = to_amdgpu_job(sched_job);
        return amdgpu_sync_get_fence(&job->ibs->sync);
 }
 
                DRM_ERROR("job is null\n");
                return NULL;
        }
-       job = (struct amdgpu_job *)sched_job;
+       job = to_amdgpu_job(sched_job);
        mutex_lock(&job->job_lock);
        r = amdgpu_ib_schedule(job->adev,
                               job->num_ibs,
                mutex_init(&job->job_lock);
                job->free_job = free_job;
                mutex_lock(&job->job_lock);
-               r = amd_sched_entity_push_job((struct amd_sched_job *)job);
+               r = amd_sched_entity_push_job(&job->base);
                if (r) {
                        mutex_unlock(&job->job_lock);
                        kfree(job);