amdgpu_userq_input_va_validate(adev, queue, args->in.rptr_va, AMDGPU_GPU_PAGE_SIZE) ||
amdgpu_userq_input_va_validate(adev, queue, args->in.wptr_va, AMDGPU_GPU_PAGE_SIZE)) {
r = -EINVAL;
- kfree(queue);
- goto unlock;
+ goto free_queue;
}
/* Convert relative doorbell offset into absolute doorbell index */
index = amdgpu_userq_get_doorbell_index(uq_mgr, &db_info, filp);
if (index == (uint64_t)-EINVAL) {
drm_file_err(uq_mgr->file, "Failed to get doorbell for queue\n");
- kfree(queue);
r = -EINVAL;
- goto unlock;
+ goto free_queue;
}
queue->doorbell_index = index;
r = amdgpu_userq_fence_driver_alloc(adev, queue);
if (r) {
drm_file_err(uq_mgr->file, "Failed to alloc fence driver\n");
- goto unlock;
+ goto free_queue;
}
r = uq_funcs->mqd_create(queue, &args->in);
if (r) {
drm_file_err(uq_mgr->file, "Failed to create Queue\n");
- amdgpu_userq_fence_driver_free(queue);
- kfree(queue);
- goto unlock;
+ goto clean_fence_driver;
}
/* don't map the queue if scheduling is halted */
r = amdgpu_userq_map_helper(queue);
if (r) {
drm_file_err(uq_mgr->file, "Failed to map Queue\n");
- uq_funcs->mqd_destroy(queue);
- amdgpu_userq_fence_driver_free(queue);
- kfree(queue);
- goto unlock;
+ down_read(&adev->reset_domain->sem);
+ goto clean_mqd;
}
}
/* Wait for mode-1 reset to complete */
down_read(&adev->reset_domain->sem);
+
r = xa_alloc(&uq_mgr->userq_xa, &qid, queue,
XA_LIMIT(1, AMDGPU_MAX_USERQ_COUNT), GFP_KERNEL);
if (r) {
if (!skip_map_queue)
amdgpu_userq_unmap_helper(queue);
- uq_funcs->mqd_destroy(queue);
- amdgpu_userq_fence_driver_free(queue);
- kfree(queue);
r = -ENOMEM;
- up_read(&adev->reset_domain->sem);
- goto unlock;
+ goto clean_mqd;
}
r = xa_err(xa_store_irq(&adev->userq_doorbell_xa, index, queue, GFP_KERNEL));
if (!skip_map_queue)
amdgpu_userq_unmap_helper(queue);
- uq_funcs->mqd_destroy(queue);
- amdgpu_userq_fence_driver_free(queue);
- kfree(queue);
- up_read(&adev->reset_domain->sem);
- goto unlock;
+ goto clean_mqd;
}
up_read(&adev->reset_domain->sem);
args->out.queue_id = qid;
atomic_inc(&uq_mgr->userq_count[queue->queue_type]);
+ mutex_unlock(&uq_mgr->userq_mutex);
+ return 0;
+clean_mqd:
+ uq_funcs->mqd_destroy(queue);
+ up_read(&adev->reset_domain->sem);
+clean_fence_driver:
+ amdgpu_userq_fence_driver_free(queue);
+free_queue:
+ kfree(queue);
unlock:
mutex_unlock(&uq_mgr->userq_mutex);