Commit | Line | Data |
---|---|---|
c1b69ed0 CZ |
1 | /* |
2 | * Copyright 2015 Advanced Micro Devices, Inc. | |
3 | * | |
4 | * Permission is hereby granted, free of charge, to any person obtaining a | |
5 | * copy of this software and associated documentation files (the "Software"), | |
6 | * to deal in the Software without restriction, including without limitation | |
7 | * the rights to use, copy, modify, merge, publish, distribute, sublicense, | |
8 | * and/or sell copies of the Software, and to permit persons to whom the | |
9 | * Software is furnished to do so, subject to the following conditions: | |
10 | * | |
11 | * The above copyright notice and this permission notice shall be included in | |
12 | * all copies or substantial portions of the Software. | |
13 | * | |
14 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | |
15 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | |
16 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | |
17 | * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR | |
18 | * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, | |
19 | * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR | |
20 | * OTHER DEALINGS IN THE SOFTWARE. | |
21 | * | |
22 | * | |
23 | */ | |
24 | #include <linux/kthread.h> | |
25 | #include <linux/wait.h> | |
26 | #include <linux/sched.h> | |
27 | #include <drm/drmP.h> | |
28 | #include "amdgpu.h" | |
29 | ||
30 | static int amdgpu_sched_prepare_job(struct amd_gpu_scheduler *sched, | |
91404fb2 | 31 | struct amd_sched_entity *entity, |
953e8fd4 | 32 | struct amd_sched_job *job) |
c1b69ed0 CZ |
33 | { |
34 | int r = 0; | |
953e8fd4 CZ |
35 | struct amdgpu_cs_parser *sched_job; |
36 | if (!job || !job->data) { | |
37 | DRM_ERROR("job is null\n"); | |
38 | return -EINVAL; | |
39 | } | |
40 | ||
41 | sched_job = (struct amdgpu_cs_parser *)job->data; | |
dd01d071 | 42 | if (sched_job->prepare_job) { |
c1b69ed0 | 43 | r = sched_job->prepare_job(sched_job); |
dd01d071 JZ |
44 | if (r) { |
45 | DRM_ERROR("Prepare job error\n"); | |
46 | schedule_work(&sched_job->job_work); | |
47 | } | |
c1b69ed0 CZ |
48 | } |
49 | return r; | |
50 | } | |
51 | ||
6f0e54a9 CK |
52 | static struct fence *amdgpu_sched_run_job(struct amd_gpu_scheduler *sched, |
53 | struct amd_sched_entity *entity, | |
54 | struct amd_sched_job *job) | |
c1b69ed0 CZ |
55 | { |
56 | int r = 0; | |
4cef9267 | 57 | struct amdgpu_cs_parser *sched_job; |
7484667c | 58 | struct amdgpu_fence *fence; |
c1b69ed0 | 59 | |
953e8fd4 | 60 | if (!job || !job->data) { |
4cef9267 | 61 | DRM_ERROR("job is null\n"); |
6f0e54a9 | 62 | return NULL; |
4cef9267 | 63 | } |
953e8fd4 | 64 | sched_job = (struct amdgpu_cs_parser *)job->data; |
c1b69ed0 CZ |
65 | mutex_lock(&sched_job->job_lock); |
66 | r = amdgpu_ib_schedule(sched_job->adev, | |
67 | sched_job->num_ibs, | |
68 | sched_job->ibs, | |
69 | sched_job->filp); | |
70 | if (r) | |
71 | goto err; | |
6f0e54a9 | 72 | fence = amdgpu_fence_ref(sched_job->ibs[sched_job->num_ibs - 1].fence); |
7484667c | 73 | |
c1b69ed0 CZ |
74 | if (sched_job->run_job) { |
75 | r = sched_job->run_job(sched_job); | |
76 | if (r) | |
77 | goto err; | |
78 | } | |
f95b7e3e | 79 | |
c1b69ed0 | 80 | mutex_unlock(&sched_job->job_lock); |
6f0e54a9 CK |
81 | return &fence->base; |
82 | ||
c1b69ed0 CZ |
83 | err: |
84 | DRM_ERROR("Run job error\n"); | |
85 | mutex_unlock(&sched_job->job_lock); | |
86 | schedule_work(&sched_job->job_work); | |
6f0e54a9 | 87 | return NULL; |
c1b69ed0 CZ |
88 | } |
89 | ||
953e8fd4 CZ |
90 | static void amdgpu_sched_process_job(struct amd_gpu_scheduler *sched, |
91 | struct amd_sched_job *job) | |
c1b69ed0 | 92 | { |
953e8fd4 | 93 | struct amdgpu_cs_parser *sched_job; |
c1b69ed0 | 94 | |
953e8fd4 CZ |
95 | if (!job || !job->data) { |
96 | DRM_ERROR("job is null\n"); | |
c1b69ed0 | 97 | return; |
953e8fd4 CZ |
98 | } |
99 | sched_job = (struct amdgpu_cs_parser *)job->data; | |
c1b69ed0 CZ |
100 | schedule_work(&sched_job->job_work); |
101 | } | |
102 | ||
103 | struct amd_sched_backend_ops amdgpu_sched_ops = { | |
104 | .prepare_job = amdgpu_sched_prepare_job, | |
105 | .run_job = amdgpu_sched_run_job, | |
106 | .process_job = amdgpu_sched_process_job | |
107 | }; | |
108 | ||
3c704e93 CZ |
109 | int amdgpu_sched_ib_submit_kernel_helper(struct amdgpu_device *adev, |
110 | struct amdgpu_ring *ring, | |
111 | struct amdgpu_ib *ibs, | |
112 | unsigned num_ibs, | |
113 | int (*free_job)(struct amdgpu_cs_parser *), | |
1763552e CZ |
114 | void *owner, |
115 | struct fence **f) | |
3c704e93 CZ |
116 | { |
117 | int r = 0; | |
118 | if (amdgpu_enable_scheduler) { | |
3c704e93 | 119 | struct amdgpu_cs_parser *sched_job = |
47f38501 | 120 | amdgpu_cs_parser_create(adev, owner, &adev->kernel_ctx, |
4af9f07c | 121 | ibs, num_ibs); |
3c704e93 CZ |
122 | if(!sched_job) { |
123 | return -ENOMEM; | |
124 | } | |
125 | sched_job->free_job = free_job; | |
f556cb0c CZ |
126 | mutex_lock(&sched_job->job_lock); |
127 | r = amd_sched_push_job(ring->scheduler, | |
128 | &adev->kernel_ctx.rings[ring->idx].entity, | |
129 | sched_job, &sched_job->s_fence); | |
130 | if (r) { | |
131 | mutex_unlock(&sched_job->job_lock); | |
132 | kfree(sched_job); | |
133 | return r; | |
134 | } | |
135 | ibs[num_ibs - 1].sequence = sched_job->s_fence->v_seq; | |
281b4223 | 136 | *f = fence_get(&sched_job->s_fence->base); |
f556cb0c CZ |
137 | mutex_unlock(&sched_job->job_lock); |
138 | } else { | |
4af9f07c | 139 | r = amdgpu_ib_schedule(adev, num_ibs, ibs, owner); |
f556cb0c CZ |
140 | if (r) |
141 | return r; | |
281b4223 | 142 | *f = fence_get(&ibs[num_ibs - 1].fence->base); |
f556cb0c | 143 | } |
1763552e | 144 | return 0; |
3c704e93 | 145 | } |