blob: 2df6fb87e7ff9bab9ebdd3a8a7f512ca002c59f1 [file] [log] [blame]
Oded Gabbay0861e412019-02-16 00:39:14 +02001// SPDX-License-Identifier: GPL-2.0
2
3/*
4 * Copyright 2016-2019 HabanaLabs, Ltd.
5 * All Rights Reserved.
6 */
7
8#include "habanalabs.h"
9
10#include <linux/slab.h>
11
12static void hl_ctx_fini(struct hl_ctx *ctx)
13{
14 struct hl_device *hdev = ctx->hdev;
Oded Gabbayeff6f4a2019-02-16 00:39:21 +020015 int i;
16
17 /*
18 * If we arrived here, there are no jobs waiting for this context
19 * on its queues so we can safely remove it.
20 * This is because for each CS, we increment the ref count and for
21 * every CS that was finished we decrement it and we won't arrive
22 * to this function unless the ref count is 0
23 */
24
25 for (i = 0 ; i < HL_MAX_PENDING_CS ; i++)
26 dma_fence_put(ctx->cs_pending[i]);
Oded Gabbay0861e412019-02-16 00:39:14 +020027
Omer Shpigelman0feaf862019-02-16 00:39:22 +020028 if (ctx->asid != HL_KERNEL_ASID_ID) {
Oded Gabbay86d53072019-07-30 11:49:36 +030029 /* The engines are stopped as there is no executing CS, but the
Omer Shpigelman89225ce2019-05-01 14:38:38 +030030 * Coresight might be still working by accessing addresses
31 * related to the stopped engines. Hence stop it explicitly.
Oded Gabbay86d53072019-07-30 11:49:36 +030032 * Stop only if this is the compute context, as there can be
33 * only one compute context
Omer Shpigelman89225ce2019-05-01 14:38:38 +030034 */
Oded Gabbay86d53072019-07-30 11:49:36 +030035 if ((hdev->in_debug) && (hdev->compute_ctx == ctx))
Oded Gabbay19734972019-05-04 17:36:06 +030036 hl_device_set_debug_mode(hdev, false);
37
Omer Shpigelman0feaf862019-02-16 00:39:22 +020038 hl_vm_ctx_fini(ctx);
Oded Gabbay0861e412019-02-16 00:39:14 +020039 hl_asid_free(hdev, ctx->asid);
Oded Gabbay7aa22272019-05-29 15:27:48 +030040 } else {
41 hl_mmu_ctx_fini(ctx);
Omer Shpigelman0feaf862019-02-16 00:39:22 +020042 }
Oded Gabbay0861e412019-02-16 00:39:14 +020043}
44
45void hl_ctx_do_release(struct kref *ref)
46{
47 struct hl_ctx *ctx;
48
49 ctx = container_of(ref, struct hl_ctx, refcount);
50
Oded Gabbay0861e412019-02-16 00:39:14 +020051 hl_ctx_fini(ctx);
52
53 if (ctx->hpriv)
54 hl_hpriv_put(ctx->hpriv);
55
56 kfree(ctx);
57}
58
59int hl_ctx_create(struct hl_device *hdev, struct hl_fpriv *hpriv)
60{
61 struct hl_ctx_mgr *mgr = &hpriv->ctx_mgr;
62 struct hl_ctx *ctx;
63 int rc;
64
65 ctx = kzalloc(sizeof(*ctx), GFP_KERNEL);
66 if (!ctx) {
67 rc = -ENOMEM;
68 goto out_err;
69 }
70
Oded Gabbayb8887512019-07-15 21:55:57 +030071 mutex_lock(&mgr->ctx_lock);
72 rc = idr_alloc(&mgr->ctx_handles, ctx, 1, 0, GFP_KERNEL);
73 mutex_unlock(&mgr->ctx_lock);
74
75 if (rc < 0) {
76 dev_err(hdev->dev, "Failed to allocate IDR for a new CTX\n");
77 goto free_ctx;
78 }
79
80 ctx->handle = rc;
81
Oded Gabbay0861e412019-02-16 00:39:14 +020082 rc = hl_ctx_init(hdev, ctx, false);
83 if (rc)
Oded Gabbayb8887512019-07-15 21:55:57 +030084 goto remove_from_idr;
Oded Gabbay0861e412019-02-16 00:39:14 +020085
86 hl_hpriv_get(hpriv);
87 ctx->hpriv = hpriv;
88
Oded Gabbay86d53072019-07-30 11:49:36 +030089 /* TODO: remove for multiple contexts per process */
Oded Gabbay0861e412019-02-16 00:39:14 +020090 hpriv->ctx = ctx;
Oded Gabbay86d53072019-07-30 11:49:36 +030091
92 /* TODO: remove the following line for multiple process support */
93 hdev->compute_ctx = ctx;
Oded Gabbay0861e412019-02-16 00:39:14 +020094
Oded Gabbay0861e412019-02-16 00:39:14 +020095 return 0;
96
Oded Gabbayb8887512019-07-15 21:55:57 +030097remove_from_idr:
98 mutex_lock(&mgr->ctx_lock);
99 idr_remove(&mgr->ctx_handles, ctx->handle);
100 mutex_unlock(&mgr->ctx_lock);
Oded Gabbay0861e412019-02-16 00:39:14 +0200101free_ctx:
102 kfree(ctx);
103out_err:
104 return rc;
105}
106
107void hl_ctx_free(struct hl_device *hdev, struct hl_ctx *ctx)
108{
109 if (kref_put(&ctx->refcount, hl_ctx_do_release) == 1)
110 return;
111
112 dev_warn(hdev->dev,
113 "Context %d closed or terminated but its CS are executing\n",
114 ctx->asid);
115}
116
117int hl_ctx_init(struct hl_device *hdev, struct hl_ctx *ctx, bool is_kernel_ctx)
118{
Omer Shpigelman0feaf862019-02-16 00:39:22 +0200119 int rc = 0;
120
Oded Gabbay0861e412019-02-16 00:39:14 +0200121 ctx->hdev = hdev;
122
123 kref_init(&ctx->refcount);
124
Oded Gabbayeff6f4a2019-02-16 00:39:21 +0200125 ctx->cs_sequence = 1;
126 spin_lock_init(&ctx->cs_lock);
Oded Gabbay027d35d2019-04-25 20:15:42 +0300127 atomic_set(&ctx->thread_ctx_switch_token, 1);
128 ctx->thread_ctx_switch_wait_token = 0;
Oded Gabbayeff6f4a2019-02-16 00:39:21 +0200129
Oded Gabbay0861e412019-02-16 00:39:14 +0200130 if (is_kernel_ctx) {
Oded Gabbay4c172bb2019-08-30 16:59:33 +0300131 ctx->asid = HL_KERNEL_ASID_ID; /* Kernel driver gets ASID 0 */
Oded Gabbay7aa22272019-05-29 15:27:48 +0300132 rc = hl_mmu_ctx_init(ctx);
133 if (rc) {
134 dev_err(hdev->dev, "Failed to init mmu ctx module\n");
135 goto mem_ctx_err;
136 }
Oded Gabbay0861e412019-02-16 00:39:14 +0200137 } else {
138 ctx->asid = hl_asid_alloc(hdev);
139 if (!ctx->asid) {
140 dev_err(hdev->dev, "No free ASID, failed to create context\n");
141 return -ENOMEM;
142 }
Omer Shpigelman0feaf862019-02-16 00:39:22 +0200143
144 rc = hl_vm_ctx_init(ctx);
145 if (rc) {
146 dev_err(hdev->dev, "Failed to init mem ctx module\n");
147 rc = -ENOMEM;
148 goto mem_ctx_err;
149 }
Oded Gabbay0861e412019-02-16 00:39:14 +0200150 }
151
Oded Gabbay0861e412019-02-16 00:39:14 +0200152 return 0;
Omer Shpigelman0feaf862019-02-16 00:39:22 +0200153
154mem_ctx_err:
155 if (ctx->asid != HL_KERNEL_ASID_ID)
156 hl_asid_free(hdev, ctx->asid);
157
158 return rc;
Oded Gabbay0861e412019-02-16 00:39:14 +0200159}
160
161void hl_ctx_get(struct hl_device *hdev, struct hl_ctx *ctx)
162{
163 kref_get(&ctx->refcount);
164}
165
166int hl_ctx_put(struct hl_ctx *ctx)
167{
168 return kref_put(&ctx->refcount, hl_ctx_do_release);
169}
170
Oded Gabbayeff6f4a2019-02-16 00:39:21 +0200171struct dma_fence *hl_ctx_get_fence(struct hl_ctx *ctx, u64 seq)
172{
173 struct hl_device *hdev = ctx->hdev;
174 struct dma_fence *fence;
175
176 spin_lock(&ctx->cs_lock);
177
178 if (seq >= ctx->cs_sequence) {
Oded Gabbay018e0e32019-12-03 10:12:10 +0200179 dev_notice_ratelimited(hdev->dev,
Oded Gabbayeff6f4a2019-02-16 00:39:21 +0200180 "Can't wait on seq %llu because current CS is at seq %llu\n",
181 seq, ctx->cs_sequence);
182 spin_unlock(&ctx->cs_lock);
183 return ERR_PTR(-EINVAL);
184 }
185
186
187 if (seq + HL_MAX_PENDING_CS < ctx->cs_sequence) {
188 dev_dbg(hdev->dev,
189 "Can't wait on seq %llu because current CS is at seq %llu (Fence is gone)\n",
190 seq, ctx->cs_sequence);
191 spin_unlock(&ctx->cs_lock);
192 return NULL;
193 }
194
195 fence = dma_fence_get(
196 ctx->cs_pending[seq & (HL_MAX_PENDING_CS - 1)]);
197 spin_unlock(&ctx->cs_lock);
198
199 return fence;
200}
201
Oded Gabbay0861e412019-02-16 00:39:14 +0200202/*
203 * hl_ctx_mgr_init - initialize the context manager
204 *
205 * @mgr: pointer to context manager structure
206 *
207 * This manager is an object inside the hpriv object of the user process.
208 * The function is called when a user process opens the FD.
209 */
210void hl_ctx_mgr_init(struct hl_ctx_mgr *mgr)
211{
212 mutex_init(&mgr->ctx_lock);
213 idr_init(&mgr->ctx_handles);
214}
215
216/*
217 * hl_ctx_mgr_fini - finalize the context manager
218 *
219 * @hdev: pointer to device structure
220 * @mgr: pointer to context manager structure
221 *
222 * This function goes over all the contexts in the manager and frees them.
223 * It is called when a process closes the FD.
224 */
225void hl_ctx_mgr_fini(struct hl_device *hdev, struct hl_ctx_mgr *mgr)
226{
227 struct hl_ctx *ctx;
228 struct idr *idp;
229 u32 id;
230
231 idp = &mgr->ctx_handles;
232
233 idr_for_each_entry(idp, ctx, id)
234 hl_ctx_free(hdev, ctx);
235
236 idr_destroy(&mgr->ctx_handles);
237 mutex_destroy(&mgr->ctx_lock);
238}