| 1 | // SPDX-License-Identifier: GPL-2.0-only |
| 2 | /* Copyright (c) 2020 NVIDIA Corporation */ |
| 3 | |
| 4 | #include <linux/host1x.h> |
| 5 | #include <linux/iommu.h> |
| 6 | #include <linux/list.h> |
| 7 | |
| 8 | #include <drm/drm_drv.h> |
| 9 | #include <drm/drm_file.h> |
| 10 | #include <drm/drm_utils.h> |
| 11 | |
| 12 | #include "drm.h" |
| 13 | #include "uapi.h" |
| 14 | |
| 15 | static void tegra_drm_mapping_release(struct kref *ref) |
| 16 | { |
| 17 | struct tegra_drm_mapping *mapping = |
| 18 | container_of(ref, struct tegra_drm_mapping, ref); |
| 19 | |
| 20 | host1x_bo_unpin(map: mapping->map); |
| 21 | host1x_bo_put(bo: mapping->bo); |
| 22 | |
| 23 | kfree(objp: mapping); |
| 24 | } |
| 25 | |
| 26 | void tegra_drm_mapping_put(struct tegra_drm_mapping *mapping) |
| 27 | { |
| 28 | kref_put(kref: &mapping->ref, release: tegra_drm_mapping_release); |
| 29 | } |
| 30 | |
| 31 | static void tegra_drm_channel_context_close(struct tegra_drm_context *context) |
| 32 | { |
| 33 | struct tegra_drm_mapping *mapping; |
| 34 | unsigned long id; |
| 35 | |
| 36 | if (context->memory_context) |
| 37 | host1x_memory_context_put(cd: context->memory_context); |
| 38 | |
| 39 | xa_for_each(&context->mappings, id, mapping) |
| 40 | tegra_drm_mapping_put(mapping); |
| 41 | |
| 42 | xa_destroy(&context->mappings); |
| 43 | |
| 44 | host1x_channel_put(channel: context->channel); |
| 45 | |
| 46 | kfree(objp: context); |
| 47 | } |
| 48 | |
| 49 | void tegra_drm_uapi_close_file(struct tegra_drm_file *file) |
| 50 | { |
| 51 | struct tegra_drm_context *context; |
| 52 | struct host1x_syncpt *sp; |
| 53 | unsigned long id; |
| 54 | |
| 55 | xa_for_each(&file->contexts, id, context) |
| 56 | tegra_drm_channel_context_close(context); |
| 57 | |
| 58 | xa_for_each(&file->syncpoints, id, sp) |
| 59 | host1x_syncpt_put(sp); |
| 60 | |
| 61 | xa_destroy(&file->contexts); |
| 62 | xa_destroy(&file->syncpoints); |
| 63 | } |
| 64 | |
| 65 | static struct tegra_drm_client *tegra_drm_find_client(struct tegra_drm *tegra, u32 class) |
| 66 | { |
| 67 | struct tegra_drm_client *client; |
| 68 | |
| 69 | list_for_each_entry(client, &tegra->clients, list) |
| 70 | if (client->base.class == class) |
| 71 | return client; |
| 72 | |
| 73 | return NULL; |
| 74 | } |
| 75 | |
| 76 | int tegra_drm_ioctl_channel_open(struct drm_device *drm, void *data, struct drm_file *file) |
| 77 | { |
| 78 | struct host1x *host = tegra_drm_to_host1x(tegra: drm->dev_private); |
| 79 | struct tegra_drm_file *fpriv = file->driver_priv; |
| 80 | struct tegra_drm *tegra = drm->dev_private; |
| 81 | struct drm_tegra_channel_open *args = data; |
| 82 | struct tegra_drm_client *client = NULL; |
| 83 | struct tegra_drm_context *context; |
| 84 | int err; |
| 85 | |
| 86 | if (args->flags) |
| 87 | return -EINVAL; |
| 88 | |
| 89 | context = kzalloc(sizeof(*context), GFP_KERNEL); |
| 90 | if (!context) |
| 91 | return -ENOMEM; |
| 92 | |
| 93 | client = tegra_drm_find_client(tegra, class: args->host1x_class); |
| 94 | if (!client) { |
| 95 | err = -ENODEV; |
| 96 | goto free; |
| 97 | } |
| 98 | |
| 99 | if (client->shared_channel) { |
| 100 | context->channel = host1x_channel_get(channel: client->shared_channel); |
| 101 | } else { |
| 102 | context->channel = host1x_channel_request(client: &client->base); |
| 103 | if (!context->channel) { |
| 104 | err = -EBUSY; |
| 105 | goto free; |
| 106 | } |
| 107 | } |
| 108 | |
| 109 | /* Only allocate context if the engine supports context isolation. */ |
| 110 | if (device_iommu_mapped(dev: client->base.dev) && client->ops->can_use_memory_ctx) { |
| 111 | bool supported; |
| 112 | |
| 113 | err = client->ops->can_use_memory_ctx(client, &supported); |
| 114 | if (err) |
| 115 | goto put_channel; |
| 116 | |
| 117 | if (supported) { |
| 118 | struct pid *pid = get_task_pid(current, type: PIDTYPE_TGID); |
| 119 | context->memory_context = host1x_memory_context_alloc( |
| 120 | host1x: host, dev: client->base.dev, pid); |
| 121 | put_pid(pid); |
| 122 | } |
| 123 | |
| 124 | if (IS_ERR(ptr: context->memory_context)) { |
| 125 | if (PTR_ERR(ptr: context->memory_context) != -EOPNOTSUPP) { |
| 126 | err = PTR_ERR(ptr: context->memory_context); |
| 127 | goto put_channel; |
| 128 | } else { |
| 129 | /* |
| 130 | * OK, HW does not support contexts or contexts |
| 131 | * are disabled. |
| 132 | */ |
| 133 | context->memory_context = NULL; |
| 134 | } |
| 135 | } |
| 136 | } |
| 137 | |
| 138 | err = xa_alloc(xa: &fpriv->contexts, id: &args->context, entry: context, XA_LIMIT(1, U32_MAX), |
| 139 | GFP_KERNEL); |
| 140 | if (err < 0) |
| 141 | goto put_memctx; |
| 142 | |
| 143 | context->client = client; |
| 144 | xa_init_flags(xa: &context->mappings, XA_FLAGS_ALLOC1); |
| 145 | |
| 146 | args->version = client->version; |
| 147 | args->capabilities = 0; |
| 148 | |
| 149 | if (device_get_dma_attr(dev: client->base.dev) == DEV_DMA_COHERENT) |
| 150 | args->capabilities |= DRM_TEGRA_CHANNEL_CAP_CACHE_COHERENT; |
| 151 | |
| 152 | return 0; |
| 153 | |
| 154 | put_memctx: |
| 155 | if (context->memory_context) |
| 156 | host1x_memory_context_put(cd: context->memory_context); |
| 157 | put_channel: |
| 158 | host1x_channel_put(channel: context->channel); |
| 159 | free: |
| 160 | kfree(objp: context); |
| 161 | |
| 162 | return err; |
| 163 | } |
| 164 | |
| 165 | int tegra_drm_ioctl_channel_close(struct drm_device *drm, void *data, struct drm_file *file) |
| 166 | { |
| 167 | struct tegra_drm_file *fpriv = file->driver_priv; |
| 168 | struct drm_tegra_channel_close *args = data; |
| 169 | struct tegra_drm_context *context; |
| 170 | |
| 171 | mutex_lock(&fpriv->lock); |
| 172 | |
| 173 | context = xa_load(&fpriv->contexts, index: args->context); |
| 174 | if (!context) { |
| 175 | mutex_unlock(lock: &fpriv->lock); |
| 176 | return -EINVAL; |
| 177 | } |
| 178 | |
| 179 | xa_erase(&fpriv->contexts, index: args->context); |
| 180 | |
| 181 | mutex_unlock(lock: &fpriv->lock); |
| 182 | |
| 183 | tegra_drm_channel_context_close(context); |
| 184 | |
| 185 | return 0; |
| 186 | } |
| 187 | |
| 188 | int tegra_drm_ioctl_channel_map(struct drm_device *drm, void *data, struct drm_file *file) |
| 189 | { |
| 190 | struct tegra_drm_file *fpriv = file->driver_priv; |
| 191 | struct drm_tegra_channel_map *args = data; |
| 192 | struct tegra_drm_mapping *mapping; |
| 193 | struct tegra_drm_context *context; |
| 194 | enum dma_data_direction direction; |
| 195 | struct device *mapping_dev; |
| 196 | int err = 0; |
| 197 | |
| 198 | if (args->flags & ~DRM_TEGRA_CHANNEL_MAP_READ_WRITE) |
| 199 | return -EINVAL; |
| 200 | |
| 201 | mutex_lock(&fpriv->lock); |
| 202 | |
| 203 | context = xa_load(&fpriv->contexts, index: args->context); |
| 204 | if (!context) { |
| 205 | mutex_unlock(lock: &fpriv->lock); |
| 206 | return -EINVAL; |
| 207 | } |
| 208 | |
| 209 | mapping = kzalloc(sizeof(*mapping), GFP_KERNEL); |
| 210 | if (!mapping) { |
| 211 | err = -ENOMEM; |
| 212 | goto unlock; |
| 213 | } |
| 214 | |
| 215 | kref_init(kref: &mapping->ref); |
| 216 | |
| 217 | if (context->memory_context) |
| 218 | mapping_dev = &context->memory_context->dev; |
| 219 | else |
| 220 | mapping_dev = context->client->base.dev; |
| 221 | |
| 222 | mapping->bo = tegra_gem_lookup(file, handle: args->handle); |
| 223 | if (!mapping->bo) { |
| 224 | err = -EINVAL; |
| 225 | goto free; |
| 226 | } |
| 227 | |
| 228 | switch (args->flags & DRM_TEGRA_CHANNEL_MAP_READ_WRITE) { |
| 229 | case DRM_TEGRA_CHANNEL_MAP_READ_WRITE: |
| 230 | direction = DMA_BIDIRECTIONAL; |
| 231 | break; |
| 232 | |
| 233 | case DRM_TEGRA_CHANNEL_MAP_WRITE: |
| 234 | direction = DMA_FROM_DEVICE; |
| 235 | break; |
| 236 | |
| 237 | case DRM_TEGRA_CHANNEL_MAP_READ: |
| 238 | direction = DMA_TO_DEVICE; |
| 239 | break; |
| 240 | |
| 241 | default: |
| 242 | err = -EINVAL; |
| 243 | goto put_gem; |
| 244 | } |
| 245 | |
| 246 | mapping->map = host1x_bo_pin(dev: mapping_dev, bo: mapping->bo, dir: direction, NULL); |
| 247 | if (IS_ERR(ptr: mapping->map)) { |
| 248 | err = PTR_ERR(ptr: mapping->map); |
| 249 | goto put_gem; |
| 250 | } |
| 251 | |
| 252 | mapping->iova = mapping->map->phys; |
| 253 | mapping->iova_end = mapping->iova + host1x_to_tegra_bo(bo: mapping->bo)->gem.size; |
| 254 | |
| 255 | err = xa_alloc(xa: &context->mappings, id: &args->mapping, entry: mapping, XA_LIMIT(1, U32_MAX), |
| 256 | GFP_KERNEL); |
| 257 | if (err < 0) |
| 258 | goto unpin; |
| 259 | |
| 260 | mutex_unlock(lock: &fpriv->lock); |
| 261 | |
| 262 | return 0; |
| 263 | |
| 264 | unpin: |
| 265 | host1x_bo_unpin(map: mapping->map); |
| 266 | put_gem: |
| 267 | host1x_bo_put(bo: mapping->bo); |
| 268 | free: |
| 269 | kfree(objp: mapping); |
| 270 | unlock: |
| 271 | mutex_unlock(lock: &fpriv->lock); |
| 272 | return err; |
| 273 | } |
| 274 | |
| 275 | int tegra_drm_ioctl_channel_unmap(struct drm_device *drm, void *data, struct drm_file *file) |
| 276 | { |
| 277 | struct tegra_drm_file *fpriv = file->driver_priv; |
| 278 | struct drm_tegra_channel_unmap *args = data; |
| 279 | struct tegra_drm_mapping *mapping; |
| 280 | struct tegra_drm_context *context; |
| 281 | |
| 282 | mutex_lock(&fpriv->lock); |
| 283 | |
| 284 | context = xa_load(&fpriv->contexts, index: args->context); |
| 285 | if (!context) { |
| 286 | mutex_unlock(lock: &fpriv->lock); |
| 287 | return -EINVAL; |
| 288 | } |
| 289 | |
| 290 | mapping = xa_erase(&context->mappings, index: args->mapping); |
| 291 | |
| 292 | mutex_unlock(lock: &fpriv->lock); |
| 293 | |
| 294 | if (!mapping) |
| 295 | return -EINVAL; |
| 296 | |
| 297 | tegra_drm_mapping_put(mapping); |
| 298 | return 0; |
| 299 | } |
| 300 | |
| 301 | int tegra_drm_ioctl_syncpoint_allocate(struct drm_device *drm, void *data, struct drm_file *file) |
| 302 | { |
| 303 | struct host1x *host1x = tegra_drm_to_host1x(tegra: drm->dev_private); |
| 304 | struct tegra_drm_file *fpriv = file->driver_priv; |
| 305 | struct drm_tegra_syncpoint_allocate *args = data; |
| 306 | struct host1x_syncpt *sp; |
| 307 | int err; |
| 308 | |
| 309 | if (args->id) |
| 310 | return -EINVAL; |
| 311 | |
| 312 | sp = host1x_syncpt_alloc(host: host1x, HOST1X_SYNCPT_CLIENT_MANAGED, current->comm); |
| 313 | if (!sp) |
| 314 | return -EBUSY; |
| 315 | |
| 316 | args->id = host1x_syncpt_id(sp); |
| 317 | |
| 318 | err = xa_insert(xa: &fpriv->syncpoints, index: args->id, entry: sp, GFP_KERNEL); |
| 319 | if (err) { |
| 320 | host1x_syncpt_put(sp); |
| 321 | return err; |
| 322 | } |
| 323 | |
| 324 | return 0; |
| 325 | } |
| 326 | |
| 327 | int tegra_drm_ioctl_syncpoint_free(struct drm_device *drm, void *data, struct drm_file *file) |
| 328 | { |
| 329 | struct tegra_drm_file *fpriv = file->driver_priv; |
| 330 | struct drm_tegra_syncpoint_allocate *args = data; |
| 331 | struct host1x_syncpt *sp; |
| 332 | |
| 333 | mutex_lock(&fpriv->lock); |
| 334 | sp = xa_erase(&fpriv->syncpoints, index: args->id); |
| 335 | mutex_unlock(lock: &fpriv->lock); |
| 336 | |
| 337 | if (!sp) |
| 338 | return -EINVAL; |
| 339 | |
| 340 | host1x_syncpt_put(sp); |
| 341 | |
| 342 | return 0; |
| 343 | } |
| 344 | |
| 345 | int tegra_drm_ioctl_syncpoint_wait(struct drm_device *drm, void *data, struct drm_file *file) |
| 346 | { |
| 347 | struct host1x *host1x = tegra_drm_to_host1x(tegra: drm->dev_private); |
| 348 | struct drm_tegra_syncpoint_wait *args = data; |
| 349 | signed long timeout_jiffies; |
| 350 | struct host1x_syncpt *sp; |
| 351 | |
| 352 | if (args->padding != 0) |
| 353 | return -EINVAL; |
| 354 | |
| 355 | sp = host1x_syncpt_get_by_id_noref(host: host1x, id: args->id); |
| 356 | if (!sp) |
| 357 | return -EINVAL; |
| 358 | |
| 359 | timeout_jiffies = drm_timeout_abs_to_jiffies(timeout_nsec: args->timeout_ns); |
| 360 | |
| 361 | return host1x_syncpt_wait(sp, thresh: args->threshold, timeout: timeout_jiffies, value: &args->value); |
| 362 | } |
| 363 | |