1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * Copyright © 2015 Broadcom
4 */
5
6/**
7 * DOC: VC4 GEM BO management support
8 *
9 * The VC4 GPU architecture (both scanout and rendering) has direct
10 * access to system memory with no MMU in between. To support it, we
11 * use the GEM DMA helper functions to allocate contiguous ranges of
12 * physical memory for our BOs.
13 *
14 * Since the DMA allocator is very slow, we keep a cache of recently
15 * freed BOs around so that the kernel's allocation of objects for 3D
16 * rendering can return quickly.
17 */
18
19#include <linux/dma-buf.h>
20
21#include <drm/drm_fourcc.h>
22#include <drm/drm_print.h>
23
24#include "vc4_drv.h"
25#include "uapi/drm/vc4_drm.h"
26
27static const struct drm_gem_object_funcs vc4_gem_object_funcs;
28
29static const char * const bo_type_names[] = {
30 "kernel",
31 "V3D",
32 "V3D shader",
33 "dumb",
34 "binner",
35 "RCL",
36 "BCL",
37 "kernel BO cache",
38};
39
40static bool is_user_label(int label)
41{
42 return label >= VC4_BO_TYPE_COUNT;
43}
44
45static void vc4_bo_stats_print(struct drm_printer *p, struct vc4_dev *vc4)
46{
47 int i;
48
49 for (i = 0; i < vc4->num_labels; i++) {
50 if (!vc4->bo_labels[i].num_allocated)
51 continue;
52
53 drm_printf(p, f: "%30s: %6dkb BOs (%d)\n",
54 vc4->bo_labels[i].name,
55 vc4->bo_labels[i].size_allocated / 1024,
56 vc4->bo_labels[i].num_allocated);
57 }
58
59 mutex_lock(&vc4->purgeable.lock);
60 if (vc4->purgeable.num)
61 drm_printf(p, f: "%30s: %6zdkb BOs (%d)\n", "userspace BO cache",
62 vc4->purgeable.size / 1024, vc4->purgeable.num);
63
64 if (vc4->purgeable.purged_num)
65 drm_printf(p, f: "%30s: %6zdkb BOs (%d)\n", "total purged BO",
66 vc4->purgeable.purged_size / 1024,
67 vc4->purgeable.purged_num);
68 mutex_unlock(lock: &vc4->purgeable.lock);
69}
70
71static int vc4_bo_stats_debugfs(struct seq_file *m, void *unused)
72{
73 struct drm_debugfs_entry *entry = m->private;
74 struct drm_device *dev = entry->dev;
75 struct vc4_dev *vc4 = to_vc4_dev(dev);
76 struct drm_printer p = drm_seq_file_printer(f: m);
77
78 vc4_bo_stats_print(p: &p, vc4);
79
80 return 0;
81}
82
83/* Takes ownership of *name and returns the appropriate slot for it in
84 * the bo_labels[] array, extending it as necessary.
85 *
86 * This is inefficient and could use a hash table instead of walking
87 * an array and strcmp()ing. However, the assumption is that user
88 * labeling will be infrequent (scanout buffers and other long-lived
89 * objects, or debug driver builds), so we can live with it for now.
90 */
91static int vc4_get_user_label(struct vc4_dev *vc4, const char *name)
92{
93 int i;
94 int free_slot = -1;
95
96 for (i = 0; i < vc4->num_labels; i++) {
97 if (!vc4->bo_labels[i].name) {
98 free_slot = i;
99 } else if (strcmp(vc4->bo_labels[i].name, name) == 0) {
100 kfree(objp: name);
101 return i;
102 }
103 }
104
105 if (free_slot != -1) {
106 WARN_ON(vc4->bo_labels[free_slot].num_allocated != 0);
107 vc4->bo_labels[free_slot].name = name;
108 return free_slot;
109 } else {
110 u32 new_label_count = vc4->num_labels + 1;
111 struct vc4_label *new_labels =
112 krealloc(vc4->bo_labels,
113 new_label_count * sizeof(*new_labels),
114 GFP_KERNEL);
115
116 if (!new_labels) {
117 kfree(objp: name);
118 return -1;
119 }
120
121 free_slot = vc4->num_labels;
122 vc4->bo_labels = new_labels;
123 vc4->num_labels = new_label_count;
124
125 vc4->bo_labels[free_slot].name = name;
126 vc4->bo_labels[free_slot].num_allocated = 0;
127 vc4->bo_labels[free_slot].size_allocated = 0;
128
129 return free_slot;
130 }
131}
132
133static void vc4_bo_set_label(struct drm_gem_object *gem_obj, int label)
134{
135 struct vc4_bo *bo = to_vc4_bo(gem_obj);
136 struct vc4_dev *vc4 = to_vc4_dev(gem_obj->dev);
137
138 lockdep_assert_held(&vc4->bo_lock);
139
140 if (label != -1) {
141 vc4->bo_labels[label].num_allocated++;
142 vc4->bo_labels[label].size_allocated += gem_obj->size;
143 }
144
145 vc4->bo_labels[bo->label].num_allocated--;
146 vc4->bo_labels[bo->label].size_allocated -= gem_obj->size;
147
148 if (vc4->bo_labels[bo->label].num_allocated == 0 &&
149 is_user_label(label: bo->label)) {
150 /* Free user BO label slots on last unreference.
151 * Slots are just where we track the stats for a given
152 * name, and once a name is unused we can reuse that
153 * slot.
154 */
155 kfree(objp: vc4->bo_labels[bo->label].name);
156 vc4->bo_labels[bo->label].name = NULL;
157 }
158
159 bo->label = label;
160}
161
162static uint32_t bo_page_index(size_t size)
163{
164 return (size / PAGE_SIZE) - 1;
165}
166
167static void vc4_bo_destroy(struct vc4_bo *bo)
168{
169 struct drm_gem_object *obj = &bo->base.base;
170 struct vc4_dev *vc4 = to_vc4_dev(obj->dev);
171
172 lockdep_assert_held(&vc4->bo_lock);
173
174 vc4_bo_set_label(gem_obj: obj, label: -1);
175
176 if (bo->validated_shader) {
177 kfree(objp: bo->validated_shader->uniform_addr_offsets);
178 kfree(objp: bo->validated_shader->texture_samples);
179 kfree(objp: bo->validated_shader);
180 bo->validated_shader = NULL;
181 }
182
183 mutex_destroy(lock: &bo->madv_lock);
184 drm_gem_dma_free(dma_obj: &bo->base);
185}
186
187static void vc4_bo_remove_from_cache(struct vc4_bo *bo)
188{
189 struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev);
190
191 lockdep_assert_held(&vc4->bo_lock);
192 list_del(entry: &bo->unref_head);
193 list_del(entry: &bo->size_head);
194}
195
196static struct list_head *vc4_get_cache_list_for_size(struct drm_device *dev,
197 size_t size)
198{
199 struct vc4_dev *vc4 = to_vc4_dev(dev);
200 uint32_t page_index = bo_page_index(size);
201
202 if (vc4->bo_cache.size_list_size <= page_index) {
203 uint32_t new_size = max(vc4->bo_cache.size_list_size * 2,
204 page_index + 1);
205 struct list_head *new_list;
206 uint32_t i;
207
208 new_list = kmalloc_array(new_size, sizeof(struct list_head),
209 GFP_KERNEL);
210 if (!new_list)
211 return NULL;
212
213 /* Rebase the old cached BO lists to their new list
214 * head locations.
215 */
216 for (i = 0; i < vc4->bo_cache.size_list_size; i++) {
217 struct list_head *old_list =
218 &vc4->bo_cache.size_list[i];
219
220 if (list_empty(head: old_list))
221 INIT_LIST_HEAD(list: &new_list[i]);
222 else
223 list_replace(old: old_list, new: &new_list[i]);
224 }
225 /* And initialize the brand new BO list heads. */
226 for (i = vc4->bo_cache.size_list_size; i < new_size; i++)
227 INIT_LIST_HEAD(list: &new_list[i]);
228
229 kfree(objp: vc4->bo_cache.size_list);
230 vc4->bo_cache.size_list = new_list;
231 vc4->bo_cache.size_list_size = new_size;
232 }
233
234 return &vc4->bo_cache.size_list[page_index];
235}
236
237static void vc4_bo_cache_purge(struct drm_device *dev)
238{
239 struct vc4_dev *vc4 = to_vc4_dev(dev);
240
241 mutex_lock(&vc4->bo_lock);
242 while (!list_empty(head: &vc4->bo_cache.time_list)) {
243 struct vc4_bo *bo = list_last_entry(&vc4->bo_cache.time_list,
244 struct vc4_bo, unref_head);
245 vc4_bo_remove_from_cache(bo);
246 vc4_bo_destroy(bo);
247 }
248 mutex_unlock(lock: &vc4->bo_lock);
249}
250
251void vc4_bo_add_to_purgeable_pool(struct vc4_bo *bo)
252{
253 struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev);
254
255 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4))
256 return;
257
258 mutex_lock(&vc4->purgeable.lock);
259 list_add_tail(new: &bo->size_head, head: &vc4->purgeable.list);
260 vc4->purgeable.num++;
261 vc4->purgeable.size += bo->base.base.size;
262 mutex_unlock(lock: &vc4->purgeable.lock);
263}
264
265static void vc4_bo_remove_from_purgeable_pool_locked(struct vc4_bo *bo)
266{
267 struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev);
268
269 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4))
270 return;
271
272 /* list_del_init() is used here because the caller might release
273 * the purgeable lock in order to acquire the madv one and update the
274 * madv status.
275 * During this short period of time a user might decide to mark
276 * the BO as unpurgeable, and if bo->madv is set to
277 * VC4_MADV_DONTNEED it will try to remove the BO from the
278 * purgeable list which will fail if the ->next/prev fields
279 * are set to LIST_POISON1/LIST_POISON2 (which is what
280 * list_del() does).
281 * Re-initializing the list element guarantees that list_del()
282 * will work correctly even if it's a NOP.
283 */
284 list_del_init(entry: &bo->size_head);
285 vc4->purgeable.num--;
286 vc4->purgeable.size -= bo->base.base.size;
287}
288
289void vc4_bo_remove_from_purgeable_pool(struct vc4_bo *bo)
290{
291 struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev);
292
293 mutex_lock(&vc4->purgeable.lock);
294 vc4_bo_remove_from_purgeable_pool_locked(bo);
295 mutex_unlock(lock: &vc4->purgeable.lock);
296}
297
298static void vc4_bo_purge(struct drm_gem_object *obj)
299{
300 struct vc4_bo *bo = to_vc4_bo(obj);
301 struct drm_device *dev = obj->dev;
302
303 WARN_ON(!mutex_is_locked(&bo->madv_lock));
304 WARN_ON(bo->madv != VC4_MADV_DONTNEED);
305
306 drm_vma_node_unmap(node: &obj->vma_node, file_mapping: dev->anon_inode->i_mapping);
307
308 dma_free_wc(dev: dev->dev, size: obj->size, cpu_addr: bo->base.vaddr, dma_addr: bo->base.dma_addr);
309 bo->base.vaddr = NULL;
310 bo->madv = __VC4_MADV_PURGED;
311}
312
313static void vc4_bo_userspace_cache_purge(struct drm_device *dev)
314{
315 struct vc4_dev *vc4 = to_vc4_dev(dev);
316
317 mutex_lock(&vc4->purgeable.lock);
318 while (!list_empty(head: &vc4->purgeable.list)) {
319 struct vc4_bo *bo = list_first_entry(&vc4->purgeable.list,
320 struct vc4_bo, size_head);
321 struct drm_gem_object *obj = &bo->base.base;
322 size_t purged_size = 0;
323
324 vc4_bo_remove_from_purgeable_pool_locked(bo);
325
326 /* Release the purgeable lock while we're purging the BO so
327 * that other people can continue inserting things in the
328 * purgeable pool without having to wait for all BOs to be
329 * purged.
330 */
331 mutex_unlock(lock: &vc4->purgeable.lock);
332 mutex_lock(&bo->madv_lock);
333
334 /* Since we released the purgeable pool lock before acquiring
335 * the BO madv one, the user may have marked the BO as WILLNEED
336 * and re-used it in the meantime.
337 * Before purging the BO we need to make sure
338 * - it is still marked as DONTNEED
339 * - it has not been re-inserted in the purgeable list
340 * - it is not used by HW blocks
341 * If one of these conditions is not met, just skip the entry.
342 */
343 if (bo->madv == VC4_MADV_DONTNEED &&
344 list_empty(head: &bo->size_head) &&
345 !refcount_read(r: &bo->usecnt)) {
346 purged_size = bo->base.base.size;
347 vc4_bo_purge(obj);
348 }
349 mutex_unlock(lock: &bo->madv_lock);
350 mutex_lock(&vc4->purgeable.lock);
351
352 if (purged_size) {
353 vc4->purgeable.purged_size += purged_size;
354 vc4->purgeable.purged_num++;
355 }
356 }
357 mutex_unlock(lock: &vc4->purgeable.lock);
358}
359
360static struct vc4_bo *vc4_bo_get_from_cache(struct drm_device *dev,
361 uint32_t size,
362 enum vc4_kernel_bo_type type)
363{
364 struct vc4_dev *vc4 = to_vc4_dev(dev);
365 uint32_t page_index = bo_page_index(size);
366 struct vc4_bo *bo = NULL;
367
368 mutex_lock(&vc4->bo_lock);
369 if (page_index >= vc4->bo_cache.size_list_size)
370 goto out;
371
372 if (list_empty(head: &vc4->bo_cache.size_list[page_index]))
373 goto out;
374
375 bo = list_first_entry(&vc4->bo_cache.size_list[page_index],
376 struct vc4_bo, size_head);
377 vc4_bo_remove_from_cache(bo);
378 kref_init(kref: &bo->base.base.refcount);
379
380out:
381 if (bo)
382 vc4_bo_set_label(gem_obj: &bo->base.base, label: type);
383 mutex_unlock(lock: &vc4->bo_lock);
384 return bo;
385}
386
387/**
388 * vc4_create_object - Implementation of driver->gem_create_object.
389 * @dev: DRM device
390 * @size: Size in bytes of the memory the object will reference
391 *
392 * This lets the DMA helpers allocate object structs for us, and keep
393 * our BO stats correct.
394 */
395struct drm_gem_object *vc4_create_object(struct drm_device *dev, size_t size)
396{
397 struct vc4_dev *vc4 = to_vc4_dev(dev);
398 struct vc4_bo *bo;
399
400 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4))
401 return ERR_PTR(error: -ENODEV);
402
403 bo = kzalloc(sizeof(*bo), GFP_KERNEL);
404 if (!bo)
405 return ERR_PTR(error: -ENOMEM);
406
407 bo->madv = VC4_MADV_WILLNEED;
408 refcount_set(r: &bo->usecnt, n: 0);
409
410 mutex_init(&bo->madv_lock);
411
412 mutex_lock(&vc4->bo_lock);
413 bo->label = VC4_BO_TYPE_KERNEL;
414 vc4->bo_labels[VC4_BO_TYPE_KERNEL].num_allocated++;
415 vc4->bo_labels[VC4_BO_TYPE_KERNEL].size_allocated += size;
416 mutex_unlock(lock: &vc4->bo_lock);
417
418 bo->base.base.funcs = &vc4_gem_object_funcs;
419
420 return &bo->base.base;
421}
422
423struct vc4_bo *vc4_bo_create(struct drm_device *dev, size_t unaligned_size,
424 bool allow_unzeroed, enum vc4_kernel_bo_type type)
425{
426 size_t size = roundup(unaligned_size, PAGE_SIZE);
427 struct vc4_dev *vc4 = to_vc4_dev(dev);
428 struct drm_gem_dma_object *dma_obj;
429 struct vc4_bo *bo;
430
431 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4))
432 return ERR_PTR(error: -ENODEV);
433
434 if (size == 0)
435 return ERR_PTR(error: -EINVAL);
436
437 /* First, try to get a vc4_bo from the kernel BO cache. */
438 bo = vc4_bo_get_from_cache(dev, size, type);
439 if (bo) {
440 if (!allow_unzeroed)
441 memset(bo->base.vaddr, 0, bo->base.base.size);
442 return bo;
443 }
444
445 dma_obj = drm_gem_dma_create(drm: dev, size);
446 if (IS_ERR(ptr: dma_obj)) {
447 /*
448 * If we've run out of DMA memory, kill the cache of
449 * DMA allocations we've got laying around and try again.
450 */
451 vc4_bo_cache_purge(dev);
452 dma_obj = drm_gem_dma_create(drm: dev, size);
453 }
454
455 if (IS_ERR(ptr: dma_obj)) {
456 /*
457 * Still not enough DMA memory, purge the userspace BO
458 * cache and retry.
459 * This is sub-optimal since we purge the whole userspace
460 * BO cache which forces user that want to re-use the BO to
461 * restore its initial content.
462 * Ideally, we should purge entries one by one and retry
463 * after each to see if DMA allocation succeeds. Or even
464 * better, try to find an entry with at least the same
465 * size.
466 */
467 vc4_bo_userspace_cache_purge(dev);
468 dma_obj = drm_gem_dma_create(drm: dev, size);
469 }
470
471 if (IS_ERR(ptr: dma_obj)) {
472 struct drm_printer p = drm_info_printer(dev: vc4->base.dev);
473 drm_err(dev, "Failed to allocate from GEM DMA helper:\n");
474 vc4_bo_stats_print(p: &p, vc4);
475 return ERR_PTR(error: -ENOMEM);
476 }
477 bo = to_vc4_bo(&dma_obj->base);
478
479 /* By default, BOs do not support the MADV ioctl. This will be enabled
480 * only on BOs that are exposed to userspace (V3D, V3D_SHADER and DUMB
481 * BOs).
482 */
483 bo->madv = __VC4_MADV_NOTSUPP;
484
485 mutex_lock(&vc4->bo_lock);
486 vc4_bo_set_label(gem_obj: &dma_obj->base, label: type);
487 mutex_unlock(lock: &vc4->bo_lock);
488
489 return bo;
490}
491
492int vc4_bo_dumb_create(struct drm_file *file_priv,
493 struct drm_device *dev,
494 struct drm_mode_create_dumb *args)
495{
496 struct vc4_dev *vc4 = to_vc4_dev(dev);
497 struct vc4_bo *bo = NULL;
498 int ret;
499
500 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4))
501 return -ENODEV;
502
503 ret = vc4_dumb_fixup_args(args);
504 if (ret)
505 return ret;
506
507 bo = vc4_bo_create(dev, unaligned_size: args->size, allow_unzeroed: false, type: VC4_BO_TYPE_DUMB);
508 if (IS_ERR(ptr: bo))
509 return PTR_ERR(ptr: bo);
510
511 bo->madv = VC4_MADV_WILLNEED;
512
513 ret = drm_gem_handle_create(file_priv, obj: &bo->base.base, handlep: &args->handle);
514 drm_gem_object_put(obj: &bo->base.base);
515
516 return ret;
517}
518
519static void vc4_bo_cache_free_old(struct drm_device *dev)
520{
521 struct vc4_dev *vc4 = to_vc4_dev(dev);
522 unsigned long expire_time = jiffies - msecs_to_jiffies(m: 1000);
523
524 lockdep_assert_held(&vc4->bo_lock);
525
526 while (!list_empty(head: &vc4->bo_cache.time_list)) {
527 struct vc4_bo *bo = list_last_entry(&vc4->bo_cache.time_list,
528 struct vc4_bo, unref_head);
529 if (time_before(expire_time, bo->free_time)) {
530 mod_timer(timer: &vc4->bo_cache.time_timer,
531 expires: round_jiffies_up(j: jiffies +
532 msecs_to_jiffies(m: 1000)));
533 return;
534 }
535
536 vc4_bo_remove_from_cache(bo);
537 vc4_bo_destroy(bo);
538 }
539}
540
541/* Called on the last userspace/kernel unreference of the BO. Returns
542 * it to the BO cache if possible, otherwise frees it.
543 */
544static void vc4_free_object(struct drm_gem_object *gem_bo)
545{
546 struct drm_device *dev = gem_bo->dev;
547 struct vc4_dev *vc4 = to_vc4_dev(dev);
548 struct vc4_bo *bo = to_vc4_bo(gem_bo);
549 struct list_head *cache_list;
550
551 /* Remove the BO from the purgeable list. */
552 mutex_lock(&bo->madv_lock);
553 if (bo->madv == VC4_MADV_DONTNEED && !refcount_read(r: &bo->usecnt))
554 vc4_bo_remove_from_purgeable_pool(bo);
555 mutex_unlock(lock: &bo->madv_lock);
556
557 mutex_lock(&vc4->bo_lock);
558 /* If the object references someone else's memory, we can't cache it.
559 */
560 if (gem_bo->import_attach) {
561 vc4_bo_destroy(bo);
562 goto out;
563 }
564
565 /* Don't cache if it was publicly named. */
566 if (gem_bo->name) {
567 vc4_bo_destroy(bo);
568 goto out;
569 }
570
571 /* If this object was partially constructed but DMA allocation
572 * had failed, just free it. Can also happen when the BO has been
573 * purged.
574 */
575 if (!bo->base.vaddr) {
576 vc4_bo_destroy(bo);
577 goto out;
578 }
579
580 cache_list = vc4_get_cache_list_for_size(dev, size: gem_bo->size);
581 if (!cache_list) {
582 vc4_bo_destroy(bo);
583 goto out;
584 }
585
586 if (bo->validated_shader) {
587 kfree(objp: bo->validated_shader->uniform_addr_offsets);
588 kfree(objp: bo->validated_shader->texture_samples);
589 kfree(objp: bo->validated_shader);
590 bo->validated_shader = NULL;
591 }
592
593 /* Reset madv and usecnt before adding the BO to the cache. */
594 bo->madv = __VC4_MADV_NOTSUPP;
595 refcount_set(r: &bo->usecnt, n: 0);
596
597 bo->t_format = false;
598 bo->free_time = jiffies;
599 list_add(new: &bo->size_head, head: cache_list);
600 list_add(new: &bo->unref_head, head: &vc4->bo_cache.time_list);
601
602 vc4_bo_set_label(gem_obj: &bo->base.base, label: VC4_BO_TYPE_KERNEL_CACHE);
603
604 vc4_bo_cache_free_old(dev);
605
606out:
607 mutex_unlock(lock: &vc4->bo_lock);
608}
609
610static void vc4_bo_cache_time_work(struct work_struct *work)
611{
612 struct vc4_dev *vc4 =
613 container_of(work, struct vc4_dev, bo_cache.time_work);
614 struct drm_device *dev = &vc4->base;
615
616 mutex_lock(&vc4->bo_lock);
617 vc4_bo_cache_free_old(dev);
618 mutex_unlock(lock: &vc4->bo_lock);
619}
620
621int vc4_bo_inc_usecnt(struct vc4_bo *bo)
622{
623 struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev);
624 int ret;
625
626 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4))
627 return -ENODEV;
628
629 /* Fast path: if the BO is already retained by someone, no need to
630 * check the madv status.
631 */
632 if (refcount_inc_not_zero(r: &bo->usecnt))
633 return 0;
634
635 mutex_lock(&bo->madv_lock);
636 switch (bo->madv) {
637 case VC4_MADV_WILLNEED:
638 if (!refcount_inc_not_zero(r: &bo->usecnt))
639 refcount_set(r: &bo->usecnt, n: 1);
640 ret = 0;
641 break;
642 case VC4_MADV_DONTNEED:
643 /* We shouldn't use a BO marked as purgeable if at least
644 * someone else retained its content by incrementing usecnt.
645 * Luckily the BO hasn't been purged yet, but something wrong
646 * is happening here. Just throw an error instead of
647 * authorizing this use case.
648 */
649 case __VC4_MADV_PURGED:
650 /* We can't use a purged BO. */
651 default:
652 /* Invalid madv value. */
653 ret = -EINVAL;
654 break;
655 }
656 mutex_unlock(lock: &bo->madv_lock);
657
658 return ret;
659}
660
661void vc4_bo_dec_usecnt(struct vc4_bo *bo)
662{
663 struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev);
664
665 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4))
666 return;
667
668 /* Fast path: if the BO is still retained by someone, no need to test
669 * the madv value.
670 */
671 if (refcount_dec_not_one(r: &bo->usecnt))
672 return;
673
674 mutex_lock(&bo->madv_lock);
675 if (refcount_dec_and_test(r: &bo->usecnt) &&
676 bo->madv == VC4_MADV_DONTNEED)
677 vc4_bo_add_to_purgeable_pool(bo);
678 mutex_unlock(lock: &bo->madv_lock);
679}
680
681static void vc4_bo_cache_time_timer(struct timer_list *t)
682{
683 struct vc4_dev *vc4 = timer_container_of(vc4, t, bo_cache.time_timer);
684
685 schedule_work(work: &vc4->bo_cache.time_work);
686}
687
688static struct dma_buf *vc4_prime_export(struct drm_gem_object *obj, int flags)
689{
690 struct vc4_bo *bo = to_vc4_bo(obj);
691 struct dma_buf *dmabuf;
692 int ret;
693
694 if (bo->validated_shader) {
695 DRM_DEBUG("Attempting to export shader BO\n");
696 return ERR_PTR(error: -EINVAL);
697 }
698
699 /* Note: as soon as the BO is exported it becomes unpurgeable, because
700 * noone ever decrements the usecnt even if the reference held by the
701 * exported BO is released. This shouldn't be a problem since we don't
702 * expect exported BOs to be marked as purgeable.
703 */
704 ret = vc4_bo_inc_usecnt(bo);
705 if (ret) {
706 drm_err(obj->dev, "Failed to increment BO usecnt\n");
707 return ERR_PTR(error: ret);
708 }
709
710 dmabuf = drm_gem_prime_export(obj, flags);
711 if (IS_ERR(ptr: dmabuf))
712 vc4_bo_dec_usecnt(bo);
713
714 return dmabuf;
715}
716
717static vm_fault_t vc4_fault(struct vm_fault *vmf)
718{
719 struct vm_area_struct *vma = vmf->vma;
720 struct drm_gem_object *obj = vma->vm_private_data;
721 struct vc4_bo *bo = to_vc4_bo(obj);
722
723 /* The only reason we would end up here is when user-space accesses
724 * BO's memory after it's been purged.
725 */
726 mutex_lock(&bo->madv_lock);
727 WARN_ON(bo->madv != __VC4_MADV_PURGED);
728 mutex_unlock(lock: &bo->madv_lock);
729
730 return VM_FAULT_SIGBUS;
731}
732
733static int vc4_gem_object_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma)
734{
735 struct vc4_bo *bo = to_vc4_bo(obj);
736
737 if (bo->validated_shader && (vma->vm_flags & VM_WRITE)) {
738 DRM_DEBUG("mmapping of shader BOs for writing not allowed.\n");
739 return -EINVAL;
740 }
741
742 if (bo->madv != VC4_MADV_WILLNEED) {
743 DRM_DEBUG("mmapping of %s BO not allowed\n",
744 bo->madv == VC4_MADV_DONTNEED ?
745 "purgeable" : "purged");
746 return -EINVAL;
747 }
748
749 return drm_gem_dma_mmap(dma_obj: &bo->base, vma);
750}
751
752static const struct vm_operations_struct vc4_vm_ops = {
753 .fault = vc4_fault,
754 .open = drm_gem_vm_open,
755 .close = drm_gem_vm_close,
756};
757
758static const struct drm_gem_object_funcs vc4_gem_object_funcs = {
759 .free = vc4_free_object,
760 .export = vc4_prime_export,
761 .get_sg_table = drm_gem_dma_object_get_sg_table,
762 .vmap = drm_gem_dma_object_vmap,
763 .mmap = vc4_gem_object_mmap,
764 .vm_ops = &vc4_vm_ops,
765};
766
767static int vc4_grab_bin_bo(struct vc4_dev *vc4, struct vc4_file *vc4file)
768{
769 if (!vc4->v3d)
770 return -ENODEV;
771
772 if (vc4file->bin_bo_used)
773 return 0;
774
775 return vc4_v3d_bin_bo_get(vc4, used: &vc4file->bin_bo_used);
776}
777
778int vc4_create_bo_ioctl(struct drm_device *dev, void *data,
779 struct drm_file *file_priv)
780{
781 struct drm_vc4_create_bo *args = data;
782 struct vc4_file *vc4file = file_priv->driver_priv;
783 struct vc4_dev *vc4 = to_vc4_dev(dev);
784 struct vc4_bo *bo = NULL;
785 int ret;
786
787 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4))
788 return -ENODEV;
789
790 ret = vc4_grab_bin_bo(vc4, vc4file);
791 if (ret)
792 return ret;
793
794 /*
795 * We can't allocate from the BO cache, because the BOs don't
796 * get zeroed, and that might leak data between users.
797 */
798 bo = vc4_bo_create(dev, unaligned_size: args->size, allow_unzeroed: false, type: VC4_BO_TYPE_V3D);
799 if (IS_ERR(ptr: bo))
800 return PTR_ERR(ptr: bo);
801
802 bo->madv = VC4_MADV_WILLNEED;
803
804 ret = drm_gem_handle_create(file_priv, obj: &bo->base.base, handlep: &args->handle);
805 drm_gem_object_put(obj: &bo->base.base);
806
807 return ret;
808}
809
810int vc4_mmap_bo_ioctl(struct drm_device *dev, void *data,
811 struct drm_file *file_priv)
812{
813 struct vc4_dev *vc4 = to_vc4_dev(dev);
814 struct drm_vc4_mmap_bo *args = data;
815 struct drm_gem_object *gem_obj;
816
817 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4))
818 return -ENODEV;
819
820 gem_obj = drm_gem_object_lookup(filp: file_priv, handle: args->handle);
821 if (!gem_obj) {
822 DRM_DEBUG("Failed to look up GEM BO %d\n", args->handle);
823 return -EINVAL;
824 }
825
826 /* The mmap offset was set up at BO allocation time. */
827 args->offset = drm_vma_node_offset_addr(node: &gem_obj->vma_node);
828
829 drm_gem_object_put(obj: gem_obj);
830 return 0;
831}
832
833int
834vc4_create_shader_bo_ioctl(struct drm_device *dev, void *data,
835 struct drm_file *file_priv)
836{
837 struct drm_vc4_create_shader_bo *args = data;
838 struct vc4_file *vc4file = file_priv->driver_priv;
839 struct vc4_dev *vc4 = to_vc4_dev(dev);
840 struct vc4_bo *bo = NULL;
841 int ret;
842
843 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4))
844 return -ENODEV;
845
846 if (args->size == 0)
847 return -EINVAL;
848
849 if (args->size % sizeof(u64) != 0)
850 return -EINVAL;
851
852 if (args->flags != 0) {
853 DRM_INFO("Unknown flags set: 0x%08x\n", args->flags);
854 return -EINVAL;
855 }
856
857 if (args->pad != 0) {
858 DRM_INFO("Pad set: 0x%08x\n", args->pad);
859 return -EINVAL;
860 }
861
862 ret = vc4_grab_bin_bo(vc4, vc4file);
863 if (ret)
864 return ret;
865
866 bo = vc4_bo_create(dev, unaligned_size: args->size, allow_unzeroed: true, type: VC4_BO_TYPE_V3D_SHADER);
867 if (IS_ERR(ptr: bo))
868 return PTR_ERR(ptr: bo);
869
870 bo->madv = VC4_MADV_WILLNEED;
871
872 if (copy_from_user(to: bo->base.vaddr,
873 from: (void __user *)(uintptr_t)args->data,
874 n: args->size)) {
875 ret = -EFAULT;
876 goto fail;
877 }
878 /* Clear the rest of the memory from allocating from the BO
879 * cache.
880 */
881 memset(bo->base.vaddr + args->size, 0,
882 bo->base.base.size - args->size);
883
884 bo->validated_shader = vc4_validate_shader(shader_obj: &bo->base);
885 if (!bo->validated_shader) {
886 ret = -EINVAL;
887 goto fail;
888 }
889
890 /* We have to create the handle after validation, to avoid
891 * races for users to do doing things like mmap the shader BO.
892 */
893 ret = drm_gem_handle_create(file_priv, obj: &bo->base.base, handlep: &args->handle);
894
895fail:
896 drm_gem_object_put(obj: &bo->base.base);
897
898 return ret;
899}
900
901/**
902 * vc4_set_tiling_ioctl() - Sets the tiling modifier for a BO.
903 * @dev: DRM device
904 * @data: ioctl argument
905 * @file_priv: DRM file for this fd
906 *
907 * The tiling state of the BO decides the default modifier of an fb if
908 * no specific modifier was set by userspace, and the return value of
909 * vc4_get_tiling_ioctl() (so that userspace can treat a BO it
910 * received from dmabuf as the same tiling format as the producer
911 * used).
912 */
913int vc4_set_tiling_ioctl(struct drm_device *dev, void *data,
914 struct drm_file *file_priv)
915{
916 struct vc4_dev *vc4 = to_vc4_dev(dev);
917 struct drm_vc4_set_tiling *args = data;
918 struct drm_gem_object *gem_obj;
919 struct vc4_bo *bo;
920 bool t_format;
921
922 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4))
923 return -ENODEV;
924
925 if (args->flags != 0)
926 return -EINVAL;
927
928 switch (args->modifier) {
929 case DRM_FORMAT_MOD_NONE:
930 t_format = false;
931 break;
932 case DRM_FORMAT_MOD_BROADCOM_VC4_T_TILED:
933 t_format = true;
934 break;
935 default:
936 return -EINVAL;
937 }
938
939 gem_obj = drm_gem_object_lookup(filp: file_priv, handle: args->handle);
940 if (!gem_obj) {
941 DRM_DEBUG("Failed to look up GEM BO %d\n", args->handle);
942 return -ENOENT;
943 }
944 bo = to_vc4_bo(gem_obj);
945 bo->t_format = t_format;
946
947 drm_gem_object_put(obj: gem_obj);
948
949 return 0;
950}
951
952/**
953 * vc4_get_tiling_ioctl() - Gets the tiling modifier for a BO.
954 * @dev: DRM device
955 * @data: ioctl argument
956 * @file_priv: DRM file for this fd
957 *
958 * Returns the tiling modifier for a BO as set by vc4_set_tiling_ioctl().
959 */
960int vc4_get_tiling_ioctl(struct drm_device *dev, void *data,
961 struct drm_file *file_priv)
962{
963 struct vc4_dev *vc4 = to_vc4_dev(dev);
964 struct drm_vc4_get_tiling *args = data;
965 struct drm_gem_object *gem_obj;
966 struct vc4_bo *bo;
967
968 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4))
969 return -ENODEV;
970
971 if (args->flags != 0 || args->modifier != 0)
972 return -EINVAL;
973
974 gem_obj = drm_gem_object_lookup(filp: file_priv, handle: args->handle);
975 if (!gem_obj) {
976 DRM_DEBUG("Failed to look up GEM BO %d\n", args->handle);
977 return -ENOENT;
978 }
979 bo = to_vc4_bo(gem_obj);
980
981 if (bo->t_format)
982 args->modifier = DRM_FORMAT_MOD_BROADCOM_VC4_T_TILED;
983 else
984 args->modifier = DRM_FORMAT_MOD_NONE;
985
986 drm_gem_object_put(obj: gem_obj);
987
988 return 0;
989}
990
991int vc4_bo_debugfs_init(struct drm_minor *minor)
992{
993 struct drm_device *drm = minor->dev;
994 struct vc4_dev *vc4 = to_vc4_dev(drm);
995
996 if (!vc4->v3d)
997 return -ENODEV;
998
999 drm_debugfs_add_file(dev: drm, name: "bo_stats", show: vc4_bo_stats_debugfs, NULL);
1000
1001 return 0;
1002}
1003
1004static void vc4_bo_cache_destroy(struct drm_device *dev, void *unused);
1005int vc4_bo_cache_init(struct drm_device *dev)
1006{
1007 struct vc4_dev *vc4 = to_vc4_dev(dev);
1008 int ret;
1009 int i;
1010
1011 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4))
1012 return -ENODEV;
1013
1014 /* Create the initial set of BO labels that the kernel will
1015 * use. This lets us avoid a bunch of string reallocation in
1016 * the kernel's draw and BO allocation paths.
1017 */
1018 vc4->bo_labels = kcalloc(VC4_BO_TYPE_COUNT, sizeof(*vc4->bo_labels),
1019 GFP_KERNEL);
1020 if (!vc4->bo_labels)
1021 return -ENOMEM;
1022 vc4->num_labels = VC4_BO_TYPE_COUNT;
1023
1024 BUILD_BUG_ON(ARRAY_SIZE(bo_type_names) != VC4_BO_TYPE_COUNT);
1025 for (i = 0; i < VC4_BO_TYPE_COUNT; i++)
1026 vc4->bo_labels[i].name = bo_type_names[i];
1027
1028 ret = drmm_mutex_init(dev, &vc4->bo_lock);
1029 if (ret) {
1030 kfree(objp: vc4->bo_labels);
1031 return ret;
1032 }
1033
1034 INIT_LIST_HEAD(list: &vc4->bo_cache.time_list);
1035
1036 INIT_WORK(&vc4->bo_cache.time_work, vc4_bo_cache_time_work);
1037 timer_setup(&vc4->bo_cache.time_timer, vc4_bo_cache_time_timer, 0);
1038
1039 return drmm_add_action_or_reset(dev, vc4_bo_cache_destroy, NULL);
1040}
1041
1042static void vc4_bo_cache_destroy(struct drm_device *dev, void *unused)
1043{
1044 struct vc4_dev *vc4 = to_vc4_dev(dev);
1045 int i;
1046
1047 timer_delete(timer: &vc4->bo_cache.time_timer);
1048 cancel_work_sync(work: &vc4->bo_cache.time_work);
1049
1050 vc4_bo_cache_purge(dev);
1051
1052 for (i = 0; i < vc4->num_labels; i++) {
1053 if (vc4->bo_labels[i].num_allocated) {
1054 drm_err(dev, "Destroying BO cache with %d %s "
1055 "BOs still allocated\n",
1056 vc4->bo_labels[i].num_allocated,
1057 vc4->bo_labels[i].name);
1058 }
1059
1060 if (is_user_label(label: i))
1061 kfree(objp: vc4->bo_labels[i].name);
1062 }
1063 kfree(objp: vc4->bo_labels);
1064}
1065
1066int vc4_label_bo_ioctl(struct drm_device *dev, void *data,
1067 struct drm_file *file_priv)
1068{
1069 struct vc4_dev *vc4 = to_vc4_dev(dev);
1070 struct drm_vc4_label_bo *args = data;
1071 char *name;
1072 struct drm_gem_object *gem_obj;
1073 int ret = 0, label;
1074
1075 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4))
1076 return -ENODEV;
1077
1078 if (!args->len)
1079 return -EINVAL;
1080
1081 name = strndup_user(u64_to_user_ptr(args->name), args->len + 1);
1082 if (IS_ERR(ptr: name))
1083 return PTR_ERR(ptr: name);
1084
1085 gem_obj = drm_gem_object_lookup(filp: file_priv, handle: args->handle);
1086 if (!gem_obj) {
1087 drm_err(dev, "Failed to look up GEM BO %d\n", args->handle);
1088 kfree(objp: name);
1089 return -ENOENT;
1090 }
1091
1092 mutex_lock(&vc4->bo_lock);
1093 label = vc4_get_user_label(vc4, name);
1094 if (label != -1)
1095 vc4_bo_set_label(gem_obj, label);
1096 else
1097 ret = -ENOMEM;
1098 mutex_unlock(lock: &vc4->bo_lock);
1099
1100 drm_gem_object_put(obj: gem_obj);
1101
1102 return ret;
1103}
1104

source code of linux/drivers/gpu/drm/vc4/vc4_bo.c