| 1 | /* SPDX-License-Identifier: MIT */ |
| 2 | /* Copyright © 2025 Intel Corporation */ |
| 3 | |
| 4 | #ifndef __I915_WAIT_UTIL_H__ |
| 5 | #define __I915_WAIT_UTIL_H__ |
| 6 | |
| 7 | #include <linux/compiler.h> |
| 8 | #include <linux/delay.h> |
| 9 | #include <linux/ktime.h> |
| 10 | #include <linux/sched/clock.h> |
| 11 | #include <linux/smp.h> |
| 12 | |
| 13 | /* |
| 14 | * __wait_for - magic wait macro |
| 15 | * |
| 16 | * Macro to help avoid open coding check/wait/timeout patterns. Note that it's |
| 17 | * important that we check the condition again after having timed out, since the |
| 18 | * timeout could be due to preemption or similar and we've never had a chance to |
| 19 | * check the condition before the timeout. |
| 20 | */ |
| 21 | #define __wait_for(OP, COND, US, Wmin, Wmax) ({ \ |
| 22 | const ktime_t end__ = ktime_add_ns(ktime_get_raw(), 1000ll * (US)); \ |
| 23 | long wait__ = (Wmin); /* recommended min for usleep is 10 us */ \ |
| 24 | int ret__; \ |
| 25 | might_sleep(); \ |
| 26 | for (;;) { \ |
| 27 | const bool expired__ = ktime_after(ktime_get_raw(), end__); \ |
| 28 | OP; \ |
| 29 | /* Guarantee COND check prior to timeout */ \ |
| 30 | barrier(); \ |
| 31 | if (COND) { \ |
| 32 | ret__ = 0; \ |
| 33 | break; \ |
| 34 | } \ |
| 35 | if (expired__) { \ |
| 36 | ret__ = -ETIMEDOUT; \ |
| 37 | break; \ |
| 38 | } \ |
| 39 | usleep_range(wait__, wait__ * 2); \ |
| 40 | if (wait__ < (Wmax)) \ |
| 41 | wait__ <<= 1; \ |
| 42 | } \ |
| 43 | ret__; \ |
| 44 | }) |
| 45 | |
| 46 | #define _wait_for(COND, US, Wmin, Wmax) __wait_for(, (COND), (US), (Wmin), \ |
| 47 | (Wmax)) |
| 48 | #define wait_for(COND, MS) _wait_for((COND), (MS) * 1000, 10, 1000) |
| 49 | |
| 50 | /* |
| 51 | * If CONFIG_PREEMPT_COUNT is disabled, in_atomic() always reports false. |
| 52 | * On PREEMPT_RT the context isn't becoming atomic because it is used in an |
| 53 | * interrupt handler or because a spinlock_t is acquired. This leads to |
| 54 | * warnings which don't occur otherwise and therefore the check is disabled. |
| 55 | */ |
| 56 | #if IS_ENABLED(CONFIG_DRM_I915_DEBUG) && IS_ENABLED(CONFIG_PREEMPT_COUNT) && !defined(CONFIG_PREEMPT_RT) |
| 57 | # define _WAIT_FOR_ATOMIC_CHECK(ATOMIC) WARN_ON_ONCE((ATOMIC) && !in_atomic()) |
| 58 | #else |
| 59 | # define _WAIT_FOR_ATOMIC_CHECK(ATOMIC) do { } while (0) |
| 60 | #endif |
| 61 | |
| 62 | #define _wait_for_atomic(COND, US, ATOMIC) \ |
| 63 | ({ \ |
| 64 | int cpu, ret, timeout = (US) * 1000; \ |
| 65 | u64 base; \ |
| 66 | _WAIT_FOR_ATOMIC_CHECK(ATOMIC); \ |
| 67 | if (!(ATOMIC)) { \ |
| 68 | preempt_disable(); \ |
| 69 | cpu = smp_processor_id(); \ |
| 70 | } \ |
| 71 | base = local_clock(); \ |
| 72 | for (;;) { \ |
| 73 | u64 now = local_clock(); \ |
| 74 | if (!(ATOMIC)) \ |
| 75 | preempt_enable(); \ |
| 76 | /* Guarantee COND check prior to timeout */ \ |
| 77 | barrier(); \ |
| 78 | if (COND) { \ |
| 79 | ret = 0; \ |
| 80 | break; \ |
| 81 | } \ |
| 82 | if (now - base >= timeout) { \ |
| 83 | ret = -ETIMEDOUT; \ |
| 84 | break; \ |
| 85 | } \ |
| 86 | cpu_relax(); \ |
| 87 | if (!(ATOMIC)) { \ |
| 88 | preempt_disable(); \ |
| 89 | if (unlikely(cpu != smp_processor_id())) { \ |
| 90 | timeout -= now - base; \ |
| 91 | cpu = smp_processor_id(); \ |
| 92 | base = local_clock(); \ |
| 93 | } \ |
| 94 | } \ |
| 95 | } \ |
| 96 | ret; \ |
| 97 | }) |
| 98 | |
| 99 | #define wait_for_us(COND, US) \ |
| 100 | ({ \ |
| 101 | int ret__; \ |
| 102 | BUILD_BUG_ON(!__builtin_constant_p(US)); \ |
| 103 | if ((US) > 10) \ |
| 104 | ret__ = _wait_for((COND), (US), 10, 10); \ |
| 105 | else \ |
| 106 | ret__ = _wait_for_atomic((COND), (US), 0); \ |
| 107 | ret__; \ |
| 108 | }) |
| 109 | |
| 110 | #define wait_for_atomic_us(COND, US) \ |
| 111 | ({ \ |
| 112 | BUILD_BUG_ON(!__builtin_constant_p(US)); \ |
| 113 | BUILD_BUG_ON((US) > 50000); \ |
| 114 | _wait_for_atomic((COND), (US), 1); \ |
| 115 | }) |
| 116 | |
| 117 | #define wait_for_atomic(COND, MS) wait_for_atomic_us((COND), (MS) * 1000) |
| 118 | |
| 119 | #endif /* __I915_WAIT_UTIL_H__ */ |
| 120 | |