1 /* $OpenBSD: workqueue.h,v 1.8 2022/03/01 04:08:04 jsg Exp $ */ 2 /* 3 * Copyright (c) 2015 Mark Kettenis 4 * 5 * Permission to use, copy, modify, and distribute this software for any 6 * purpose with or without fee is hereby granted, provided that the above 7 * copyright notice and this permission notice appear in all copies. 8 * 9 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES 10 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF 11 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR 12 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES 13 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN 14 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF 15 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. 16 */ 17 18 #ifndef _LINUX_WORKQUEUE_H 19 #define _LINUX_WORKQUEUE_H 20 21 #include <sys/param.h> 22 #include <sys/systm.h> 23 #include <sys/task.h> 24 #include <sys/timeout.h> 25 #include <linux/bitops.h> 26 #include <linux/atomic.h> 27 #include <linux/rcupdate.h> 28 #include <linux/kernel.h> 29 #include <linux/lockdep.h> 30 #include <linux/timer.h> 31 32 struct workqueue_struct; 33 34 extern struct workqueue_struct *system_wq; 35 extern struct workqueue_struct *system_highpri_wq; 36 extern struct workqueue_struct *system_unbound_wq; 37 extern struct workqueue_struct *system_long_wq; 38 39 #define WQ_HIGHPRI 1 40 #define WQ_FREEZABLE 2 41 #define WQ_UNBOUND 4 42 43 #define WQ_UNBOUND_MAX_ACTIVE 4 /* matches nthreads in drm_linux.c */ 44 45 static inline struct workqueue_struct * 46 alloc_workqueue(const char *name, int flags, int max_active) 47 { 48 struct taskq *tq = taskq_create(name, 1, IPL_TTY, 0); 49 return (struct workqueue_struct *)tq; 50 } 51 52 static inline struct workqueue_struct * 53 alloc_ordered_workqueue(const char *name, int flags) 54 { 55 struct taskq *tq = taskq_create(name, 1, IPL_TTY, 0); 56 return (struct workqueue_struct *)tq; 57 } 58 59 static inline struct workqueue_struct * 60 create_singlethread_workqueue(const char *name) 61 { 62 struct taskq *tq = taskq_create(name, 1, IPL_TTY, 0); 63 return (struct workqueue_struct *)tq; 64 } 65 66 static inline void 67 destroy_workqueue(struct workqueue_struct *wq) 68 { 69 taskq_destroy((struct taskq *)wq); 70 } 71 72 struct work_struct { 73 struct task task; 74 struct taskq *tq; 75 }; 76 77 typedef void (*work_func_t)(struct work_struct *); 78 79 static inline void 80 INIT_WORK(struct work_struct *work, work_func_t func) 81 { 82 work->tq = NULL; 83 task_set(&work->task, (void (*)(void *))func, work); 84 } 85 86 #define INIT_WORK_ONSTACK(x, y) INIT_WORK((x), (y)) 87 88 static inline bool 89 queue_work(struct workqueue_struct *wq, struct work_struct *work) 90 { 91 work->tq = (struct taskq *)wq; 92 return task_add(work->tq, &work->task); 93 } 94 95 static inline void 96 cancel_work_sync(struct work_struct *work) 97 { 98 if (work->tq != NULL) 99 task_del(work->tq, &work->task); 100 } 101 102 #define work_pending(work) task_pending(&(work)->task) 103 104 struct delayed_work { 105 struct work_struct work; 106 struct timeout to; 107 struct taskq *tq; 108 }; 109 110 #define system_power_efficient_wq ((struct workqueue_struct *)systq) 111 112 static inline struct delayed_work * 113 to_delayed_work(struct work_struct *work) 114 { 115 return container_of(work, struct delayed_work, work); 116 } 117 118 static void 119 __delayed_work_tick(void *arg) 120 { 121 struct delayed_work *dwork = arg; 122 123 task_add(dwork->tq, &dwork->work.task); 124 } 125 126 static inline void 127 INIT_DELAYED_WORK(struct delayed_work *dwork, work_func_t func) 128 { 129 INIT_WORK(&dwork->work, func); 130 timeout_set(&dwork->to, __delayed_work_tick, &dwork->work); 131 } 132 133 static inline void 134 INIT_DELAYED_WORK_ONSTACK(struct delayed_work *dwork, work_func_t func) 135 { 136 INIT_WORK(&dwork->work, func); 137 timeout_set(&dwork->to, __delayed_work_tick, &dwork->work); 138 } 139 140 #define __DELAYED_WORK_INITIALIZER(dw, fn, flags) { \ 141 .to = TIMEOUT_INITIALIZER(__delayed_work_tick, &(dw)), \ 142 .tq = NULL, \ 143 .work.tq = NULL, \ 144 .work.task = TASK_INITIALIZER((void (*)(void *))(fn), &(dw).work) \ 145 } 146 147 static inline bool 148 schedule_work(struct work_struct *work) 149 { 150 work->tq = (struct taskq *)system_wq; 151 return task_add(work->tq, &work->task); 152 } 153 154 static inline bool 155 schedule_delayed_work(struct delayed_work *dwork, int jiffies) 156 { 157 dwork->tq = (struct taskq *)system_wq; 158 return timeout_add(&dwork->to, jiffies); 159 } 160 161 static inline bool 162 queue_delayed_work(struct workqueue_struct *wq, 163 struct delayed_work *dwork, int jiffies) 164 { 165 dwork->tq = (struct taskq *)wq; 166 return timeout_add(&dwork->to, jiffies); 167 } 168 169 static inline bool 170 mod_delayed_work(struct workqueue_struct *wq, 171 struct delayed_work *dwork, int jiffies) 172 { 173 dwork->tq = (struct taskq *)wq; 174 return (timeout_add(&dwork->to, jiffies) == 0); 175 } 176 177 static inline bool 178 cancel_delayed_work(struct delayed_work *dwork) 179 { 180 if (dwork->tq == NULL) 181 return false; 182 if (timeout_del(&dwork->to)) 183 return true; 184 return task_del(dwork->tq, &dwork->work.task); 185 } 186 187 static inline bool 188 cancel_delayed_work_sync(struct delayed_work *dwork) 189 { 190 if (dwork->tq == NULL) 191 return false; 192 if (timeout_del(&dwork->to)) 193 return true; 194 return task_del(dwork->tq, &dwork->work.task); 195 } 196 197 static inline bool 198 delayed_work_pending(struct delayed_work *dwork) 199 { 200 if (timeout_pending(&dwork->to)) 201 return true; 202 return task_pending(&dwork->work.task); 203 } 204 205 void flush_workqueue(struct workqueue_struct *); 206 bool flush_work(struct work_struct *); 207 bool flush_delayed_work(struct delayed_work *); 208 209 static inline void 210 flush_scheduled_work(void) 211 { 212 flush_workqueue(system_wq); 213 } 214 215 static inline void 216 drain_workqueue(struct workqueue_struct *wq) 217 { 218 flush_workqueue(wq); 219 } 220 221 static inline void 222 destroy_work_on_stack(struct work_struct *work) 223 { 224 if (work->tq) 225 task_del(work->tq, &work->task); 226 } 227 228 static inline void 229 destroy_delayed_work_on_stack(struct delayed_work *dwork) 230 { 231 } 232 233 struct rcu_work { 234 struct work_struct work; 235 struct rcu_head rcu; 236 }; 237 238 static inline void 239 INIT_RCU_WORK(struct rcu_work *work, work_func_t func) 240 { 241 INIT_WORK(&work->work, func); 242 } 243 244 static inline bool 245 queue_rcu_work(struct workqueue_struct *wq, struct rcu_work *work) 246 { 247 return queue_work(wq, &work->work); 248 } 249 250 #endif 251