xref: /netbsd-src/external/bsd/libevent/dist/epoll.c (revision bdc22b2e01993381dcefeff2bc9b56ca75a4235c)
1 /*	$NetBSD: epoll.c,v 1.1.1.2 2017/01/31 21:14:52 christos Exp $	*/
2 /*
3  * Copyright 2000-2007 Niels Provos <provos@citi.umich.edu>
4  * Copyright 2007-2012 Niels Provos, Nick Mathewson
5  *
6  * Redistribution and use in source and binary forms, with or without
7  * modification, are permitted provided that the following conditions
8  * are met:
9  * 1. Redistributions of source code must retain the above copyright
10  *    notice, this list of conditions and the following disclaimer.
11  * 2. Redistributions in binary form must reproduce the above copyright
12  *    notice, this list of conditions and the following disclaimer in the
13  *    documentation and/or other materials provided with the distribution.
14  * 3. The name of the author may not be used to endorse or promote products
15  *    derived from this software without specific prior written permission.
16  *
17  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
18  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
19  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
20  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
21  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
22  * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
26  * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27  */
28 #include "event2/event-config.h"
29 #include <sys/cdefs.h>
30 __RCSID("$NetBSD: epoll.c,v 1.1.1.2 2017/01/31 21:14:52 christos Exp $");
31 #include "evconfig-private.h"
32 
33 #ifdef EVENT__HAVE_EPOLL
34 
35 #include <stdint.h>
36 #include <sys/types.h>
37 #include <sys/resource.h>
38 #ifdef EVENT__HAVE_SYS_TIME_H
39 #include <sys/time.h>
40 #endif
41 #include <sys/queue.h>
42 #include <sys/epoll.h>
43 #include <signal.h>
44 #include <limits.h>
45 #include <stdio.h>
46 #include <stdlib.h>
47 #include <string.h>
48 #include <unistd.h>
49 #include <errno.h>
50 #ifdef EVENT__HAVE_FCNTL_H
51 #include <fcntl.h>
52 #endif
53 #ifdef EVENT__HAVE_SYS_TIMERFD_H
54 #include <sys/timerfd.h>
55 #endif
56 
57 #include "event-internal.h"
58 #include "evsignal-internal.h"
59 #include "event2/thread.h"
60 #include "evthread-internal.h"
61 #include "log-internal.h"
62 #include "evmap-internal.h"
63 #include "changelist-internal.h"
64 #include "time-internal.h"
65 
66 /* Since Linux 2.6.17, epoll is able to report about peer half-closed connection
67    using special EPOLLRDHUP flag on a read event.
68 */
69 #if !defined(EPOLLRDHUP)
70 #define EPOLLRDHUP 0
71 #define EARLY_CLOSE_IF_HAVE_RDHUP 0
72 #else
73 #define EARLY_CLOSE_IF_HAVE_RDHUP EV_FEATURE_EARLY_CLOSE
74 #endif
75 
76 #include "epolltable-internal.h"
77 
78 #if defined(EVENT__HAVE_SYS_TIMERFD_H) &&			  \
79 	defined(EVENT__HAVE_TIMERFD_CREATE) &&			  \
80 	defined(HAVE_POSIX_MONOTONIC) && defined(TFD_NONBLOCK) && \
81 	defined(TFD_CLOEXEC)
82 /* Note that we only use timerfd if TFD_NONBLOCK and TFD_CLOEXEC are available
83    and working.  This means that we can't support it on 2.6.25 (where timerfd
84    was introduced) or 2.6.26, since 2.6.27 introduced those flags.
85  */
86 #define USING_TIMERFD
87 #endif
88 
89 struct epollop {
90 	struct epoll_event *events;
91 	int nevents;
92 	int epfd;
93 #ifdef USING_TIMERFD
94 	int timerfd;
95 #endif
96 };
97 
98 static void *epoll_init(struct event_base *);
99 static int epoll_dispatch(struct event_base *, struct timeval *);
100 static void epoll_dealloc(struct event_base *);
101 
102 static const struct eventop epollops_changelist = {
103 	"epoll (with changelist)",
104 	epoll_init,
105 	event_changelist_add_,
106 	event_changelist_del_,
107 	epoll_dispatch,
108 	epoll_dealloc,
109 	1, /* need reinit */
110 	EV_FEATURE_ET|EV_FEATURE_O1| EARLY_CLOSE_IF_HAVE_RDHUP,
111 	EVENT_CHANGELIST_FDINFO_SIZE
112 };
113 
114 
115 static int epoll_nochangelist_add(struct event_base *base, evutil_socket_t fd,
116     short old, short events, void *p);
117 static int epoll_nochangelist_del(struct event_base *base, evutil_socket_t fd,
118     short old, short events, void *p);
119 
120 const struct eventop epollops = {
121 	"epoll",
122 	epoll_init,
123 	epoll_nochangelist_add,
124 	epoll_nochangelist_del,
125 	epoll_dispatch,
126 	epoll_dealloc,
127 	1, /* need reinit */
128 	EV_FEATURE_ET|EV_FEATURE_O1|EV_FEATURE_EARLY_CLOSE,
129 	0
130 };
131 
132 #define INITIAL_NEVENT 32
133 #define MAX_NEVENT 4096
134 
135 /* On Linux kernels at least up to 2.6.24.4, epoll can't handle timeout
136  * values bigger than (LONG_MAX - 999ULL)/HZ.  HZ in the wild can be
137  * as big as 1000, and LONG_MAX can be as small as (1<<31)-1, so the
138  * largest number of msec we can support here is 2147482.  Let's
139  * round that down by 47 seconds.
140  */
141 #define MAX_EPOLL_TIMEOUT_MSEC (35*60*1000)
142 
143 static void *
144 epoll_init(struct event_base *base)
145 {
146 	int epfd = -1;
147 	struct epollop *epollop;
148 
149 #ifdef EVENT__HAVE_EPOLL_CREATE1
150 	/* First, try the shiny new epoll_create1 interface, if we have it. */
151 	epfd = epoll_create1(EPOLL_CLOEXEC);
152 #endif
153 	if (epfd == -1) {
154 		/* Initialize the kernel queue using the old interface.  (The
155 		size field is ignored   since 2.6.8.) */
156 		if ((epfd = epoll_create(32000)) == -1) {
157 			if (errno != ENOSYS)
158 				event_warn("epoll_create");
159 			return (NULL);
160 		}
161 		evutil_make_socket_closeonexec(epfd);
162 	}
163 
164 	if (!(epollop = mm_calloc(1, sizeof(struct epollop)))) {
165 		close(epfd);
166 		return (NULL);
167 	}
168 
169 	epollop->epfd = epfd;
170 
171 	/* Initialize fields */
172 	epollop->events = mm_calloc(INITIAL_NEVENT, sizeof(struct epoll_event));
173 	if (epollop->events == NULL) {
174 		mm_free(epollop);
175 		close(epfd);
176 		return (NULL);
177 	}
178 	epollop->nevents = INITIAL_NEVENT;
179 
180 	if ((base->flags & EVENT_BASE_FLAG_EPOLL_USE_CHANGELIST) != 0 ||
181 	    ((base->flags & EVENT_BASE_FLAG_IGNORE_ENV) == 0 &&
182 		evutil_getenv_("EVENT_EPOLL_USE_CHANGELIST") != NULL)) {
183 
184 		base->evsel = &epollops_changelist;
185 	}
186 
187 #ifdef USING_TIMERFD
188 	/*
189 	  The epoll interface ordinarily gives us one-millisecond precision,
190 	  so on Linux it makes perfect sense to use the CLOCK_MONOTONIC_COARSE
191 	  timer.  But when the user has set the new PRECISE_TIMER flag for an
192 	  event_base, we can try to use timerfd to give them finer granularity.
193 	*/
194 	if ((base->flags & EVENT_BASE_FLAG_PRECISE_TIMER) &&
195 	    base->monotonic_timer.monotonic_clock == CLOCK_MONOTONIC) {
196 		int fd;
197 		fd = epollop->timerfd = timerfd_create(CLOCK_MONOTONIC, TFD_NONBLOCK|TFD_CLOEXEC);
198 		if (epollop->timerfd >= 0) {
199 			struct epoll_event epev;
200 			memset(&epev, 0, sizeof(epev));
201 			epev.data.fd = epollop->timerfd;
202 			epev.events = EPOLLIN;
203 			if (epoll_ctl(epollop->epfd, EPOLL_CTL_ADD, fd, &epev) < 0) {
204 				event_warn("epoll_ctl(timerfd)");
205 				close(fd);
206 				epollop->timerfd = -1;
207 			}
208 		} else {
209 			if (errno != EINVAL && errno != ENOSYS) {
210 				/* These errors probably mean that we were
211 				 * compiled with timerfd/TFD_* support, but
212 				 * we're running on a kernel that lacks those.
213 				 */
214 				event_warn("timerfd_create");
215 			}
216 			epollop->timerfd = -1;
217 		}
218 	} else {
219 		epollop->timerfd = -1;
220 	}
221 #endif
222 
223 	evsig_init_(base);
224 
225 	return (epollop);
226 }
227 
228 static const char *
229 change_to_string(int change)
230 {
231 	change &= (EV_CHANGE_ADD|EV_CHANGE_DEL);
232 	if (change == EV_CHANGE_ADD) {
233 		return "add";
234 	} else if (change == EV_CHANGE_DEL) {
235 		return "del";
236 	} else if (change == 0) {
237 		return "none";
238 	} else {
239 		return "???";
240 	}
241 }
242 
243 static const char *
244 epoll_op_to_string(int op)
245 {
246 	return op == EPOLL_CTL_ADD?"ADD":
247 	    op == EPOLL_CTL_DEL?"DEL":
248 	    op == EPOLL_CTL_MOD?"MOD":
249 	    "???";
250 }
251 
252 #define PRINT_CHANGES(op, events, ch, status)  \
253 	"Epoll %s(%d) on fd %d " status ". "       \
254 	"Old events were %d; "                     \
255 	"read change was %d (%s); "                \
256 	"write change was %d (%s); "               \
257 	"close change was %d (%s)",                \
258 	epoll_op_to_string(op),                    \
259 	events,                                    \
260 	ch->fd,                                    \
261 	ch->old_events,                            \
262 	ch->read_change,                           \
263 	change_to_string(ch->read_change),         \
264 	ch->write_change,                          \
265 	change_to_string(ch->write_change),        \
266 	ch->close_change,                          \
267 	change_to_string(ch->close_change)
268 
269 static int
270 epoll_apply_one_change(struct event_base *base,
271     struct epollop *epollop,
272     const struct event_change *ch)
273 {
274 	struct epoll_event epev;
275 	int op, events = 0;
276 	int idx;
277 
278 	idx = EPOLL_OP_TABLE_INDEX(ch);
279 	op = epoll_op_table[idx].op;
280 	events = epoll_op_table[idx].events;
281 
282 	if (!events) {
283 		EVUTIL_ASSERT(op == 0);
284 		return 0;
285 	}
286 
287 	if ((ch->read_change|ch->write_change) & EV_CHANGE_ET)
288 		events |= EPOLLET;
289 
290 	memset(&epev, 0, sizeof(epev));
291 	epev.data.fd = ch->fd;
292 	epev.events = events;
293 	if (epoll_ctl(epollop->epfd, op, ch->fd, &epev) == 0) {
294 		event_debug((PRINT_CHANGES(op, epev.events, ch, "okay")));
295 		return 0;
296 	}
297 
298 	switch (op) {
299 	case EPOLL_CTL_MOD:
300 		if (errno == ENOENT) {
301 			/* If a MOD operation fails with ENOENT, the
302 			 * fd was probably closed and re-opened.  We
303 			 * should retry the operation as an ADD.
304 			 */
305 			if (epoll_ctl(epollop->epfd, EPOLL_CTL_ADD, ch->fd, &epev) == -1) {
306 				event_warn("Epoll MOD(%d) on %d retried as ADD; that failed too",
307 				    (int)epev.events, ch->fd);
308 				return -1;
309 			} else {
310 				event_debug(("Epoll MOD(%d) on %d retried as ADD; succeeded.",
311 					(int)epev.events,
312 					ch->fd));
313 				return 0;
314 			}
315 		}
316 		break;
317 	case EPOLL_CTL_ADD:
318 		if (errno == EEXIST) {
319 			/* If an ADD operation fails with EEXIST,
320 			 * either the operation was redundant (as with a
321 			 * precautionary add), or we ran into a fun
322 			 * kernel bug where using dup*() to duplicate the
323 			 * same file into the same fd gives you the same epitem
324 			 * rather than a fresh one.  For the second case,
325 			 * we must retry with MOD. */
326 			if (epoll_ctl(epollop->epfd, EPOLL_CTL_MOD, ch->fd, &epev) == -1) {
327 				event_warn("Epoll ADD(%d) on %d retried as MOD; that failed too",
328 				    (int)epev.events, ch->fd);
329 				return -1;
330 			} else {
331 				event_debug(("Epoll ADD(%d) on %d retried as MOD; succeeded.",
332 					(int)epev.events,
333 					ch->fd));
334 				return 0;
335 			}
336 		}
337 		break;
338 	case EPOLL_CTL_DEL:
339 		if (errno == ENOENT || errno == EBADF || errno == EPERM) {
340 			/* If a delete fails with one of these errors,
341 			 * that's fine too: we closed the fd before we
342 			 * got around to calling epoll_dispatch. */
343 			event_debug(("Epoll DEL(%d) on fd %d gave %s: DEL was unnecessary.",
344 				(int)epev.events,
345 				ch->fd,
346 				strerror(errno)));
347 			return 0;
348 		}
349 		break;
350 	default:
351 		break;
352 	}
353 
354 	event_warn(PRINT_CHANGES(op, epev.events, ch, "failed"));
355 	return -1;
356 }
357 
358 static int
359 epoll_apply_changes(struct event_base *base)
360 {
361 	struct event_changelist *changelist = &base->changelist;
362 	struct epollop *epollop = base->evbase;
363 	struct event_change *ch;
364 
365 	int r = 0;
366 	int i;
367 
368 	for (i = 0; i < changelist->n_changes; ++i) {
369 		ch = &changelist->changes[i];
370 		if (epoll_apply_one_change(base, epollop, ch) < 0)
371 			r = -1;
372 	}
373 
374 	return (r);
375 }
376 
377 static int
378 epoll_nochangelist_add(struct event_base *base, evutil_socket_t fd,
379     short old, short events, void *p)
380 {
381 	struct event_change ch;
382 	ch.fd = fd;
383 	ch.old_events = old;
384 	ch.read_change = ch.write_change = ch.close_change = 0;
385 	if (events & EV_WRITE)
386 		ch.write_change = EV_CHANGE_ADD |
387 		    (events & EV_ET);
388 	if (events & EV_READ)
389 		ch.read_change = EV_CHANGE_ADD |
390 		    (events & EV_ET);
391 	if (events & EV_CLOSED)
392 		ch.close_change = EV_CHANGE_ADD |
393 		    (events & EV_ET);
394 
395 	return epoll_apply_one_change(base, base->evbase, &ch);
396 }
397 
398 static int
399 epoll_nochangelist_del(struct event_base *base, evutil_socket_t fd,
400     short old, short events, void *p)
401 {
402 	struct event_change ch;
403 	ch.fd = fd;
404 	ch.old_events = old;
405 	ch.read_change = ch.write_change = ch.close_change = 0;
406 	if (events & EV_WRITE)
407 		ch.write_change = EV_CHANGE_DEL;
408 	if (events & EV_READ)
409 		ch.read_change = EV_CHANGE_DEL;
410 	if (events & EV_CLOSED)
411 		ch.close_change = EV_CHANGE_DEL;
412 
413 	return epoll_apply_one_change(base, base->evbase, &ch);
414 }
415 
416 static int
417 epoll_dispatch(struct event_base *base, struct timeval *tv)
418 {
419 	struct epollop *epollop = base->evbase;
420 	struct epoll_event *events = epollop->events;
421 	int i, res;
422 	long timeout = -1;
423 
424 #ifdef USING_TIMERFD
425 	if (epollop->timerfd >= 0) {
426 		struct itimerspec is;
427 		is.it_interval.tv_sec = 0;
428 		is.it_interval.tv_nsec = 0;
429 		if (tv == NULL) {
430 			/* No timeout; disarm the timer. */
431 			is.it_value.tv_sec = 0;
432 			is.it_value.tv_nsec = 0;
433 		} else {
434 			if (tv->tv_sec == 0 && tv->tv_usec == 0) {
435 				/* we need to exit immediately; timerfd can't
436 				 * do that. */
437 				timeout = 0;
438 			}
439 			is.it_value.tv_sec = tv->tv_sec;
440 			is.it_value.tv_nsec = tv->tv_usec * 1000;
441 		}
442 		/* TODO: we could avoid unnecessary syscalls here by only
443 		   calling timerfd_settime when the top timeout changes, or
444 		   when we're called with a different timeval.
445 		*/
446 		if (timerfd_settime(epollop->timerfd, 0, &is, NULL) < 0) {
447 			event_warn("timerfd_settime");
448 		}
449 	} else
450 #endif
451 	if (tv != NULL) {
452 		timeout = evutil_tv_to_msec_(tv);
453 		if (timeout < 0 || timeout > MAX_EPOLL_TIMEOUT_MSEC) {
454 			/* Linux kernels can wait forever if the timeout is
455 			 * too big; see comment on MAX_EPOLL_TIMEOUT_MSEC. */
456 			timeout = MAX_EPOLL_TIMEOUT_MSEC;
457 		}
458 	}
459 
460 	epoll_apply_changes(base);
461 	event_changelist_remove_all_(&base->changelist, base);
462 
463 	EVBASE_RELEASE_LOCK(base, th_base_lock);
464 
465 	res = epoll_wait(epollop->epfd, events, epollop->nevents, timeout);
466 
467 	EVBASE_ACQUIRE_LOCK(base, th_base_lock);
468 
469 	if (res == -1) {
470 		if (errno != EINTR) {
471 			event_warn("epoll_wait");
472 			return (-1);
473 		}
474 
475 		return (0);
476 	}
477 
478 	event_debug(("%s: epoll_wait reports %d", __func__, res));
479 	EVUTIL_ASSERT(res <= epollop->nevents);
480 
481 	for (i = 0; i < res; i++) {
482 		int what = events[i].events;
483 		short ev = 0;
484 #ifdef USING_TIMERFD
485 		if (events[i].data.fd == epollop->timerfd)
486 			continue;
487 #endif
488 
489 		if (what & (EPOLLHUP|EPOLLERR)) {
490 			ev = EV_READ | EV_WRITE;
491 		} else {
492 			if (what & EPOLLIN)
493 				ev |= EV_READ;
494 			if (what & EPOLLOUT)
495 				ev |= EV_WRITE;
496 			if (what & EPOLLRDHUP)
497 				ev |= EV_CLOSED;
498 		}
499 
500 		if (!ev)
501 			continue;
502 
503 		evmap_io_active_(base, events[i].data.fd, ev | EV_ET);
504 	}
505 
506 	if (res == epollop->nevents && epollop->nevents < MAX_NEVENT) {
507 		/* We used all of the event space this time.  We should
508 		   be ready for more events next time. */
509 		int new_nevents = epollop->nevents * 2;
510 		struct epoll_event *new_events;
511 
512 		new_events = mm_realloc(epollop->events,
513 		    new_nevents * sizeof(struct epoll_event));
514 		if (new_events) {
515 			epollop->events = new_events;
516 			epollop->nevents = new_nevents;
517 		}
518 	}
519 
520 	return (0);
521 }
522 
523 
524 static void
525 epoll_dealloc(struct event_base *base)
526 {
527 	struct epollop *epollop = base->evbase;
528 
529 	evsig_dealloc_(base);
530 	if (epollop->events)
531 		mm_free(epollop->events);
532 	if (epollop->epfd >= 0)
533 		close(epollop->epfd);
534 #ifdef USING_TIMERFD
535 	if (epollop->timerfd >= 0)
536 		close(epollop->timerfd);
537 #endif
538 
539 	memset(epollop, 0, sizeof(struct epollop));
540 	mm_free(epollop);
541 }
542 
543 #endif /* EVENT__HAVE_EPOLL */
544