tor-browser

The Tor Browser
git clone https://git.dasho.dev/tor-browser.git
Log | Files | Refs | README | LICENSE

epoll.c (14881B)


      1 /*
      2 * Copyright 2000-2007 Niels Provos <provos@citi.umich.edu>
      3 * Copyright 2007-2012 Niels Provos, Nick Mathewson
      4 *
      5 * Redistribution and use in source and binary forms, with or without
      6 * modification, are permitted provided that the following conditions
      7 * are met:
      8 * 1. Redistributions of source code must retain the above copyright
      9 *    notice, this list of conditions and the following disclaimer.
     10 * 2. Redistributions in binary form must reproduce the above copyright
     11 *    notice, this list of conditions and the following disclaimer in the
     12 *    documentation and/or other materials provided with the distribution.
     13 * 3. The name of the author may not be used to endorse or promote products
     14 *    derived from this software without specific prior written permission.
     15 *
     16 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
     17 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
     18 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
     19 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
     20 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
     21 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
     22 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
     23 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
     24 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
     25 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
     26 */
     27 #include "event2/event-config.h"
     28 #include "evconfig-private.h"
     29 
     30 #ifdef EVENT__HAVE_EPOLL
     31 
     32 #include <stdint.h>
     33 #include <sys/types.h>
     34 #include <sys/resource.h>
     35 #ifdef EVENT__HAVE_SYS_TIME_H
     36 #include <sys/time.h>
     37 #endif
     38 #include <sys/queue.h>
     39 #include <sys/epoll.h>
     40 #include <signal.h>
     41 #include <limits.h>
     42 #include <stdio.h>
     43 #include <stdlib.h>
     44 #include <string.h>
     45 #include <unistd.h>
     46 #include <errno.h>
     47 #ifdef EVENT__HAVE_FCNTL_H
     48 #include <fcntl.h>
     49 #endif
     50 #ifdef EVENT__HAVE_SYS_TIMERFD_H
     51 #include <sys/timerfd.h>
     52 #endif
     53 
     54 #include "event-internal.h"
     55 #include "evsignal-internal.h"
     56 #include "event2/thread.h"
     57 #include "evthread-internal.h"
     58 #include "log-internal.h"
     59 #include "evmap-internal.h"
     60 #include "changelist-internal.h"
     61 #include "time-internal.h"
     62 
     63 /* Since Linux 2.6.17, epoll is able to report about peer half-closed connection
     64   using special EPOLLRDHUP flag on a read event.
     65 */
     66 #if !defined(EPOLLRDHUP)
     67 #define EPOLLRDHUP 0
     68 #define EARLY_CLOSE_IF_HAVE_RDHUP 0
     69 #else
     70 #define EARLY_CLOSE_IF_HAVE_RDHUP EV_FEATURE_EARLY_CLOSE
     71 #endif
     72 
     73 #include "epolltable-internal.h"
     74 
     75 #if defined(EVENT__HAVE_SYS_TIMERFD_H) &&			  \
     76 defined(EVENT__HAVE_TIMERFD_CREATE) &&			  \
     77 defined(HAVE_POSIX_MONOTONIC) && defined(TFD_NONBLOCK) && \
     78 defined(TFD_CLOEXEC)
     79 /* Note that we only use timerfd if TFD_NONBLOCK and TFD_CLOEXEC are available
     80   and working.  This means that we can't support it on 2.6.25 (where timerfd
     81   was introduced) or 2.6.26, since 2.6.27 introduced those flags.
     82 */
     83 #define USING_TIMERFD
     84 #endif
     85 
     86 struct epollop {
     87 struct epoll_event *events;
     88 int nevents;
     89 int epfd;
     90 #ifdef USING_TIMERFD
     91 int timerfd;
     92 #endif
     93 };
     94 
     95 static void *epoll_init(struct event_base *);
     96 static int epoll_dispatch(struct event_base *, struct timeval *);
     97 static void epoll_dealloc(struct event_base *);
     98 
     99 static const struct eventop epollops_changelist = {
    100 "epoll (with changelist)",
    101 epoll_init,
    102 event_changelist_add_,
    103 event_changelist_del_,
    104 epoll_dispatch,
    105 epoll_dealloc,
    106 1, /* need reinit */
    107 EV_FEATURE_ET|EV_FEATURE_O1| EARLY_CLOSE_IF_HAVE_RDHUP,
    108 EVENT_CHANGELIST_FDINFO_SIZE
    109 };
    110 
    111 
    112 static int epoll_nochangelist_add(struct event_base *base, evutil_socket_t fd,
    113    short old, short events, void *p);
    114 static int epoll_nochangelist_del(struct event_base *base, evutil_socket_t fd,
    115    short old, short events, void *p);
    116 
    117 const struct eventop epollops = {
    118 "epoll",
    119 epoll_init,
    120 epoll_nochangelist_add,
    121 epoll_nochangelist_del,
    122 epoll_dispatch,
    123 epoll_dealloc,
    124 1, /* need reinit */
    125 EV_FEATURE_ET|EV_FEATURE_O1|EV_FEATURE_EARLY_CLOSE,
    126 0
    127 };
    128 
    129 #define INITIAL_NEVENT 32
    130 #define MAX_NEVENT 4096
    131 
    132 /* On Linux kernels at least up to 2.6.24.4, epoll can't handle timeout
    133 * values bigger than (LONG_MAX - 999ULL)/HZ.  HZ in the wild can be
    134 * as big as 1000, and LONG_MAX can be as small as (1<<31)-1, so the
    135 * largest number of msec we can support here is 2147482.  Let's
    136 * round that down by 47 seconds.
    137 */
    138 #define MAX_EPOLL_TIMEOUT_MSEC (35*60*1000)
    139 
    140 static void *
    141 epoll_init(struct event_base *base)
    142 {
    143 int epfd = -1;
    144 struct epollop *epollop;
    145 
    146 #ifdef EVENT__HAVE_EPOLL_CREATE1
    147 /* First, try the shiny new epoll_create1 interface, if we have it. */
    148 epfd = epoll_create1(EPOLL_CLOEXEC);
    149 #endif
    150 if (epfd == -1) {
    151 	/* Initialize the kernel queue using the old interface.  (The
    152 	size field is ignored   since 2.6.8.) */
    153 	if ((epfd = epoll_create(32000)) == -1) {
    154 		if (errno != ENOSYS)
    155 			event_warn("epoll_create");
    156 		return (NULL);
    157 	}
    158 	evutil_make_socket_closeonexec(epfd);
    159 }
    160 
    161 if (!(epollop = mm_calloc(1, sizeof(struct epollop)))) {
    162 	close(epfd);
    163 	return (NULL);
    164 }
    165 
    166 epollop->epfd = epfd;
    167 
    168 /* Initialize fields */
    169 epollop->events = mm_calloc(INITIAL_NEVENT, sizeof(struct epoll_event));
    170 if (epollop->events == NULL) {
    171 	mm_free(epollop);
    172 	close(epfd);
    173 	return (NULL);
    174 }
    175 epollop->nevents = INITIAL_NEVENT;
    176 
    177 if ((base->flags & EVENT_BASE_FLAG_EPOLL_USE_CHANGELIST) != 0 ||
    178     ((base->flags & EVENT_BASE_FLAG_IGNORE_ENV) == 0 &&
    179 	evutil_getenv_("EVENT_EPOLL_USE_CHANGELIST") != NULL)) {
    180 
    181 	base->evsel = &epollops_changelist;
    182 }
    183 
    184 #ifdef USING_TIMERFD
    185 /*
    186   The epoll interface ordinarily gives us one-millisecond precision,
    187   so on Linux it makes perfect sense to use the CLOCK_MONOTONIC_COARSE
    188   timer.  But when the user has set the new PRECISE_TIMER flag for an
    189   event_base, we can try to use timerfd to give them finer granularity.
    190 */
    191 if ((base->flags & EVENT_BASE_FLAG_PRECISE_TIMER) &&
    192     base->monotonic_timer.monotonic_clock == CLOCK_MONOTONIC) {
    193 	int fd;
    194 	fd = epollop->timerfd = timerfd_create(CLOCK_MONOTONIC, TFD_NONBLOCK|TFD_CLOEXEC);
    195 	if (epollop->timerfd >= 0) {
    196 		struct epoll_event epev;
    197 		memset(&epev, 0, sizeof(epev));
    198 		epev.data.fd = epollop->timerfd;
    199 		epev.events = EPOLLIN;
    200 		if (epoll_ctl(epollop->epfd, EPOLL_CTL_ADD, fd, &epev) < 0) {
    201 			event_warn("epoll_ctl(timerfd)");
    202 			close(fd);
    203 			epollop->timerfd = -1;
    204 		}
    205 	} else {
    206 		if (errno != EINVAL && errno != ENOSYS) {
    207 			/* These errors probably mean that we were
    208 			 * compiled with timerfd/TFD_* support, but
    209 			 * we're running on a kernel that lacks those.
    210 			 */
    211 			event_warn("timerfd_create");
    212 		}
    213 		epollop->timerfd = -1;
    214 	}
    215 } else {
    216 	epollop->timerfd = -1;
    217 }
    218 #endif
    219 
    220 evsig_init_(base);
    221 
    222 return (epollop);
    223 }
    224 
    225 static const char *
    226 change_to_string(int change)
    227 {
    228 change &= (EV_CHANGE_ADD|EV_CHANGE_DEL);
    229 if (change == EV_CHANGE_ADD) {
    230 	return "add";
    231 } else if (change == EV_CHANGE_DEL) {
    232 	return "del";
    233 } else if (change == 0) {
    234 	return "none";
    235 } else {
    236 	return "???";
    237 }
    238 }
    239 
    240 static const char *
    241 epoll_op_to_string(int op)
    242 {
    243 return op == EPOLL_CTL_ADD?"ADD":
    244     op == EPOLL_CTL_DEL?"DEL":
    245     op == EPOLL_CTL_MOD?"MOD":
    246     "???";
    247 }
    248 
    249 #define PRINT_CHANGES(op, events, ch, status)  \
    250 "Epoll %s(%d) on fd %d " status ". "       \
    251 "Old events were %d; "                     \
    252 "read change was %d (%s); "                \
    253 "write change was %d (%s); "               \
    254 "close change was %d (%s)",                \
    255 epoll_op_to_string(op),                    \
    256 events,                                    \
    257 ch->fd,                                    \
    258 ch->old_events,                            \
    259 ch->read_change,                           \
    260 change_to_string(ch->read_change),         \
    261 ch->write_change,                          \
    262 change_to_string(ch->write_change),        \
    263 ch->close_change,                          \
    264 change_to_string(ch->close_change)
    265 
    266 static int
    267 epoll_apply_one_change(struct event_base *base,
    268    struct epollop *epollop,
    269    const struct event_change *ch)
    270 {
    271 struct epoll_event epev;
    272 int op, events = 0;
    273 int idx;
    274 
    275 idx = EPOLL_OP_TABLE_INDEX(ch);
    276 op = epoll_op_table[idx].op;
    277 events = epoll_op_table[idx].events;
    278 
    279 if (!events) {
    280 	EVUTIL_ASSERT(op == 0);
    281 	return 0;
    282 }
    283 
    284 if ((ch->read_change|ch->write_change|ch->close_change) & EV_CHANGE_ET)
    285 	events |= EPOLLET;
    286 
    287 memset(&epev, 0, sizeof(epev));
    288 epev.data.fd = ch->fd;
    289 epev.events = events;
    290 if (epoll_ctl(epollop->epfd, op, ch->fd, &epev) == 0) {
    291 	event_debug((PRINT_CHANGES(op, epev.events, ch, "okay")));
    292 	return 0;
    293 }
    294 
    295 switch (op) {
    296 case EPOLL_CTL_MOD:
    297 	if (errno == ENOENT) {
    298 		/* If a MOD operation fails with ENOENT, the
    299 		 * fd was probably closed and re-opened.  We
    300 		 * should retry the operation as an ADD.
    301 		 */
    302 		if (epoll_ctl(epollop->epfd, EPOLL_CTL_ADD, ch->fd, &epev) == -1) {
    303 			event_warn("Epoll MOD(%d) on %d retried as ADD; that failed too",
    304 			    (int)epev.events, ch->fd);
    305 			return -1;
    306 		} else {
    307 			event_debug(("Epoll MOD(%d) on %d retried as ADD; succeeded.",
    308 				(int)epev.events,
    309 				ch->fd));
    310 			return 0;
    311 		}
    312 	}
    313 	break;
    314 case EPOLL_CTL_ADD:
    315 	if (errno == EEXIST) {
    316 		/* If an ADD operation fails with EEXIST,
    317 		 * either the operation was redundant (as with a
    318 		 * precautionary add), or we ran into a fun
    319 		 * kernel bug where using dup*() to duplicate the
    320 		 * same file into the same fd gives you the same epitem
    321 		 * rather than a fresh one.  For the second case,
    322 		 * we must retry with MOD. */
    323 		if (epoll_ctl(epollop->epfd, EPOLL_CTL_MOD, ch->fd, &epev) == -1) {
    324 			event_warn("Epoll ADD(%d) on %d retried as MOD; that failed too",
    325 			    (int)epev.events, ch->fd);
    326 			return -1;
    327 		} else {
    328 			event_debug(("Epoll ADD(%d) on %d retried as MOD; succeeded.",
    329 				(int)epev.events,
    330 				ch->fd));
    331 			return 0;
    332 		}
    333 	}
    334 	break;
    335 case EPOLL_CTL_DEL:
    336 	if (errno == ENOENT || errno == EBADF || errno == EPERM) {
    337 		/* If a delete fails with one of these errors,
    338 		 * that's fine too: we closed the fd before we
    339 		 * got around to calling epoll_dispatch. */
    340 		event_debug(("Epoll DEL(%d) on fd %d gave %s: DEL was unnecessary.",
    341 			(int)epev.events,
    342 			ch->fd,
    343 			strerror(errno)));
    344 		return 0;
    345 	}
    346 	break;
    347 default:
    348 	break;
    349 }
    350 
    351 event_warn(PRINT_CHANGES(op, epev.events, ch, "failed"));
    352 return -1;
    353 }
    354 
    355 static int
    356 epoll_apply_changes(struct event_base *base)
    357 {
    358 struct event_changelist *changelist = &base->changelist;
    359 struct epollop *epollop = base->evbase;
    360 struct event_change *ch;
    361 
    362 int r = 0;
    363 int i;
    364 
    365 for (i = 0; i < changelist->n_changes; ++i) {
    366 	ch = &changelist->changes[i];
    367 	if (epoll_apply_one_change(base, epollop, ch) < 0)
    368 		r = -1;
    369 }
    370 
    371 return (r);
    372 }
    373 
    374 static int
    375 epoll_nochangelist_add(struct event_base *base, evutil_socket_t fd,
    376    short old, short events, void *p)
    377 {
    378 struct event_change ch;
    379 ch.fd = fd;
    380 ch.old_events = old;
    381 ch.read_change = ch.write_change = ch.close_change = 0;
    382 if (events & EV_WRITE)
    383 	ch.write_change = EV_CHANGE_ADD |
    384 	    (events & EV_ET);
    385 if (events & EV_READ)
    386 	ch.read_change = EV_CHANGE_ADD |
    387 	    (events & EV_ET);
    388 if (events & EV_CLOSED)
    389 	ch.close_change = EV_CHANGE_ADD |
    390 	    (events & EV_ET);
    391 
    392 return epoll_apply_one_change(base, base->evbase, &ch);
    393 }
    394 
    395 static int
    396 epoll_nochangelist_del(struct event_base *base, evutil_socket_t fd,
    397    short old, short events, void *p)
    398 {
    399 struct event_change ch;
    400 ch.fd = fd;
    401 ch.old_events = old;
    402 ch.read_change = ch.write_change = ch.close_change = 0;
    403 if (events & EV_WRITE)
    404 	ch.write_change = EV_CHANGE_DEL |
    405 	    (events & EV_ET);
    406 if (events & EV_READ)
    407 	ch.read_change = EV_CHANGE_DEL |
    408 	    (events & EV_ET);
    409 if (events & EV_CLOSED)
    410 	ch.close_change = EV_CHANGE_DEL |
    411 	    (events & EV_ET);
    412 
    413 return epoll_apply_one_change(base, base->evbase, &ch);
    414 }
    415 
    416 static int
    417 epoll_dispatch(struct event_base *base, struct timeval *tv)
    418 {
    419 struct epollop *epollop = base->evbase;
    420 struct epoll_event *events = epollop->events;
    421 int i, res;
    422 long timeout = -1;
    423 
    424 #ifdef USING_TIMERFD
    425 if (epollop->timerfd >= 0) {
    426 	struct itimerspec is;
    427 	is.it_interval.tv_sec = 0;
    428 	is.it_interval.tv_nsec = 0;
    429 	if (tv == NULL) {
    430 		/* No timeout; disarm the timer. */
    431 		is.it_value.tv_sec = 0;
    432 		is.it_value.tv_nsec = 0;
    433 	} else {
    434 		if (tv->tv_sec == 0 && tv->tv_usec == 0) {
    435 			/* we need to exit immediately; timerfd can't
    436 			 * do that. */
    437 			timeout = 0;
    438 		}
    439 		is.it_value.tv_sec = tv->tv_sec;
    440 		is.it_value.tv_nsec = tv->tv_usec * 1000;
    441 	}
    442 	/* TODO: we could avoid unnecessary syscalls here by only
    443 	   calling timerfd_settime when the top timeout changes, or
    444 	   when we're called with a different timeval.
    445 	*/
    446 	if (timerfd_settime(epollop->timerfd, 0, &is, NULL) < 0) {
    447 		event_warn("timerfd_settime");
    448 	}
    449 } else
    450 #endif
    451 if (tv != NULL) {
    452 	timeout = evutil_tv_to_msec_(tv);
    453 	if (timeout < 0 || timeout > MAX_EPOLL_TIMEOUT_MSEC) {
    454 		/* Linux kernels can wait forever if the timeout is
    455 		 * too big; see comment on MAX_EPOLL_TIMEOUT_MSEC. */
    456 		timeout = MAX_EPOLL_TIMEOUT_MSEC;
    457 	}
    458 }
    459 
    460 epoll_apply_changes(base);
    461 event_changelist_remove_all_(&base->changelist, base);
    462 
    463 EVBASE_RELEASE_LOCK(base, th_base_lock);
    464 
    465 res = epoll_wait(epollop->epfd, events, epollop->nevents, timeout);
    466 
    467 EVBASE_ACQUIRE_LOCK(base, th_base_lock);
    468 
    469 if (res == -1) {
    470 	if (errno != EINTR) {
    471 		event_warn("epoll_wait");
    472 		return (-1);
    473 	}
    474 
    475 	return (0);
    476 }
    477 
    478 event_debug(("%s: epoll_wait reports %d", __func__, res));
    479 EVUTIL_ASSERT(res <= epollop->nevents);
    480 
    481 for (i = 0; i < res; i++) {
    482 	int what = events[i].events;
    483 	short ev = 0;
    484 #ifdef USING_TIMERFD
    485 	if (events[i].data.fd == epollop->timerfd)
    486 		continue;
    487 #endif
    488 
    489 	if (what & EPOLLERR) {
    490 		ev = EV_READ | EV_WRITE;
    491 	} else if ((what & EPOLLHUP) && !(what & EPOLLRDHUP)) {
    492 		ev = EV_READ | EV_WRITE;
    493 	} else {
    494 		if (what & EPOLLIN)
    495 			ev |= EV_READ;
    496 		if (what & EPOLLOUT)
    497 			ev |= EV_WRITE;
    498 		if (what & EPOLLRDHUP)
    499 			ev |= EV_CLOSED;
    500 	}
    501 
    502 	if (!ev)
    503 		continue;
    504 
    505 	evmap_io_active_(base, events[i].data.fd, ev | EV_ET);
    506 }
    507 
    508 if (res == epollop->nevents && epollop->nevents < MAX_NEVENT) {
    509 	/* We used all of the event space this time.  We should
    510 	   be ready for more events next time. */
    511 	int new_nevents = epollop->nevents * 2;
    512 	struct epoll_event *new_events;
    513 
    514 	new_events = mm_realloc(epollop->events,
    515 	    new_nevents * sizeof(struct epoll_event));
    516 	if (new_events) {
    517 		epollop->events = new_events;
    518 		epollop->nevents = new_nevents;
    519 	}
    520 }
    521 
    522 return (0);
    523 }
    524 
    525 
    526 static void
    527 epoll_dealloc(struct event_base *base)
    528 {
    529 struct epollop *epollop = base->evbase;
    530 
    531 evsig_dealloc_(base);
    532 if (epollop->events)
    533 	mm_free(epollop->events);
    534 if (epollop->epfd >= 0)
    535 	close(epollop->epfd);
    536 #ifdef USING_TIMERFD
    537 if (epollop->timerfd >= 0)
    538 	close(epollop->timerfd);
    539 #endif
    540 
    541 memset(epollop, 0, sizeof(struct epollop));
    542 mm_free(epollop);
    543 }
    544 
    545 #endif /* EVENT__HAVE_EPOLL */