Added epoll(7) support to SILC Scheduler.
authorPekka Riikonen <priikone@silcnet.org>
Sat, 23 Dec 2006 17:57:00 +0000 (17:57 +0000)
committerPekka Riikonen <priikone@silcnet.org>
Sat, 23 Dec 2006 17:57:00 +0000 (17:57 +0000)
lib/silcutil/silclog.h
lib/silcutil/silcschedule.c
lib/silcutil/silcschedule.h
lib/silcutil/silcschedule_i.h
lib/silcutil/unix/silcunixnet.c
lib/silcutil/unix/silcunixschedule.c
lib/silcutil/unix/silcunixsocketstream.c

index 0b12a2a4c300e021940cff04f8590b6ffdfa8d79..60949270eb1b8587f51f251a326ff0a36ad65a44 100644 (file)
@@ -56,7 +56,7 @@ typedef enum {
  * SYNOPSIS
  *
  *    typedef SilcBool (*SilcLogCb)(SilcLogType type, char *message,
- *                              void *context);
+ *                                  void *context);
  *
  * DESCRIPTION
  *
@@ -79,7 +79,7 @@ typedef SilcBool (*SilcLogCb)(SilcLogType type, char *message, void *context);
  * SYNOPSIS
  *
  *    typedef SilcBool (*SilcLogDebugCb)(char *file, char *function, int line,
- *                                   char *message, void *context);
+ *                                       char *message, void *context);
  *
  * DESCRIPTION
  *
@@ -98,16 +98,17 @@ typedef SilcBool (*SilcLogCb)(SilcLogType type, char *message, void *context);
  *
  ***/
 typedef SilcBool (*SilcLogDebugCb)(char *file, char *function, int line,
-                              char *message, void *context);
+                                  char *message, void *context);
 
 /****f* silcutil/SilcLogAPI/SilcLogHexdumpCb
  *
  * SYNOPSIS
  *
- *    typedef SilcBool (*SilcDebugHexdumpCb)(char *file, char *function, int line,
- *                                       unsigned char *data,
- *                                       SilcUInt32 data_len,
- *                                       char *message, void *context;
+ *    typedef SilcBool
+ *    (*SilcDebugHexdumpCb)(char *file, char *function, int line,
+ *                          unsigned char *data,
+ *                          SilcUInt32 data_len,
+ *                          char *message, void *context;
  *
  * DESCRIPTION
  *
@@ -128,8 +129,8 @@ typedef SilcBool (*SilcLogDebugCb)(char *file, char *function, int line,
  *
  ***/
 typedef SilcBool (*SilcLogHexdumpCb)(char *file, char *function, int line,
-                                unsigned char *data, SilcUInt32 data_len,
-                                char *message, void *context);
+                                    unsigned char *data, SilcUInt32 data_len,
+                                    char *message, void *context);
 
 /* Macros */
 
@@ -294,8 +295,8 @@ typedef SilcBool (*SilcLogHexdumpCb)(char *file, char *function, int line,
  * DESCRIPTION
  *
  *    Assert macro that prints error message to stderr and calls abort()
- *    if the `expression' is is false (ie. compares equal to zero).  If
- *    SILC_DEBUG is not defined this macro as no effect.
+ *    if the `expression' is false (ie. compares equal to zero).  If
+ *    SILC_DEBUG is not defined this macro has no effect.
  *
  * SOURCE
  */
index cd5a32030506a105799c1403f4c4f080f1607446..2e546ee8576a6ee919bf97e9902268c3431a14d9 100644 (file)
@@ -45,44 +45,34 @@ static void silc_schedule_fd_destructor(void *key, void *context,
 
 static void silc_schedule_dispatch_fd(SilcSchedule schedule)
 {
-  SilcHashTableList htl;
-  SilcTask t;
   SilcTaskFd task;
-  SilcUInt32 fd;
+  SilcTask t;
 
-  silc_hash_table_list(schedule->fd_queue, &htl);
-  while (silc_likely(silc_hash_table_get(&htl, (void **)&fd,
-                                        (void **)&task))) {
+  /* The dispatch list includes only valid tasks, and tasks that have
+     something to dispatch.  Dispatching is atomic; no matter if another
+     thread invalidates a task when we unlock, we dispatch to completion. */
+  SILC_SCHEDULE_UNLOCK(schedule);
+  silc_list_start(schedule->fd_dispatch);
+  while ((task = silc_list_get(schedule->fd_dispatch))) {
     t = (SilcTask)task;
 
-    if (silc_unlikely(!t->valid)) {
-      silc_schedule_task_remove(schedule, t);
-      continue;
-    }
-    if (!task->revents || !task->events)
-      continue;
-
     /* Is the task ready for reading */
-    if (task->revents & SILC_TASK_READ) {
-      SILC_SCHEDULE_UNLOCK(schedule);
+    if (task->revents & SILC_TASK_READ)
       t->callback(schedule, schedule->app_context, SILC_TASK_READ,
                  task->fd, t->context);
-      SILC_SCHEDULE_LOCK(schedule);
-    }
 
     /* Is the task ready for writing */
-    if (t->valid && task->revents & SILC_TASK_WRITE) {
-      SILC_SCHEDULE_UNLOCK(schedule);
+    if (t->valid && task->revents & SILC_TASK_WRITE)
       t->callback(schedule, schedule->app_context, SILC_TASK_WRITE,
                  task->fd, t->context);
-      SILC_SCHEDULE_LOCK(schedule);
-    }
-
-    /* Remove if task was invalidated in the task callback */
-    if (silc_unlikely(!t->valid))
-      silc_schedule_task_remove(schedule, t);
   }
-  silc_hash_table_list_reset(&htl);
+  SILC_SCHEDULE_LOCK(schedule);
+
+  /* Remove invalidated tasks */
+  silc_list_start(schedule->fd_dispatch);
+  while ((task = silc_list_get(schedule->fd_dispatch)))
+    if (silc_unlikely(!task->header.valid))
+      silc_schedule_task_remove(schedule, (SilcTask)task);
 }
 
 /* Executes all tasks whose timeout has expired. The task is removed from
@@ -338,8 +328,8 @@ SilcSchedule silc_schedule_init(int max_tasks, void *app_context)
   if (!schedule->fd_queue)
     return NULL;
 
-  silc_list_init(schedule->timeout_queue, struct SilcTaskTimeoutStruct, next);
-  silc_list_init(schedule->free_tasks, struct SilcTaskTimeoutStruct, next);
+  silc_list_init(schedule->timeout_queue, struct SilcTaskStruct, next);
+  silc_list_init(schedule->free_tasks, struct SilcTaskStruct, next);
 
   schedule->app_context = app_context;
   schedule->valid = TRUE;
@@ -462,7 +452,7 @@ static SilcBool silc_schedule_iterate(SilcSchedule schedule, int timeout_usecs)
        of the selected file descriptors change status or the selected
        timeout expires. */
     SILC_LOG_DEBUG(("Select"));
-    ret = schedule_ops.select(schedule, schedule->internal);
+    ret = schedule_ops.schedule(schedule, schedule->internal);
 
     if (silc_likely(ret == 0)) {
       /* Timeout */
@@ -812,7 +802,8 @@ void silc_schedule_set_listen_fd(SilcSchedule schedule, SilcUInt32 fd,
   if (silc_hash_table_find(schedule->fd_queue, SILC_32_TO_PTR(fd),
                           NULL, (void **)&task)) {
     task->events = mask;
-    if (silc_unlikely(send_events)) {
+    schedule_ops.schedule_fd(schedule, schedule->internal, task, mask);
+    if (silc_unlikely(send_events) && mask) {
       task->revents = mask;
       silc_schedule_dispatch_fd(schedule);
     }
@@ -821,6 +812,26 @@ void silc_schedule_set_listen_fd(SilcSchedule schedule, SilcUInt32 fd,
   SILC_SCHEDULE_UNLOCK(schedule);
 }
 
+/* Returns the file descriptors current requested event mask. */
+
+SilcTaskEvent silc_schedule_get_fd_events(SilcSchedule schedule,
+                                         SilcUInt32 fd)
+{
+  SilcTaskFd task;
+  SilcTaskEvent event = 0;
+
+  if (silc_unlikely(!schedule->valid))
+    return 0;
+
+  SILC_SCHEDULE_LOCK(schedule);
+  if (silc_hash_table_find(schedule->fd_queue, SILC_32_TO_PTR(fd),
+                          NULL, (void **)&task))
+    event = task->events;
+  SILC_SCHEDULE_UNLOCK(schedule);
+
+  return event;
+}
+
 /* Removes a file descriptor from listen list. */
 
 void silc_schedule_unset_listen_fd(SilcSchedule schedule, SilcUInt32 fd)
index 637f058101bf28f52e3587df8caef94e1b772b86..44060fdca04ea11c72d4fd469c8b13621e1e58b3 100644 (file)
@@ -542,6 +542,22 @@ void silc_schedule_task_del_by_all(SilcSchedule schedule, int fd,
 void silc_schedule_set_listen_fd(SilcSchedule schedule, SilcUInt32 fd,
                                 SilcTaskEvent mask, SilcBool send_events);
 
+/****f* silcutil/SilcScheduleAPI/silc_schedule_get_fd_events
+ *
+ * SYNOPSIS
+ *
+ *    SilcTaskEvent silc_schedule_get_fd_events(SilcSchedule schedule,
+ *                                              SilcUInt32 fd);
+ *
+ * DESCRIPTION
+ *
+ *    Returns the file descriptor `fd' current requested events mask,
+ *    or 0 on error.
+ *
+ ***/
+SilcTaskEvent silc_schedule_get_fd_events(SilcSchedule schedule,
+                                         SilcUInt32 fd);
+
 /****f* silcutil/SilcScheduleAPI/silc_schedule_unset_listen_fd
  *
  * SYNOPSIS
index 3c1598c7fe0901bd429bdaa017a283e8aea8b429..04976c36dbe1e7ee9f1bcd3571cb4b9930ed53db 100644 (file)
@@ -29,6 +29,7 @@
 
 /* Task header */
 struct SilcTaskStruct {
+  struct SilcTaskStruct *next;
   SilcTaskCallback callback;
   void *context;
   unsigned int type    : 1;    /* 0 = fd, 1 = timeout */
@@ -38,15 +39,15 @@ struct SilcTaskStruct {
 /* Timeout task */
 typedef struct SilcTaskTimeoutStruct {
   struct SilcTaskStruct header;
-  struct SilcTaskTimeoutStruct *next;
   struct timeval timeout;
 } *SilcTaskTimeout;
 
 /* Fd task */
-typedef struct {
+typedef struct SilcTaskFdStruct {
   struct SilcTaskStruct header;
-  unsigned int events  : 15;
-  unsigned int revents : 15;
+  unsigned int scheduled  : 1;
+  unsigned int events     : 14;
+  unsigned int revents    : 15;
   SilcUInt32 fd;
 } *SilcTaskFd;
 
@@ -55,11 +56,12 @@ struct SilcScheduleStruct {
   void *internal;
   void *app_context;              /* Application specific context */
   SilcHashTable fd_queue;         /* FD task queue */
+  SilcList fd_dispatch;                   /* Dispatched FDs */
   SilcList timeout_queue;         /* Timeout queue */
   SilcList free_tasks;            /* Timeout task freelist */
   SilcMutex lock;                 /* Scheduler lock */
   struct timeval timeout;         /* Current timeout */
-  unsigned int max_tasks     : 28; /* Max FD tasks */
+  unsigned int max_tasks     : 29; /* Max FD tasks */
   unsigned int has_timeout   : 1;  /* Set if timeout is set */
   unsigned int valid         : 1;  /* Set if scheduler is valid */
   unsigned int signal_tasks  : 1;  /* Set if to dispatch signals */
@@ -70,13 +72,13 @@ struct SilcScheduleStruct {
    synchronise signals with SILC Scheduler. */
 #define SILC_SCHEDULE_LOCK(schedule)                           \
 do {                                                           \
-  schedule_ops.signals_block(schedule, schedule->internal);    \
   silc_mutex_lock(schedule->lock);                             \
+  schedule_ops.signals_block(schedule, schedule->internal);    \
 } while (0)
 #define SILC_SCHEDULE_UNLOCK(schedule)                         \
 do {                                                           \
-  silc_mutex_unlock(schedule->lock);                           \
   schedule_ops.signals_unblock(schedule, schedule->internal);  \
+  silc_mutex_unlock(schedule->lock);                           \
 } while (0)
 
 /* Platform specific scheduler operations */
@@ -90,8 +92,14 @@ typedef struct {
   /* Uninitializes the platform specific scheduler context. */
   void (*uninit)(SilcSchedule schedule, void *context);
 
-  /* System specific select(). Returns same values as normal select(). */
-  int (*select)(SilcSchedule schedule, void *context);
+  /* System specific waiter. This must fill the schedule->fd_dispatch queue
+     with valid tasks that has something to dispatch, when this returns. */
+  int (*schedule)(SilcSchedule schedule, void *context);
+
+  /* Schedule `task' with events `event_mask'. Zero `event_mask'
+     unschedules the task. */
+  SilcBool (*schedule_fd)(SilcSchedule schedule, void *context,
+                         SilcTaskFd task, SilcTaskEvent event_mask);
 
   /* Wakes up the scheduler. This is platform specific routine */
   void (*wakeup)(SilcSchedule schedule, void *context);
index 294de846616c039241eb2df672e76316c9ef3790..bc138cd943bd27a42060c6b8d71638c476f0d71f 100644 (file)
@@ -438,8 +438,10 @@ int silc_net_udp_send(SilcStream stream,
   }
 
   SILC_LOG_DEBUG(("Sent data %d bytes", ret));
-  silc_schedule_set_listen_fd(sock->schedule, sock->sock,
-                             SILC_TASK_READ, FALSE);
+  if (silc_schedule_get_fd_events(sock->schedule, sock->sock) &
+      SILC_TASK_WRITE)
+    silc_schedule_set_listen_fd(sock->schedule, sock->sock,
+                               SILC_TASK_READ, FALSE);
 
   return ret;
 }
@@ -451,7 +453,7 @@ typedef struct {
   SilcSocketStreamStatus stream_status;
   SilcStream stream;
   SilcFSMStruct fsm;
-  SilcFSMSemaStruct sema;
+  SilcFSMEventStruct sema;
   SilcAsyncOperation op;
   SilcAsyncOperation sop;
   char *local_ip;
@@ -473,7 +475,7 @@ SILC_FSM_STATE(silc_net_connect_st_finish);
 SILC_TASK_CALLBACK(silc_net_connect_wait)
 {
   SilcNetConnect conn = context;
-  SILC_FSM_SEMA_POST(&conn->sema);
+  SILC_FSM_EVENT_SIGNAL(&conn->sema);
   silc_schedule_task_del_by_fd(schedule, conn->sock);
 }
 
@@ -487,7 +489,7 @@ SILC_FSM_STATE(silc_net_connect_st_start)
   if (conn->aborted) {
     /** Aborted */
     silc_fsm_next(fsm, silc_net_connect_st_finish);
-    return SILC_FSM_CONTINUE;
+    SILC_FSM_CONTINUE;
   }
 
   /* Do host lookup */
@@ -500,14 +502,14 @@ SILC_FSM_STATE(silc_net_connect_st_start)
     /** Network unreachable */
     conn->status = SILC_NET_HOST_UNREACHABLE;
     silc_fsm_next(fsm, silc_net_connect_st_finish);
-    return SILC_FSM_CONTINUE;
+    SILC_FSM_CONTINUE;
   }
 
   /* Set sockaddr for this connection */
   if (!silc_net_set_sockaddr(&desthost, conn->ip_addr, conn->port)) {
     /** Sockaddr failed */
     silc_fsm_next(fsm, silc_net_connect_st_finish);
-    return SILC_FSM_CONTINUE;
+    SILC_FSM_CONTINUE;
   }
 
   /* Create the connection socket */
@@ -523,7 +525,7 @@ SILC_FSM_STATE(silc_net_connect_st_start)
     /** Cannot create socket */
     SILC_LOG_ERROR(("Cannot create socket: %s", strerror(errno)));
     silc_fsm_next(fsm, silc_net_connect_st_finish);
-    return SILC_FSM_CONTINUE;
+    SILC_FSM_CONTINUE;
   }
 
   /* Bind to the local address if provided */
@@ -557,7 +559,7 @@ SILC_FSM_STATE(silc_net_connect_st_start)
       /** Cannot connect to remote host */
       SILC_LOG_ERROR(("Cannot connect to remote host: %s", strerror(errno)));
       silc_fsm_next(fsm, silc_net_connect_st_finish);
-      return SILC_FSM_CONTINUE;
+      SILC_FSM_CONTINUE;
     }
   }
 
@@ -573,13 +575,13 @@ SILC_FSM_STATE(silc_net_connect_st_start)
 
   /** Wait for connection */
   silc_fsm_next(fsm, silc_net_connect_st_connected);
-  silc_fsm_sema_init(&conn->sema, fsm, 0);
+  silc_fsm_event_init(&conn->sema, fsm);
   silc_schedule_task_add_fd(silc_fsm_get_schedule(fsm), sock,
                            silc_net_connect_wait, conn);
   silc_schedule_set_listen_fd(silc_fsm_get_schedule(fsm), sock,
                              SILC_TASK_WRITE, FALSE);
-  SILC_FSM_SEMA_WAIT(&conn->sema);
-  return SILC_FSM_CONTINUE;
+  SILC_FSM_EVENT_WAIT(&conn->sema);
+  SILC_FSM_CONTINUE;
 }
 
 static void silc_net_connect_wait_stream(SilcSocketStreamStatus status,
@@ -600,7 +602,7 @@ SILC_FSM_STATE(silc_net_connect_st_connected)
   if (conn->aborted) {
     /** Aborted */
     silc_fsm_next(fsm, silc_net_connect_st_finish);
-    return SILC_FSM_CONTINUE;
+    SILC_FSM_CONTINUE;
   }
 
   ret = silc_net_get_socket_opt(conn->sock, SOL_SOCKET, SO_ERROR,
@@ -616,7 +618,7 @@ SILC_FSM_STATE(silc_net_connect_st_connected)
       conn->retry--;
       silc_net_close_connection(conn->sock);
       silc_fsm_next(fsm, silc_net_connect_st_start);
-      return SILC_FSM_CONTINUE;
+      SILC_FSM_CONTINUE;
     }
 
 #if defined(ECONNREFUSED)
@@ -635,7 +637,7 @@ SILC_FSM_STATE(silc_net_connect_st_connected)
     /** Connecting failed */
     SILC_LOG_DEBUG(("Connecting failed"));
     silc_fsm_next(fsm, silc_net_connect_st_finish);
-    return SILC_FSM_CONTINUE;
+    SILC_FSM_CONTINUE;
   }
 
   /** Connection created */
@@ -653,7 +655,7 @@ SILC_FSM_STATE(silc_net_connect_st_stream)
   if (conn->aborted) {
     /** Aborted */
     silc_fsm_next(fsm, silc_net_connect_st_finish);
-    return SILC_FSM_CONTINUE;
+    SILC_FSM_CONTINUE;
   }
 
   if (conn->stream_status != SILC_SOCKET_OK) {
@@ -665,7 +667,7 @@ SILC_FSM_STATE(silc_net_connect_st_stream)
     else
       conn->status = SILC_NET_ERROR;
     silc_fsm_next(fsm, silc_net_connect_st_finish);
-    return SILC_FSM_CONTINUE;
+    SILC_FSM_CONTINUE;
   }
 
   /* Set stream information */
@@ -677,7 +679,7 @@ SILC_FSM_STATE(silc_net_connect_st_stream)
   SILC_LOG_DEBUG(("Connected successfully"));
   conn->status = SILC_NET_OK;
   silc_fsm_next(fsm, silc_net_connect_st_finish);
-  return SILC_FSM_CONTINUE;
+  SILC_FSM_CONTINUE;
 }
 
 SILC_FSM_STATE(silc_net_connect_st_finish)
@@ -693,7 +695,7 @@ SILC_FSM_STATE(silc_net_connect_st_finish)
       silc_async_free(conn->sop);
   }
 
-  return SILC_FSM_FINISH;
+  SILC_FSM_FINISH;
 }
 
 static void silc_net_connect_abort(SilcAsyncOperation op, void *context)
index 914b42fef891a8f2a6391e927425ac951cafb834..d169ee5222c12b77d3881e293f797d2838ed2a48 100644 (file)
@@ -20,7 +20,9 @@
 
 #include "silc.h"
 
-#if defined(HAVE_POLL) && defined(HAVE_SETRLIMIT) && defined(RLIMIT_NOFILE)
+#if defined(HAVE_EPOLL_WAIT)
+#include <sys/epoll.h>
+#elif defined(HAVE_POLL) && defined(HAVE_SETRLIMIT) && defined(RLIMIT_NOFILE)
 #include <poll.h>
 #endif
 
@@ -28,7 +30,11 @@ const SilcScheduleOps schedule_ops;
 
 /* Internal context. */
 typedef struct {
-#if defined(HAVE_POLL) && defined(HAVE_SETRLIMIT) && defined(RLIMIT_NOFILE)
+#if defined(HAVE_EPOLL_WAIT)
+  struct epoll_event *fds;
+  SilcUInt32 fds_count;
+  int epfd;
+#elif defined(HAVE_POLL) && defined(HAVE_SETRLIMIT) && defined(RLIMIT_NOFILE)
   struct rlimit nofile;
   struct pollfd *fds;
   SilcUInt32 fds_count;
@@ -51,7 +57,59 @@ typedef struct {
 #define SIGNAL_COUNT 32
 SilcUnixSignal signal_call[SIGNAL_COUNT];
 
-#if defined(HAVE_POLL) && defined(HAVE_SETRLIMIT) && defined(RLIMIT_NOFILE)
+#if defined(HAVE_EPOLL_WAIT)
+
+/* Linux's fast epoll system (level triggered) */
+
+int silc_epoll(SilcSchedule schedule, void *context)
+{
+  SilcUnixScheduler internal = context;
+  SilcTaskFd task;
+  struct epoll_event *fds = internal->fds;
+  SilcUInt32 fds_count = internal->fds_count;
+  int ret, i, timeout = -1;
+
+  /* Allocate larger fd table if needed */
+  i = silc_hash_table_count(schedule->fd_queue);
+  if (i > fds_count) {
+    fds = silc_realloc(internal->fds, sizeof(*internal->fds) *
+                      (fds_count + (i / 2)));
+    if (silc_likely(fds)) {
+      internal->fds = fds;
+      internal->fds_count = fds_count = fds_count + (i / 2);
+    }
+  }
+
+  if (schedule->has_timeout)
+    timeout = ((schedule->timeout.tv_sec * 1000) +
+              (schedule->timeout.tv_usec / 1000));
+
+  SILC_SCHEDULE_UNLOCK(schedule);
+  ret = epoll_wait(internal->epfd, fds, fds_count, timeout);
+  SILC_SCHEDULE_LOCK(schedule);
+  if (ret <= 0)
+    return ret;
+
+  silc_list_init(schedule->fd_dispatch, struct SilcTaskStruct, next);
+
+  for (i = 0; i < ret; i++) {
+    task = fds[i].data.ptr;
+    task->revents = 0;
+    if (!task->header.valid || !task->events) {
+      epoll_ctl(internal->epfd, EPOLL_CTL_DEL, task->fd, &fds[i]);
+      continue;
+    }
+    if (fds[i].events & EPOLLIN)
+      task->revents |= SILC_TASK_READ;
+    if (fds[i].events & EPOLLOUT)
+      task->revents |= SILC_TASK_WRITE;
+    silc_list_add(schedule->fd_dispatch, task);
+  }
+
+  return ret;
+}
+
+#elif defined(HAVE_POLL) && defined(HAVE_SETRLIMIT) && defined(RLIMIT_NOFILE)
 
 /* Calls normal poll() system call. */
 
@@ -97,6 +155,7 @@ int silc_poll(SilcSchedule schedule, void *context)
     i++;
   }
   silc_hash_table_list_reset(&htl);
+  silc_list_init(schedule->fd_dispatch, struct SilcTaskStruct, next);
 
   if (schedule->has_timeout)
     timeout = ((schedule->timeout.tv_sec * 1000) +
@@ -115,12 +174,15 @@ int silc_poll(SilcSchedule schedule, void *context)
     if (!silc_hash_table_find(schedule->fd_queue, SILC_32_TO_PTR(fds[i].fd),
                              NULL, (void **)&task))
       continue;
+    if (!task->header.valid || !task->events)
+      continue;
 
     fd = fds[i].revents;
     if (fd & (POLLIN | POLLPRI | POLLERR | POLLHUP | POLLNVAL))
       task->revents |= SILC_TASK_READ;
     if (fd & POLLOUT)
       task->revents |= SILC_TASK_WRITE;
+    silc_list_add(schedule->fd_dispatch, task);
   }
 
   return ret;
@@ -161,6 +223,7 @@ int silc_select(SilcSchedule schedule, void *context)
     task->revents = 0;
   }
   silc_hash_table_list_reset(&htl);
+  silc_list_init(schedule->fd_dispatch, struct SilcTaskStruct, next);
 
   SILC_SCHEDULE_UNLOCK(schedule);
   ret = select(max_fd + 1, &in, &out, NULL, (schedule->has_timeout ?
@@ -171,7 +234,7 @@ int silc_select(SilcSchedule schedule, void *context)
 
   silc_hash_table_list(schedule->fd_queue, &htl);
   while (silc_hash_table_get(&htl, (void **)&fd, (void **)&task)) {
-    if (!task->events)
+    if (!task->header.valid || !task->events)
       continue;
 
 #ifdef FD_SETSIZE
@@ -183,6 +246,7 @@ int silc_select(SilcSchedule schedule, void *context)
       task->revents |= SILC_TASK_READ;
     if (FD_ISSET(fd, &out))
       task->revents |= SILC_TASK_WRITE;
+    silc_list_add(schedule->fd_dispatch, task);
   }
   silc_hash_table_list_reset(&htl);
 
@@ -191,6 +255,44 @@ int silc_select(SilcSchedule schedule, void *context)
 
 #endif /* HAVE_POLL && HAVE_SETRLIMIT && RLIMIT_NOFILE */
 
+/* Schedule `task' with events `event_mask'. Zero `event_mask' unschedules. */
+
+SilcBool silc_schedule_internal_schedule_fd(SilcSchedule schedule,
+                                           void *context,
+                                           SilcTaskFd task,
+                                           SilcTaskEvent event_mask)
+{
+#if defined(HAVE_EPOLL_WAIT)
+  SilcUnixScheduler internal = (SilcUnixScheduler)context;
+  struct epoll_event event;
+
+  event.events = 0;
+  if (task->events & SILC_TASK_READ)
+    event.events |= (EPOLLIN | EPOLLPRI);
+  if (task->events & SILC_TASK_WRITE)
+    event.events |= EPOLLOUT;
+
+  /* Zero mask unschedules task */
+  if (silc_unlikely(!event.events)) {
+    epoll_ctl(internal->epfd, EPOLL_CTL_DEL, task->fd, &event);
+    return TRUE;
+  }
+
+  /* Schedule the task */
+  if (silc_unlikely(!task->scheduled)) {
+    event.data.ptr = task;
+    epoll_ctl(internal->epfd, EPOLL_CTL_ADD, task->fd, &event);
+    task->scheduled = TRUE;
+    return TRUE;
+  }
+
+  /* Schedule for specific mask */
+  event.data.ptr = task;
+  epoll_ctl(internal->epfd, EPOLL_CTL_MOD, task->fd, &event);
+#endif /* HAVE_EPOLL_WAIT */
+  return TRUE;
+}
+
 #ifdef SILC_THREADS
 
 SILC_TASK_CALLBACK(silc_schedule_wakeup_cb)
@@ -220,7 +322,17 @@ void *silc_schedule_internal_init(SilcSchedule schedule,
   if (!internal)
     return NULL;
 
-#if defined(HAVE_POLL) && defined(HAVE_SETRLIMIT) && defined(RLIMIT_NOFILE)
+#if defined(HAVE_EPOLL_WAIT)
+  internal->epfd = epoll_create(4);
+  if (internal->epfd < 0)
+    return NULL;
+  internal->fds = silc_calloc(4, sizeof(*internal->fds));
+  if (!internal->fds) {
+    close(internal->epfd);
+    return NULL;
+  }
+  internal->fds_count = 4;
+#elif defined(HAVE_POLL) && defined(HAVE_SETRLIMIT) && defined(RLIMIT_NOFILE)
   getrlimit(RLIMIT_NOFILE, &internal->nofile);
 
   if (schedule->max_tasks > 0) {
@@ -291,7 +403,10 @@ void silc_schedule_internal_uninit(SilcSchedule schedule, void *context)
   close(internal->wakeup_pipe[1]);
 #endif
 
-#if defined(HAVE_POLL) && defined(HAVE_SETRLIMIT) && defined(RLIMIT_NOFILE)
+#if defined(HAVE_EPOLL_WAIT)
+  close(internal->epfd);
+  silc_free(internal->fds);
+#elif defined(HAVE_POLL) && defined(HAVE_SETRLIMIT) && defined(RLIMIT_NOFILE)
   silc_free(internal->fds);
 #endif /* HAVE_POLL && HAVE_SETRLIMIT && RLIMIT_NOFILE */
 
@@ -449,11 +564,14 @@ const SilcScheduleOps schedule_ops =
 {
   silc_schedule_internal_init,
   silc_schedule_internal_uninit,
-#if defined(HAVE_POLL) && defined(HAVE_SETRLIMIT) && defined(RLIMIT_NOFILE)
+#if defined(HAVE_EPOLL_WAIT)
+  silc_epoll,
+#elif defined(HAVE_POLL) && defined(HAVE_SETRLIMIT) && defined(RLIMIT_NOFILE)
   silc_poll,
 #else
   silc_select,
 #endif /* HAVE_POLL && HAVE_SETRLIMIT && RLIMIT_NOFILE */
+  silc_schedule_internal_schedule_fd,
   silc_schedule_internal_wakeup,
   silc_schedule_internal_signal_register,
   silc_schedule_internal_signal_unregister,
index 8b2f7ff44d2b488e459c38d837bf828dd1068b11..e171369b693a52fa3b00b5631eb7e63c27d4ff98 100644 (file)
@@ -168,8 +168,10 @@ int silc_socket_stream_write(SilcStream stream, const unsigned char *data,
   }
 
   SILC_LOG_DEBUG(("Wrote data %d bytes", ret));
-  silc_schedule_set_listen_fd(sock->schedule, sock->sock,
-                             SILC_TASK_READ, FALSE);
+  if (silc_schedule_get_fd_events(sock->schedule, sock->sock) &
+      SILC_TASK_WRITE)
+    silc_schedule_set_listen_fd(sock->schedule, sock->sock,
+                               SILC_TASK_READ, FALSE);
 
   return ret;
 }