diff mbox series

[meta-oe] nodejs: upgrade 20.11.0 -> 20.11.1

Message ID 20240305022433.2792409-1-changqing.li@windriver.com
State Accepted
Headers show
Series [meta-oe] nodejs: upgrade 20.11.0 -> 20.11.1 | expand

Commit Message

Changqing Li March 5, 2024, 2:24 a.m. UTC
From: Changqing Li <changqing.li@windriver.com>

Drop patches for revert io_uring support in libuv:
0001-Revert-io_uring-changes-from-libuv-1.46.0.patch
0002-Revert-io_uring-changes-from-libuv-1.45.0.patch

Change to just always disable io_uring in libuv, in this way, we don't have to
pick out io_uring related changes and revert them when internal libuv
upgraded.

Signed-off-by: Changqing Li <changqing.li@windriver.com>
---
 ...t-io_uring-changes-from-libuv-1.46.0.patch | 3414 -----------------
 ...ps-disable-io_uring-support-in-libuv.patch |   35 +
 ...t-io_uring-changes-from-libuv-1.45.0.patch | 1803 ---------
 .../{nodejs_20.11.0.bb => nodejs_20.11.1.bb}  |    5 +-
 4 files changed, 37 insertions(+), 5220 deletions(-)
 delete mode 100644 meta-oe/recipes-devtools/nodejs/nodejs/0001-Revert-io_uring-changes-from-libuv-1.46.0.patch
 create mode 100644 meta-oe/recipes-devtools/nodejs/nodejs/0001-deps-disable-io_uring-support-in-libuv.patch
 delete mode 100644 meta-oe/recipes-devtools/nodejs/nodejs/0002-Revert-io_uring-changes-from-libuv-1.45.0.patch
 rename meta-oe/recipes-devtools/nodejs/{nodejs_20.11.0.bb => nodejs_20.11.1.bb} (96%)
diff mbox series

Patch

diff --git a/meta-oe/recipes-devtools/nodejs/nodejs/0001-Revert-io_uring-changes-from-libuv-1.46.0.patch b/meta-oe/recipes-devtools/nodejs/nodejs/0001-Revert-io_uring-changes-from-libuv-1.46.0.patch
deleted file mode 100644
index d98463d0d..000000000
--- a/meta-oe/recipes-devtools/nodejs/nodejs/0001-Revert-io_uring-changes-from-libuv-1.46.0.patch
+++ /dev/null
@@ -1,3414 +0,0 @@ 
-From 937dfa70104b7401a7570a98cf6e0a534e250557 Mon Sep 17 00:00:00 2001
-From: Martin Jansa <martin.jansa@gmail.com>
-Date: Wed, 18 Oct 2023 22:43:55 +0200
-Subject: [PATCH] Revert io_uring changes from libuv-1.46.0
-
-Included in nodejs-20.4.0 with the libuv upgrade to 1.46.0 in:
-https://github.com/nodejs/node/commit/6199e1946c9abd59bd71a61870a4f6c85e100b18
-
-Revert "linux: work around io_uring IORING_OP_CLOSE bug (#4059)"
-This reverts commit 1752791c9ea89dbf54e2a20a9d9f899119a2d179.
-Revert "linux: fs_read to use io_uring if iovcnt > IOV_MAX (#4023)"
-This reverts commit e7b96331703e929e75d93c574573c9736e34b0c0.
-Revert "linux: add some more iouring backed fs ops (#4012)"
-This reverts commit 962b8e626ceaaf296eecab1d008e8b70dc6da5e7.
-Revert "android: disable io_uring support (#4016)"
-This reverts commit 281e6185cc9e77c582ff6ca9e4c00d57e6b90d95.
-Revert "unix,win: replace QUEUE with struct uv__queue (#4022)"
-This reverts commit 1b01b786c0180d29f07dccbb98001a2b3148828a.
-Revert "fs: use WTF-8 on Windows (#2970)"
-This reverts commit 8f32a14afaaa47514a7d28e1e069a8329e2dd939.
-
-Dropped deps/uv/docs deps/uv/test changes as these dirs aren't included
-in nodejs tarballs.
-
-Signed-off-by: Martin Jansa <martin.jansa@gmail.com>
----
-Upstream-Status: Inappropriate [OE specific]
-
- deps/uv/include/uv.h                     |  10 +-
- deps/uv/include/uv/darwin.h              |   6 +-
- deps/uv/include/uv/linux.h               |   2 +-
- deps/uv/include/uv/threadpool.h          |   2 +-
- deps/uv/include/uv/unix.h                |  44 ++--
- deps/uv/include/uv/win.h                 |   4 +-
- deps/uv/src/queue.h                      | 154 ++++++-----
- deps/uv/src/threadpool.c                 |  72 +++---
- deps/uv/src/unix/aix.c                   |  14 +-
- deps/uv/src/unix/async.c                 |  52 ++--
- deps/uv/src/unix/core.c                  |  53 ++--
- deps/uv/src/unix/fs.c                    |  41 +--
- deps/uv/src/unix/fsevents.c              |  84 +++---
- deps/uv/src/unix/internal.h              |  11 -
- deps/uv/src/unix/kqueue.c                |  18 +-
- deps/uv/src/unix/linux.c                 | 257 +++----------------
- deps/uv/src/unix/loop-watcher.c          |  20 +-
- deps/uv/src/unix/loop.c                  |  28 +-
- deps/uv/src/unix/os390-syscalls.c        |  24 +-
- deps/uv/src/unix/os390-syscalls.h        |   2 +-
- deps/uv/src/unix/os390.c                 |  14 +-
- deps/uv/src/unix/pipe.c                  |   2 +-
- deps/uv/src/unix/posix-poll.c            |  14 +-
- deps/uv/src/unix/process.c               |  38 +--
- deps/uv/src/unix/signal.c                |   8 +-
- deps/uv/src/unix/stream.c                |  56 ++--
- deps/uv/src/unix/sunos.c                 |  18 +-
- deps/uv/src/unix/tcp.c                   |   4 +-
- deps/uv/src/unix/tty.c                   |   2 +-
- deps/uv/src/unix/udp.c                   |  76 +++---
- deps/uv/src/uv-common.c                  |  28 +-
- deps/uv/src/uv-common.h                  |   3 +-
- deps/uv/src/win/core.c                   |   6 +-
- deps/uv/src/win/fs.c                     | 311 ++++++++---------------
- deps/uv/src/win/handle-inl.h             |   2 +-
- deps/uv/src/win/pipe.c                   |  26 +-
- deps/uv/src/win/tcp.c                    |   4 +-
- deps/uv/src/win/udp.c                    |   4 +-
- deps/uv/test/test-fs.c                   |  53 ----
- deps/uv/test/test-list.h                 |   2 -
- deps/uv/test/test-queue-foreach-delete.c |   5 +-
- deps/uv/test/test-threadpool-cancel.c    |   5 -
- 42 files changed, 625 insertions(+), 954 deletions(-)
-
-diff --git a/deps/uv/include/uv.h b/deps/uv/include/uv.h
-index 02397dd0fdd..1af95289776 100644
---- a/deps/uv/include/uv.h
-+++ b/deps/uv/include/uv.h
-@@ -59,12 +59,6 @@ extern "C" {
- #include <stdio.h>
- #include <stdint.h>
- 
--/* Internal type, do not use. */
--struct uv__queue {
--  struct uv__queue* next;
--  struct uv__queue* prev;
--};
--
- #if defined(_WIN32)
- # include "uv/win.h"
- #else
-@@ -466,7 +460,7 @@ struct uv_shutdown_s {
-   uv_handle_type type;                                                        \
-   /* private */                                                               \
-   uv_close_cb close_cb;                                                       \
--  struct uv__queue handle_queue;                                              \
-+  void* handle_queue[2];                                                      \
-   union {                                                                     \
-     int fd;                                                                   \
-     void* reserved[4];                                                        \
-@@ -1870,7 +1864,7 @@ struct uv_loop_s {
-   void* data;
-   /* Loop reference counting. */
-   unsigned int active_handles;
--  struct uv__queue handle_queue;
-+  void* handle_queue[2];
-   union {
-     void* unused;
-     unsigned int count;
-diff --git a/deps/uv/include/uv/darwin.h b/deps/uv/include/uv/darwin.h
-index 06962bfda80..d226415820b 100644
---- a/deps/uv/include/uv/darwin.h
-+++ b/deps/uv/include/uv/darwin.h
-@@ -40,7 +40,7 @@
-   void* cf_state;                                                             \
-   uv_mutex_t cf_mutex;                                                        \
-   uv_sem_t cf_sem;                                                            \
--  struct uv__queue cf_signals;                                                \
-+  void* cf_signals[2];                                                        \
- 
- #define UV_PLATFORM_FS_EVENT_FIELDS                                           \
-   uv__io_t event_watcher;                                                     \
-@@ -48,8 +48,8 @@
-   int realpath_len;                                                           \
-   int cf_flags;                                                               \
-   uv_async_t* cf_cb;                                                          \
--  struct uv__queue cf_events;                                                 \
--  struct uv__queue cf_member;                                                 \
-+  void* cf_events[2];                                                         \
-+  void* cf_member[2];                                                         \
-   int cf_error;                                                               \
-   uv_mutex_t cf_mutex;                                                        \
- 
-diff --git a/deps/uv/include/uv/linux.h b/deps/uv/include/uv/linux.h
-index 9f22f8cf726..9b38405a190 100644
---- a/deps/uv/include/uv/linux.h
-+++ b/deps/uv/include/uv/linux.h
-@@ -28,7 +28,7 @@
-   int inotify_fd;                                                             \
- 
- #define UV_PLATFORM_FS_EVENT_FIELDS                                           \
--  struct uv__queue watchers;                                                  \
-+  void* watchers[2];                                                          \
-   int wd;                                                                     \
- 
- #endif /* UV_LINUX_H */
-diff --git a/deps/uv/include/uv/threadpool.h b/deps/uv/include/uv/threadpool.h
-index 24ce916fda4..9708ebdd530 100644
---- a/deps/uv/include/uv/threadpool.h
-+++ b/deps/uv/include/uv/threadpool.h
-@@ -31,7 +31,7 @@ struct uv__work {
-   void (*work)(struct uv__work *w);
-   void (*done)(struct uv__work *w, int status);
-   struct uv_loop_s* loop;
--  struct uv__queue wq;
-+  void* wq[2];
- };
- 
- #endif /* UV_THREADPOOL_H_ */
-diff --git a/deps/uv/include/uv/unix.h b/deps/uv/include/uv/unix.h
-index 09f88a56742..95447b8dd67 100644
---- a/deps/uv/include/uv/unix.h
-+++ b/deps/uv/include/uv/unix.h
-@@ -92,8 +92,8 @@ typedef struct uv__io_s uv__io_t;
- 
- struct uv__io_s {
-   uv__io_cb cb;
--  struct uv__queue pending_queue;
--  struct uv__queue watcher_queue;
-+  void* pending_queue[2];
-+  void* watcher_queue[2];
-   unsigned int pevents; /* Pending event mask i.e. mask at next tick. */
-   unsigned int events;  /* Current event mask. */
-   int fd;
-@@ -220,21 +220,21 @@ typedef struct {
- #define UV_LOOP_PRIVATE_FIELDS                                                \
-   unsigned long flags;                                                        \
-   int backend_fd;                                                             \
--  struct uv__queue pending_queue;                                             \
--  struct uv__queue watcher_queue;                                             \
-+  void* pending_queue[2];                                                     \
-+  void* watcher_queue[2];                                                     \
-   uv__io_t** watchers;                                                        \
-   unsigned int nwatchers;                                                     \
-   unsigned int nfds;                                                          \
--  struct uv__queue wq;                                                        \
-+  void* wq[2];                                                                \
-   uv_mutex_t wq_mutex;                                                        \
-   uv_async_t wq_async;                                                        \
-   uv_rwlock_t cloexec_lock;                                                   \
-   uv_handle_t* closing_handles;                                               \
--  struct uv__queue process_handles;                                           \
--  struct uv__queue prepare_handles;                                           \
--  struct uv__queue check_handles;                                             \
--  struct uv__queue idle_handles;                                              \
--  struct uv__queue async_handles;                                             \
-+  void* process_handles[2];                                                   \
-+  void* prepare_handles[2];                                                   \
-+  void* check_handles[2];                                                     \
-+  void* idle_handles[2];                                                      \
-+  void* async_handles[2];                                                     \
-   void (*async_unused)(void);  /* TODO(bnoordhuis) Remove in libuv v2. */     \
-   uv__io_t async_io_watcher;                                                  \
-   int async_wfd;                                                              \
-@@ -257,7 +257,7 @@ typedef struct {
- #define UV_PRIVATE_REQ_TYPES /* empty */
- 
- #define UV_WRITE_PRIVATE_FIELDS                                               \
--  struct uv__queue queue;                                                     \
-+  void* queue[2];                                                             \
-   unsigned int write_index;                                                   \
-   uv_buf_t* bufs;                                                             \
-   unsigned int nbufs;                                                         \
-@@ -265,12 +265,12 @@ typedef struct {
-   uv_buf_t bufsml[4];                                                         \
- 
- #define UV_CONNECT_PRIVATE_FIELDS                                             \
--  struct uv__queue queue;                                                     \
-+  void* queue[2];                                                             \
- 
- #define UV_SHUTDOWN_PRIVATE_FIELDS /* empty */
- 
- #define UV_UDP_SEND_PRIVATE_FIELDS                                            \
--  struct uv__queue queue;                                                     \
-+  void* queue[2];                                                             \
-   struct sockaddr_storage addr;                                               \
-   unsigned int nbufs;                                                         \
-   uv_buf_t* bufs;                                                             \
-@@ -286,8 +286,8 @@ typedef struct {
-   uv_connect_t *connect_req;                                                  \
-   uv_shutdown_t *shutdown_req;                                                \
-   uv__io_t io_watcher;                                                        \
--  struct uv__queue write_queue;                                               \
--  struct uv__queue write_completed_queue;                                     \
-+  void* write_queue[2];                                                       \
-+  void* write_completed_queue[2];                                             \
-   uv_connection_cb connection_cb;                                             \
-   int delayed_error;                                                          \
-   int accepted_fd;                                                            \
-@@ -300,8 +300,8 @@ typedef struct {
-   uv_alloc_cb alloc_cb;                                                       \
-   uv_udp_recv_cb recv_cb;                                                     \
-   uv__io_t io_watcher;                                                        \
--  struct uv__queue write_queue;                                               \
--  struct uv__queue write_completed_queue;                                     \
-+  void* write_queue[2];                                                       \
-+  void* write_completed_queue[2];                                             \
- 
- #define UV_PIPE_PRIVATE_FIELDS                                                \
-   const char* pipe_fname; /* NULL or strdup'ed */
-@@ -311,19 +311,19 @@ typedef struct {
- 
- #define UV_PREPARE_PRIVATE_FIELDS                                             \
-   uv_prepare_cb prepare_cb;                                                   \
--  struct uv__queue queue;                                                     \
-+  void* queue[2];                                                             \
- 
- #define UV_CHECK_PRIVATE_FIELDS                                               \
-   uv_check_cb check_cb;                                                       \
--  struct uv__queue queue;                                                     \
-+  void* queue[2];                                                             \
- 
- #define UV_IDLE_PRIVATE_FIELDS                                                \
-   uv_idle_cb idle_cb;                                                         \
--  struct uv__queue queue;                                                     \
-+  void* queue[2];                                                             \
- 
- #define UV_ASYNC_PRIVATE_FIELDS                                               \
-   uv_async_cb async_cb;                                                       \
--  struct uv__queue queue;                                                     \
-+  void* queue[2];                                                             \
-   int pending;                                                                \
- 
- #define UV_TIMER_PRIVATE_FIELDS                                               \
-@@ -352,7 +352,7 @@ typedef struct {
-   int retcode;
- 
- #define UV_PROCESS_PRIVATE_FIELDS                                             \
--  struct uv__queue queue;                                                     \
-+  void* queue[2];                                                             \
-   int status;                                                                 \
- 
- #define UV_FS_PRIVATE_FIELDS                                                  \
-diff --git a/deps/uv/include/uv/win.h b/deps/uv/include/uv/win.h
-index 6f8c47298e4..92a95fa15f1 100644
---- a/deps/uv/include/uv/win.h
-+++ b/deps/uv/include/uv/win.h
-@@ -357,7 +357,7 @@ typedef struct {
-   /* Counter to started timer */                                              \
-   uint64_t timer_counter;                                                     \
-   /* Threadpool */                                                            \
--  struct uv__queue wq;                                                        \
-+  void* wq[2];                                                                \
-   uv_mutex_t wq_mutex;                                                        \
-   uv_async_t wq_async;
- 
-@@ -486,7 +486,7 @@ typedef struct {
-     uint32_t payload_remaining;                                               \
-     uint64_t dummy; /* TODO: retained for ABI compat; remove this in v2.x. */ \
-   } ipc_data_frame;                                                           \
--  struct uv__queue ipc_xfer_queue;                                            \
-+  void* ipc_xfer_queue[2];                                                    \
-   int ipc_xfer_queue_length;                                                  \
-   uv_write_t* non_overlapped_writes_tail;                                     \
-   CRITICAL_SECTION readfile_thread_lock;                                      \
-diff --git a/deps/uv/src/queue.h b/deps/uv/src/queue.h
-index 5f8489e9bc5..ff3540a0a51 100644
---- a/deps/uv/src/queue.h
-+++ b/deps/uv/src/queue.h
-@@ -18,73 +18,91 @@
- 
- #include <stddef.h>
- 
--#define uv__queue_data(pointer, type, field)                                  \
--  ((type*) ((char*) (pointer) - offsetof(type, field)))
--
--#define uv__queue_foreach(q, h)                                               \
--  for ((q) = (h)->next; (q) != (h); (q) = (q)->next)
--
--static inline void uv__queue_init(struct uv__queue* q) {
--  q->next = q;
--  q->prev = q;
--}
--
--static inline int uv__queue_empty(const struct uv__queue* q) {
--  return q == q->next;
--}
--
--static inline struct uv__queue* uv__queue_head(const struct uv__queue* q) {
--  return q->next;
--}
--
--static inline struct uv__queue* uv__queue_next(const struct uv__queue* q) {
--  return q->next;
--}
--
--static inline void uv__queue_add(struct uv__queue* h, struct uv__queue* n) {
--  h->prev->next = n->next;
--  n->next->prev = h->prev;
--  h->prev = n->prev;
--  h->prev->next = h;
--}
--
--static inline void uv__queue_split(struct uv__queue* h,
--                                   struct uv__queue* q,
--                                   struct uv__queue* n) {
--  n->prev = h->prev;
--  n->prev->next = n;
--  n->next = q;
--  h->prev = q->prev;
--  h->prev->next = h;
--  q->prev = n;
--}
--
--static inline void uv__queue_move(struct uv__queue* h, struct uv__queue* n) {
--  if (uv__queue_empty(h))
--    uv__queue_init(n);
--  else
--    uv__queue_split(h, h->next, n);
--}
--
--static inline void uv__queue_insert_head(struct uv__queue* h,
--                                         struct uv__queue* q) {
--  q->next = h->next;
--  q->prev = h;
--  q->next->prev = q;
--  h->next = q;
--}
--
--static inline void uv__queue_insert_tail(struct uv__queue* h,
--                                         struct uv__queue* q) {
--  q->next = h;
--  q->prev = h->prev;
--  q->prev->next = q;
--  h->prev = q;
--}
--
--static inline void uv__queue_remove(struct uv__queue* q) {
--  q->prev->next = q->next;
--  q->next->prev = q->prev;
--}
-+typedef void *QUEUE[2];
-+
-+/* Private macros. */
-+#define QUEUE_NEXT(q)       (*(QUEUE **) &((*(q))[0]))
-+#define QUEUE_PREV(q)       (*(QUEUE **) &((*(q))[1]))
-+#define QUEUE_PREV_NEXT(q)  (QUEUE_NEXT(QUEUE_PREV(q)))
-+#define QUEUE_NEXT_PREV(q)  (QUEUE_PREV(QUEUE_NEXT(q)))
-+
-+/* Public macros. */
-+#define QUEUE_DATA(ptr, type, field)                                          \
-+  ((type *) ((char *) (ptr) - offsetof(type, field)))
-+
-+/* Important note: mutating the list while QUEUE_FOREACH is
-+ * iterating over its elements results in undefined behavior.
-+ */
-+#define QUEUE_FOREACH(q, h)                                                   \
-+  for ((q) = QUEUE_NEXT(h); (q) != (h); (q) = QUEUE_NEXT(q))
-+
-+#define QUEUE_EMPTY(q)                                                        \
-+  ((const QUEUE *) (q) == (const QUEUE *) QUEUE_NEXT(q))
-+
-+#define QUEUE_HEAD(q)                                                         \
-+  (QUEUE_NEXT(q))
-+
-+#define QUEUE_INIT(q)                                                         \
-+  do {                                                                        \
-+    QUEUE_NEXT(q) = (q);                                                      \
-+    QUEUE_PREV(q) = (q);                                                      \
-+  }                                                                           \
-+  while (0)
-+
-+#define QUEUE_ADD(h, n)                                                       \
-+  do {                                                                        \
-+    QUEUE_PREV_NEXT(h) = QUEUE_NEXT(n);                                       \
-+    QUEUE_NEXT_PREV(n) = QUEUE_PREV(h);                                       \
-+    QUEUE_PREV(h) = QUEUE_PREV(n);                                            \
-+    QUEUE_PREV_NEXT(h) = (h);                                                 \
-+  }                                                                           \
-+  while (0)
-+
-+#define QUEUE_SPLIT(h, q, n)                                                  \
-+  do {                                                                        \
-+    QUEUE_PREV(n) = QUEUE_PREV(h);                                            \
-+    QUEUE_PREV_NEXT(n) = (n);                                                 \
-+    QUEUE_NEXT(n) = (q);                                                      \
-+    QUEUE_PREV(h) = QUEUE_PREV(q);                                            \
-+    QUEUE_PREV_NEXT(h) = (h);                                                 \
-+    QUEUE_PREV(q) = (n);                                                      \
-+  }                                                                           \
-+  while (0)
-+
-+#define QUEUE_MOVE(h, n)                                                      \
-+  do {                                                                        \
-+    if (QUEUE_EMPTY(h))                                                       \
-+      QUEUE_INIT(n);                                                          \
-+    else {                                                                    \
-+      QUEUE* q = QUEUE_HEAD(h);                                               \
-+      QUEUE_SPLIT(h, q, n);                                                   \
-+    }                                                                         \
-+  }                                                                           \
-+  while (0)
-+
-+#define QUEUE_INSERT_HEAD(h, q)                                               \
-+  do {                                                                        \
-+    QUEUE_NEXT(q) = QUEUE_NEXT(h);                                            \
-+    QUEUE_PREV(q) = (h);                                                      \
-+    QUEUE_NEXT_PREV(q) = (q);                                                 \
-+    QUEUE_NEXT(h) = (q);                                                      \
-+  }                                                                           \
-+  while (0)
-+
-+#define QUEUE_INSERT_TAIL(h, q)                                               \
-+  do {                                                                        \
-+    QUEUE_NEXT(q) = (h);                                                      \
-+    QUEUE_PREV(q) = QUEUE_PREV(h);                                            \
-+    QUEUE_PREV_NEXT(q) = (q);                                                 \
-+    QUEUE_PREV(h) = (q);                                                      \
-+  }                                                                           \
-+  while (0)
-+
-+#define QUEUE_REMOVE(q)                                                       \
-+  do {                                                                        \
-+    QUEUE_PREV_NEXT(q) = QUEUE_NEXT(q);                                       \
-+    QUEUE_NEXT_PREV(q) = QUEUE_PREV(q);                                       \
-+  }                                                                           \
-+  while (0)
- 
- #endif /* QUEUE_H_ */
-diff --git a/deps/uv/src/threadpool.c b/deps/uv/src/threadpool.c
-index dbef67f2f10..51962bf0021 100644
---- a/deps/uv/src/threadpool.c
-+++ b/deps/uv/src/threadpool.c
-@@ -37,10 +37,10 @@ static unsigned int slow_io_work_running;
- static unsigned int nthreads;
- static uv_thread_t* threads;
- static uv_thread_t default_threads[4];
--static struct uv__queue exit_message;
--static struct uv__queue wq;
--static struct uv__queue run_slow_work_message;
--static struct uv__queue slow_io_pending_wq;
-+static QUEUE exit_message;
-+static QUEUE wq;
-+static QUEUE run_slow_work_message;
-+static QUEUE slow_io_pending_wq;
- 
- static unsigned int slow_work_thread_threshold(void) {
-   return (nthreads + 1) / 2;
-@@ -56,7 +56,7 @@ static void uv__cancelled(struct uv__work* w) {
-  */
- static void worker(void* arg) {
-   struct uv__work* w;
--  struct uv__queue* q;
-+  QUEUE* q;
-   int is_slow_work;
- 
-   uv_sem_post((uv_sem_t*) arg);
-@@ -68,49 +68,49 @@ static void worker(void* arg) {
- 
-     /* Keep waiting while either no work is present or only slow I/O
-        and we're at the threshold for that. */
--    while (uv__queue_empty(&wq) ||
--           (uv__queue_head(&wq) == &run_slow_work_message &&
--            uv__queue_next(&run_slow_work_message) == &wq &&
-+    while (QUEUE_EMPTY(&wq) ||
-+           (QUEUE_HEAD(&wq) == &run_slow_work_message &&
-+            QUEUE_NEXT(&run_slow_work_message) == &wq &&
-             slow_io_work_running >= slow_work_thread_threshold())) {
-       idle_threads += 1;
-       uv_cond_wait(&cond, &mutex);
-       idle_threads -= 1;
-     }
- 
--    q = uv__queue_head(&wq);
-+    q = QUEUE_HEAD(&wq);
-     if (q == &exit_message) {
-       uv_cond_signal(&cond);
-       uv_mutex_unlock(&mutex);
-       break;
-     }
- 
--    uv__queue_remove(q);
--    uv__queue_init(q);  /* Signal uv_cancel() that the work req is executing. */
-+    QUEUE_REMOVE(q);
-+    QUEUE_INIT(q);  /* Signal uv_cancel() that the work req is executing. */
- 
-     is_slow_work = 0;
-     if (q == &run_slow_work_message) {
-       /* If we're at the slow I/O threshold, re-schedule until after all
-          other work in the queue is done. */
-       if (slow_io_work_running >= slow_work_thread_threshold()) {
--        uv__queue_insert_tail(&wq, q);
-+        QUEUE_INSERT_TAIL(&wq, q);
-         continue;
-       }
- 
-       /* If we encountered a request to run slow I/O work but there is none
-          to run, that means it's cancelled => Start over. */
--      if (uv__queue_empty(&slow_io_pending_wq))
-+      if (QUEUE_EMPTY(&slow_io_pending_wq))
-         continue;
- 
-       is_slow_work = 1;
-       slow_io_work_running++;
- 
--      q = uv__queue_head(&slow_io_pending_wq);
--      uv__queue_remove(q);
--      uv__queue_init(q);
-+      q = QUEUE_HEAD(&slow_io_pending_wq);
-+      QUEUE_REMOVE(q);
-+      QUEUE_INIT(q);
- 
-       /* If there is more slow I/O work, schedule it to be run as well. */
--      if (!uv__queue_empty(&slow_io_pending_wq)) {
--        uv__queue_insert_tail(&wq, &run_slow_work_message);
-+      if (!QUEUE_EMPTY(&slow_io_pending_wq)) {
-+        QUEUE_INSERT_TAIL(&wq, &run_slow_work_message);
-         if (idle_threads > 0)
-           uv_cond_signal(&cond);
-       }
-@@ -118,13 +118,13 @@ static void worker(void* arg) {
- 
-     uv_mutex_unlock(&mutex);
- 
--    w = uv__queue_data(q, struct uv__work, wq);
-+    w = QUEUE_DATA(q, struct uv__work, wq);
-     w->work(w);
- 
-     uv_mutex_lock(&w->loop->wq_mutex);
-     w->work = NULL;  /* Signal uv_cancel() that the work req is done
-                         executing. */
--    uv__queue_insert_tail(&w->loop->wq, &w->wq);
-+    QUEUE_INSERT_TAIL(&w->loop->wq, &w->wq);
-     uv_async_send(&w->loop->wq_async);
-     uv_mutex_unlock(&w->loop->wq_mutex);
- 
-@@ -139,12 +139,12 @@ static void worker(void* arg) {
- }
- 
- 
--static void post(struct uv__queue* q, enum uv__work_kind kind) {
-+static void post(QUEUE* q, enum uv__work_kind kind) {
-   uv_mutex_lock(&mutex);
-   if (kind == UV__WORK_SLOW_IO) {
-     /* Insert into a separate queue. */
--    uv__queue_insert_tail(&slow_io_pending_wq, q);
--    if (!uv__queue_empty(&run_slow_work_message)) {
-+    QUEUE_INSERT_TAIL(&slow_io_pending_wq, q);
-+    if (!QUEUE_EMPTY(&run_slow_work_message)) {
-       /* Running slow I/O tasks is already scheduled => Nothing to do here.
-          The worker that runs said other task will schedule this one as well. */
-       uv_mutex_unlock(&mutex);
-@@ -153,7 +153,7 @@ static void post(struct uv__queue* q, enum uv__work_kind kind) {
-     q = &run_slow_work_message;
-   }
- 
--  uv__queue_insert_tail(&wq, q);
-+  QUEUE_INSERT_TAIL(&wq, q);
-   if (idle_threads > 0)
-     uv_cond_signal(&cond);
-   uv_mutex_unlock(&mutex);
-@@ -220,9 +220,9 @@ static void init_threads(void) {
-   if (uv_mutex_init(&mutex))
-     abort();
- 
--  uv__queue_init(&wq);
--  uv__queue_init(&slow_io_pending_wq);
--  uv__queue_init(&run_slow_work_message);
-+  QUEUE_INIT(&wq);
-+  QUEUE_INIT(&slow_io_pending_wq);
-+  QUEUE_INIT(&run_slow_work_message);
- 
-   if (uv_sem_init(&sem, 0))
-     abort();
-@@ -285,9 +285,9 @@ static int uv__work_cancel(uv_loop_t* loop, uv_req_t* req, struct uv__work* w) {
-   uv_mutex_lock(&mutex);
-   uv_mutex_lock(&w->loop->wq_mutex);
- 
--  cancelled = !uv__queue_empty(&w->wq) && w->work != NULL;
-+  cancelled = !QUEUE_EMPTY(&w->wq) && w->work != NULL;
-   if (cancelled)
--    uv__queue_remove(&w->wq);
-+    QUEUE_REMOVE(&w->wq);
- 
-   uv_mutex_unlock(&w->loop->wq_mutex);
-   uv_mutex_unlock(&mutex);
-@@ -297,7 +297,7 @@ static int uv__work_cancel(uv_loop_t* loop, uv_req_t* req, struct uv__work* w) {
- 
-   w->work = uv__cancelled;
-   uv_mutex_lock(&loop->wq_mutex);
--  uv__queue_insert_tail(&loop->wq, &w->wq);
-+  QUEUE_INSERT_TAIL(&loop->wq, &w->wq);
-   uv_async_send(&loop->wq_async);
-   uv_mutex_unlock(&loop->wq_mutex);
- 
-@@ -308,21 +308,21 @@ static int uv__work_cancel(uv_loop_t* loop, uv_req_t* req, struct uv__work* w) {
- void uv__work_done(uv_async_t* handle) {
-   struct uv__work* w;
-   uv_loop_t* loop;
--  struct uv__queue* q;
--  struct uv__queue wq;
-+  QUEUE* q;
-+  QUEUE wq;
-   int err;
-   int nevents;
- 
-   loop = container_of(handle, uv_loop_t, wq_async);
-   uv_mutex_lock(&loop->wq_mutex);
--  uv__queue_move(&loop->wq, &wq);
-+  QUEUE_MOVE(&loop->wq, &wq);
-   uv_mutex_unlock(&loop->wq_mutex);
- 
-   nevents = 0;
- 
--  while (!uv__queue_empty(&wq)) {
--    q = uv__queue_head(&wq);
--    uv__queue_remove(q);
-+  while (!QUEUE_EMPTY(&wq)) {
-+    q = QUEUE_HEAD(&wq);
-+    QUEUE_REMOVE(q);
- 
-     w = container_of(q, struct uv__work, wq);
-     err = (w->work == uv__cancelled) ? UV_ECANCELED : 0;
-diff --git a/deps/uv/src/unix/aix.c b/deps/uv/src/unix/aix.c
-index 3af3009a216..f1afbed49ec 100644
---- a/deps/uv/src/unix/aix.c
-+++ b/deps/uv/src/unix/aix.c
-@@ -136,7 +136,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-   struct pollfd pqry;
-   struct pollfd* pe;
-   struct poll_ctl pc;
--  struct uv__queue* q;
-+  QUEUE* q;
-   uv__io_t* w;
-   uint64_t base;
-   uint64_t diff;
-@@ -151,18 +151,18 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-   int reset_timeout;
- 
-   if (loop->nfds == 0) {
--    assert(uv__queue_empty(&loop->watcher_queue));
-+    assert(QUEUE_EMPTY(&loop->watcher_queue));
-     return;
-   }
- 
-   lfields = uv__get_internal_fields(loop);
- 
--  while (!uv__queue_empty(&loop->watcher_queue)) {
--    q = uv__queue_head(&loop->watcher_queue);
--    uv__queue_remove(q);
--    uv__queue_init(q);
-+  while (!QUEUE_EMPTY(&loop->watcher_queue)) {
-+    q = QUEUE_HEAD(&loop->watcher_queue);
-+    QUEUE_REMOVE(q);
-+    QUEUE_INIT(q);
- 
--    w = uv__queue_data(q, uv__io_t, watcher_queue);
-+    w = QUEUE_DATA(q, uv__io_t, watcher_queue);
-     assert(w->pevents != 0);
-     assert(w->fd >= 0);
-     assert(w->fd < (int) loop->nwatchers);
-diff --git a/deps/uv/src/unix/async.c b/deps/uv/src/unix/async.c
-index 0ff2669e30a..5751b6d02be 100644
---- a/deps/uv/src/unix/async.c
-+++ b/deps/uv/src/unix/async.c
-@@ -55,7 +55,7 @@ int uv_async_init(uv_loop_t* loop, uv_async_t* handle, uv_async_cb async_cb) {
-   handle->pending = 0;
-   handle->u.fd = 0; /* This will be used as a busy flag. */
- 
--  uv__queue_insert_tail(&loop->async_handles, &handle->queue);
-+  QUEUE_INSERT_TAIL(&loop->async_handles, &handle->queue);
-   uv__handle_start(handle);
- 
-   return 0;
-@@ -124,7 +124,7 @@ static void uv__async_spin(uv_async_t* handle) {
- 
- void uv__async_close(uv_async_t* handle) {
-   uv__async_spin(handle);
--  uv__queue_remove(&handle->queue);
-+  QUEUE_REMOVE(&handle->queue);
-   uv__handle_stop(handle);
- }
- 
-@@ -132,8 +132,8 @@ void uv__async_close(uv_async_t* handle) {
- static void uv__async_io(uv_loop_t* loop, uv__io_t* w, unsigned int events) {
-   char buf[1024];
-   ssize_t r;
--  struct uv__queue queue;
--  struct uv__queue* q;
-+  QUEUE queue;
-+  QUEUE* q;
-   uv_async_t* h;
-   _Atomic int *pending;
- 
-@@ -157,13 +157,13 @@ static void uv__async_io(uv_loop_t* loop, uv__io_t* w, unsigned int events) {
-     abort();
-   }
- 
--  uv__queue_move(&loop->async_handles, &queue);
--  while (!uv__queue_empty(&queue)) {
--    q = uv__queue_head(&queue);
--    h = uv__queue_data(q, uv_async_t, queue);
-+  QUEUE_MOVE(&loop->async_handles, &queue);
-+  while (!QUEUE_EMPTY(&queue)) {
-+    q = QUEUE_HEAD(&queue);
-+    h = QUEUE_DATA(q, uv_async_t, queue);
- 
--    uv__queue_remove(q);
--    uv__queue_insert_tail(&loop->async_handles, q);
-+    QUEUE_REMOVE(q);
-+    QUEUE_INSERT_TAIL(&loop->async_handles, q);
- 
-     /* Atomically fetch and clear pending flag */
-     pending = (_Atomic int*) &h->pending;
-@@ -241,8 +241,8 @@ static int uv__async_start(uv_loop_t* loop) {
- 
- 
- void uv__async_stop(uv_loop_t* loop) {
--  struct uv__queue queue;
--  struct uv__queue* q;
-+  QUEUE queue;
-+  QUEUE* q;
-   uv_async_t* h;
- 
-   if (loop->async_io_watcher.fd == -1)
-@@ -251,13 +251,13 @@ void uv__async_stop(uv_loop_t* loop) {
-   /* Make sure no other thread is accessing the async handle fd after the loop
-    * cleanup.
-    */
--  uv__queue_move(&loop->async_handles, &queue);
--  while (!uv__queue_empty(&queue)) {
--    q = uv__queue_head(&queue);
--    h = uv__queue_data(q, uv_async_t, queue);
-+  QUEUE_MOVE(&loop->async_handles, &queue);
-+  while (!QUEUE_EMPTY(&queue)) {
-+    q = QUEUE_HEAD(&queue);
-+    h = QUEUE_DATA(q, uv_async_t, queue);
- 
--    uv__queue_remove(q);
--    uv__queue_insert_tail(&loop->async_handles, q);
-+    QUEUE_REMOVE(q);
-+    QUEUE_INSERT_TAIL(&loop->async_handles, q);
- 
-     uv__async_spin(h);
-   }
-@@ -275,20 +275,20 @@ void uv__async_stop(uv_loop_t* loop) {
- 
- 
- int uv__async_fork(uv_loop_t* loop) {
--  struct uv__queue queue;
--  struct uv__queue* q;
-+  QUEUE queue;
-+  QUEUE* q;
-   uv_async_t* h;
- 
-   if (loop->async_io_watcher.fd == -1) /* never started */
-     return 0;
- 
--  uv__queue_move(&loop->async_handles, &queue);
--  while (!uv__queue_empty(&queue)) {
--    q = uv__queue_head(&queue);
--    h = uv__queue_data(q, uv_async_t, queue);
-+  QUEUE_MOVE(&loop->async_handles, &queue);
-+  while (!QUEUE_EMPTY(&queue)) {
-+    q = QUEUE_HEAD(&queue);
-+    h = QUEUE_DATA(q, uv_async_t, queue);
- 
--    uv__queue_remove(q);
--    uv__queue_insert_tail(&loop->async_handles, q);
-+    QUEUE_REMOVE(q);
-+    QUEUE_INSERT_TAIL(&loop->async_handles, q);
- 
-     /* The state of any thread that set pending is now likely corrupt in this
-      * child because the user called fork, so just clear these flags and move
-diff --git a/deps/uv/src/unix/core.c b/deps/uv/src/unix/core.c
-index 25c5181f370..55aa962787e 100644
---- a/deps/uv/src/unix/core.c
-+++ b/deps/uv/src/unix/core.c
-@@ -344,7 +344,7 @@ static void uv__finish_close(uv_handle_t* handle) {
-   }
- 
-   uv__handle_unref(handle);
--  uv__queue_remove(&handle->handle_queue);
-+  QUEUE_REMOVE(&handle->handle_queue);
- 
-   if (handle->close_cb) {
-     handle->close_cb(handle);
-@@ -380,7 +380,7 @@ int uv_backend_fd(const uv_loop_t* loop) {
- static int uv__loop_alive(const uv_loop_t* loop) {
-   return uv__has_active_handles(loop) ||
-          uv__has_active_reqs(loop) ||
--         !uv__queue_empty(&loop->pending_queue) ||
-+         !QUEUE_EMPTY(&loop->pending_queue) ||
-          loop->closing_handles != NULL;
- }
- 
-@@ -389,8 +389,8 @@ static int uv__backend_timeout(const uv_loop_t* loop) {
-   if (loop->stop_flag == 0 &&
-       /* uv__loop_alive(loop) && */
-       (uv__has_active_handles(loop) || uv__has_active_reqs(loop)) &&
--      uv__queue_empty(&loop->pending_queue) &&
--      uv__queue_empty(&loop->idle_handles) &&
-+      QUEUE_EMPTY(&loop->pending_queue) &&
-+      QUEUE_EMPTY(&loop->idle_handles) &&
-       (loop->flags & UV_LOOP_REAP_CHILDREN) == 0 &&
-       loop->closing_handles == NULL)
-     return uv__next_timeout(loop);
-@@ -399,7 +399,7 @@ static int uv__backend_timeout(const uv_loop_t* loop) {
- 
- 
- int uv_backend_timeout(const uv_loop_t* loop) {
--  if (uv__queue_empty(&loop->watcher_queue))
-+  if (QUEUE_EMPTY(&loop->watcher_queue))
-     return uv__backend_timeout(loop);
-   /* Need to call uv_run to update the backend fd state. */
-   return 0;
-@@ -431,8 +431,7 @@ int uv_run(uv_loop_t* loop, uv_run_mode mode) {
- 
-   while (r != 0 && loop->stop_flag == 0) {
-     can_sleep =
--        uv__queue_empty(&loop->pending_queue) &&
--        uv__queue_empty(&loop->idle_handles);
-+        QUEUE_EMPTY(&loop->pending_queue) && QUEUE_EMPTY(&loop->idle_handles);
- 
-     uv__run_pending(loop);
-     uv__run_idle(loop);
-@@ -448,7 +447,7 @@ int uv_run(uv_loop_t* loop, uv_run_mode mode) {
- 
-     /* Process immediate callbacks (e.g. write_cb) a small fixed number of
-      * times to avoid loop starvation.*/
--    for (r = 0; r < 8 && !uv__queue_empty(&loop->pending_queue); r++)
-+    for (r = 0; r < 8 && !QUEUE_EMPTY(&loop->pending_queue); r++)
-       uv__run_pending(loop);
- 
-     /* Run one final update on the provider_idle_time in case uv__io_poll
-@@ -827,17 +826,17 @@ int uv_fileno(const uv_handle_t* handle, uv_os_fd_t* fd) {
- 
- 
- static void uv__run_pending(uv_loop_t* loop) {
--  struct uv__queue* q;
--  struct uv__queue pq;
-+  QUEUE* q;
-+  QUEUE pq;
-   uv__io_t* w;
- 
--  uv__queue_move(&loop->pending_queue, &pq);
-+  QUEUE_MOVE(&loop->pending_queue, &pq);
- 
--  while (!uv__queue_empty(&pq)) {
--    q = uv__queue_head(&pq);
--    uv__queue_remove(q);
--    uv__queue_init(q);
--    w = uv__queue_data(q, uv__io_t, pending_queue);
-+  while (!QUEUE_EMPTY(&pq)) {
-+    q = QUEUE_HEAD(&pq);
-+    QUEUE_REMOVE(q);
-+    QUEUE_INIT(q);
-+    w = QUEUE_DATA(q, uv__io_t, pending_queue);
-     w->cb(loop, w, POLLOUT);
-   }
- }
-@@ -892,8 +891,8 @@ static void maybe_resize(uv_loop_t* loop, unsigned int len) {
- void uv__io_init(uv__io_t* w, uv__io_cb cb, int fd) {
-   assert(cb != NULL);
-   assert(fd >= -1);
--  uv__queue_init(&w->pending_queue);
--  uv__queue_init(&w->watcher_queue);
-+  QUEUE_INIT(&w->pending_queue);
-+  QUEUE_INIT(&w->watcher_queue);
-   w->cb = cb;
-   w->fd = fd;
-   w->events = 0;
-@@ -919,8 +918,8 @@ void uv__io_start(uv_loop_t* loop, uv__io_t* w, unsigned int events) {
-     return;
- #endif
- 
--  if (uv__queue_empty(&w->watcher_queue))
--    uv__queue_insert_tail(&loop->watcher_queue, &w->watcher_queue);
-+  if (QUEUE_EMPTY(&w->watcher_queue))
-+    QUEUE_INSERT_TAIL(&loop->watcher_queue, &w->watcher_queue);
- 
-   if (loop->watchers[w->fd] == NULL) {
-     loop->watchers[w->fd] = w;
-@@ -945,8 +944,8 @@ void uv__io_stop(uv_loop_t* loop, uv__io_t* w, unsigned int events) {
-   w->pevents &= ~events;
- 
-   if (w->pevents == 0) {
--    uv__queue_remove(&w->watcher_queue);
--    uv__queue_init(&w->watcher_queue);
-+    QUEUE_REMOVE(&w->watcher_queue);
-+    QUEUE_INIT(&w->watcher_queue);
-     w->events = 0;
- 
-     if (w == loop->watchers[w->fd]) {
-@@ -955,14 +954,14 @@ void uv__io_stop(uv_loop_t* loop, uv__io_t* w, unsigned int events) {
-       loop->nfds--;
-     }
-   }
--  else if (uv__queue_empty(&w->watcher_queue))
--    uv__queue_insert_tail(&loop->watcher_queue, &w->watcher_queue);
-+  else if (QUEUE_EMPTY(&w->watcher_queue))
-+    QUEUE_INSERT_TAIL(&loop->watcher_queue, &w->watcher_queue);
- }
- 
- 
- void uv__io_close(uv_loop_t* loop, uv__io_t* w) {
-   uv__io_stop(loop, w, POLLIN | POLLOUT | UV__POLLRDHUP | UV__POLLPRI);
--  uv__queue_remove(&w->pending_queue);
-+  QUEUE_REMOVE(&w->pending_queue);
- 
-   /* Remove stale events for this file descriptor */
-   if (w->fd != -1)
-@@ -971,8 +970,8 @@ void uv__io_close(uv_loop_t* loop, uv__io_t* w) {
- 
- 
- void uv__io_feed(uv_loop_t* loop, uv__io_t* w) {
--  if (uv__queue_empty(&w->pending_queue))
--    uv__queue_insert_tail(&loop->pending_queue, &w->pending_queue);
-+  if (QUEUE_EMPTY(&w->pending_queue))
-+    QUEUE_INSERT_TAIL(&loop->pending_queue, &w->pending_queue);
- }
- 
- 
-diff --git a/deps/uv/src/unix/fs.c b/deps/uv/src/unix/fs.c
-index 6b051c124f2..00d385c24b7 100644
---- a/deps/uv/src/unix/fs.c
-+++ b/deps/uv/src/unix/fs.c
-@@ -62,6 +62,7 @@
- 
- #if defined(__linux__)
- # include <sys/sendfile.h>
-+# include <sys/utsname.h>
- #endif
- 
- #if defined(__sun)
-@@ -903,6 +904,31 @@ out:
- 
- 
- #ifdef __linux__
-+static unsigned uv__kernel_version(void) {
-+  static _Atomic unsigned cached_version;
-+  struct utsname u;
-+  unsigned version;
-+  unsigned major;
-+  unsigned minor;
-+  unsigned patch;
-+
-+  version = atomic_load_explicit(&cached_version, memory_order_relaxed);
-+  if (version != 0)
-+    return version;
-+
-+  if (-1 == uname(&u))
-+    return 0;
-+
-+  if (3 != sscanf(u.release, "%u.%u.%u", &major, &minor, &patch))
-+    return 0;
-+
-+  version = major * 65536 + minor * 256 + patch;
-+  atomic_store_explicit(&cached_version, version, memory_order_relaxed);
-+
-+  return version;
-+}
-+
-+
- /* Pre-4.20 kernels have a bug where CephFS uses the RADOS copy-from command
-  * in copy_file_range() when it shouldn't. There is no workaround except to
-  * fall back to a regular copy.
-@@ -1905,9 +1931,6 @@ int uv_fs_link(uv_loop_t* loop,
-                uv_fs_cb cb) {
-   INIT(LINK);
-   PATH2;
--  if (cb != NULL)
--    if (uv__iou_fs_link(loop, req))
--      return 0;
-   POST;
- }
- 
-@@ -1920,9 +1943,6 @@ int uv_fs_mkdir(uv_loop_t* loop,
-   INIT(MKDIR);
-   PATH;
-   req->mode = mode;
--  if (cb != NULL)
--    if (uv__iou_fs_mkdir(loop, req))
--      return 0;
-   POST;
- }
- 
-@@ -2074,9 +2094,6 @@ int uv_fs_rename(uv_loop_t* loop,
-                  uv_fs_cb cb) {
-   INIT(RENAME);
-   PATH2;
--  if (cb != NULL)
--    if (uv__iou_fs_rename(loop, req))
--      return 0;
-   POST;
- }
- 
-@@ -2123,9 +2140,6 @@ int uv_fs_symlink(uv_loop_t* loop,
-   INIT(SYMLINK);
-   PATH2;
-   req->flags = flags;
--  if (cb != NULL)
--    if (uv__iou_fs_symlink(loop, req))
--      return 0;
-   POST;
- }
- 
-@@ -2133,9 +2147,6 @@ int uv_fs_symlink(uv_loop_t* loop,
- int uv_fs_unlink(uv_loop_t* loop, uv_fs_t* req, const char* path, uv_fs_cb cb) {
-   INIT(UNLINK);
-   PATH;
--  if (cb != NULL)
--    if (uv__iou_fs_unlink(loop, req))
--      return 0;
-   POST;
- }
- 
-diff --git a/deps/uv/src/unix/fsevents.c b/deps/uv/src/unix/fsevents.c
-index df703f3635f..0535b4547aa 100644
---- a/deps/uv/src/unix/fsevents.c
-+++ b/deps/uv/src/unix/fsevents.c
-@@ -80,13 +80,13 @@ enum uv__cf_loop_signal_type_e {
- typedef enum uv__cf_loop_signal_type_e uv__cf_loop_signal_type_t;
- 
- struct uv__cf_loop_signal_s {
--  struct uv__queue member;
-+  QUEUE member;
-   uv_fs_event_t* handle;
-   uv__cf_loop_signal_type_t type;
- };
- 
- struct uv__fsevents_event_s {
--  struct uv__queue member;
-+  QUEUE member;
-   int events;
-   char path[1];
- };
-@@ -98,7 +98,7 @@ struct uv__cf_loop_state_s {
-   FSEventStreamRef fsevent_stream;
-   uv_sem_t fsevent_sem;
-   uv_mutex_t fsevent_mutex;
--  struct uv__queue fsevent_handles;
-+  void* fsevent_handles[2];
-   unsigned int fsevent_handle_count;
- };
- 
-@@ -150,22 +150,22 @@ static void (*pFSEventStreamStop)(FSEventStreamRef);
- 
- #define UV__FSEVENTS_PROCESS(handle, block)                                   \
-     do {                                                                      \
--      struct uv__queue events;                                                \
--      struct uv__queue* q;                                                    \
-+      QUEUE events;                                                           \
-+      QUEUE* q;                                                               \
-       uv__fsevents_event_t* event;                                            \
-       int err;                                                                \
-       uv_mutex_lock(&(handle)->cf_mutex);                                     \
-       /* Split-off all events and empty original queue */                     \
--      uv__queue_move(&(handle)->cf_events, &events);                          \
-+      QUEUE_MOVE(&(handle)->cf_events, &events);                              \
-       /* Get error (if any) and zero original one */                          \
-       err = (handle)->cf_error;                                               \
-       (handle)->cf_error = 0;                                                 \
-       uv_mutex_unlock(&(handle)->cf_mutex);                                   \
-       /* Loop through events, deallocating each after processing */           \
--      while (!uv__queue_empty(&events)) {                                     \
--        q = uv__queue_head(&events);                                          \
--        event = uv__queue_data(q, uv__fsevents_event_t, member);              \
--        uv__queue_remove(q);                                                  \
-+      while (!QUEUE_EMPTY(&events)) {                                         \
-+        q = QUEUE_HEAD(&events);                                              \
-+        event = QUEUE_DATA(q, uv__fsevents_event_t, member);                  \
-+        QUEUE_REMOVE(q);                                                      \
-         /* NOTE: Checking uv__is_active() is required here, because handle    \
-          * callback may close handle and invoking it after it will lead to    \
-          * incorrect behaviour */                                             \
-@@ -193,14 +193,14 @@ static void uv__fsevents_cb(uv_async_t* cb) {
- 
- /* Runs in CF thread, pushed event into handle's event list */
- static void uv__fsevents_push_event(uv_fs_event_t* handle,
--                                    struct uv__queue* events,
-+                                    QUEUE* events,
-                                     int err) {
-   assert(events != NULL || err != 0);
-   uv_mutex_lock(&handle->cf_mutex);
- 
-   /* Concatenate two queues */
-   if (events != NULL)
--    uv__queue_add(&handle->cf_events, events);
-+    QUEUE_ADD(&handle->cf_events, events);
- 
-   /* Propagate error */
-   if (err != 0)
-@@ -224,12 +224,12 @@ static void uv__fsevents_event_cb(const FSEventStreamRef streamRef,
-   char* path;
-   char* pos;
-   uv_fs_event_t* handle;
--  struct uv__queue* q;
-+  QUEUE* q;
-   uv_loop_t* loop;
-   uv__cf_loop_state_t* state;
-   uv__fsevents_event_t* event;
-   FSEventStreamEventFlags flags;
--  struct uv__queue head;
-+  QUEUE head;
- 
-   loop = info;
-   state = loop->cf_state;
-@@ -238,9 +238,9 @@ static void uv__fsevents_event_cb(const FSEventStreamRef streamRef,
- 
-   /* For each handle */
-   uv_mutex_lock(&state->fsevent_mutex);
--  uv__queue_foreach(q, &state->fsevent_handles) {
--    handle = uv__queue_data(q, uv_fs_event_t, cf_member);
--    uv__queue_init(&head);
-+  QUEUE_FOREACH(q, &state->fsevent_handles) {
-+    handle = QUEUE_DATA(q, uv_fs_event_t, cf_member);
-+    QUEUE_INIT(&head);
- 
-     /* Process and filter out events */
-     for (i = 0; i < numEvents; i++) {
-@@ -318,10 +318,10 @@ static void uv__fsevents_event_cb(const FSEventStreamRef streamRef,
-           event->events = UV_CHANGE;
-       }
- 
--      uv__queue_insert_tail(&head, &event->member);
-+      QUEUE_INSERT_TAIL(&head, &event->member);
-     }
- 
--    if (!uv__queue_empty(&head))
-+    if (!QUEUE_EMPTY(&head))
-       uv__fsevents_push_event(handle, &head, 0);
-   }
-   uv_mutex_unlock(&state->fsevent_mutex);
-@@ -403,7 +403,7 @@ static void uv__fsevents_destroy_stream(uv__cf_loop_state_t* state) {
- static void uv__fsevents_reschedule(uv__cf_loop_state_t* state,
-                                     uv_loop_t* loop,
-                                     uv__cf_loop_signal_type_t type) {
--  struct uv__queue* q;
-+  QUEUE* q;
-   uv_fs_event_t* curr;
-   CFArrayRef cf_paths;
-   CFStringRef* paths;
-@@ -446,9 +446,9 @@ static void uv__fsevents_reschedule(uv__cf_loop_state_t* state,
- 
-     q = &state->fsevent_handles;
-     for (; i < path_count; i++) {
--      q = uv__queue_next(q);
-+      q = QUEUE_NEXT(q);
-       assert(q != &state->fsevent_handles);
--      curr = uv__queue_data(q, uv_fs_event_t, cf_member);
-+      curr = QUEUE_DATA(q, uv_fs_event_t, cf_member);
- 
-       assert(curr->realpath != NULL);
-       paths[i] =
-@@ -486,8 +486,8 @@ final:
- 
-     /* Broadcast error to all handles */
-     uv_mutex_lock(&state->fsevent_mutex);
--    uv__queue_foreach(q, &state->fsevent_handles) {
--      curr = uv__queue_data(q, uv_fs_event_t, cf_member);
-+    QUEUE_FOREACH(q, &state->fsevent_handles) {
-+      curr = QUEUE_DATA(q, uv_fs_event_t, cf_member);
-       uv__fsevents_push_event(curr, NULL, err);
-     }
-     uv_mutex_unlock(&state->fsevent_mutex);
-@@ -606,7 +606,7 @@ static int uv__fsevents_loop_init(uv_loop_t* loop) {
-   if (err)
-     goto fail_sem_init;
- 
--  uv__queue_init(&loop->cf_signals);
-+  QUEUE_INIT(&loop->cf_signals);
- 
-   err = uv_sem_init(&state->fsevent_sem, 0);
-   if (err)
-@@ -616,7 +616,7 @@ static int uv__fsevents_loop_init(uv_loop_t* loop) {
-   if (err)
-     goto fail_fsevent_mutex_init;
- 
--  uv__queue_init(&state->fsevent_handles);
-+  QUEUE_INIT(&state->fsevent_handles);
-   state->fsevent_need_reschedule = 0;
-   state->fsevent_handle_count = 0;
- 
-@@ -675,7 +675,7 @@ fail_mutex_init:
- void uv__fsevents_loop_delete(uv_loop_t* loop) {
-   uv__cf_loop_signal_t* s;
-   uv__cf_loop_state_t* state;
--  struct uv__queue* q;
-+  QUEUE* q;
- 
-   if (loop->cf_state == NULL)
-     return;
-@@ -688,10 +688,10 @@ void uv__fsevents_loop_delete(uv_loop_t* loop) {
-   uv_mutex_destroy(&loop->cf_mutex);
- 
-   /* Free any remaining data */
--  while (!uv__queue_empty(&loop->cf_signals)) {
--    q = uv__queue_head(&loop->cf_signals);
--    s = uv__queue_data(q, uv__cf_loop_signal_t, member);
--    uv__queue_remove(q);
-+  while (!QUEUE_EMPTY(&loop->cf_signals)) {
-+    q = QUEUE_HEAD(&loop->cf_signals);
-+    s = QUEUE_DATA(q, uv__cf_loop_signal_t, member);
-+    QUEUE_REMOVE(q);
-     uv__free(s);
-   }
- 
-@@ -735,22 +735,22 @@ static void* uv__cf_loop_runner(void* arg) {
- static void uv__cf_loop_cb(void* arg) {
-   uv_loop_t* loop;
-   uv__cf_loop_state_t* state;
--  struct uv__queue* item;
--  struct uv__queue split_head;
-+  QUEUE* item;
-+  QUEUE split_head;
-   uv__cf_loop_signal_t* s;
- 
-   loop = arg;
-   state = loop->cf_state;
- 
-   uv_mutex_lock(&loop->cf_mutex);
--  uv__queue_move(&loop->cf_signals, &split_head);
-+  QUEUE_MOVE(&loop->cf_signals, &split_head);
-   uv_mutex_unlock(&loop->cf_mutex);
- 
--  while (!uv__queue_empty(&split_head)) {
--    item = uv__queue_head(&split_head);
--    uv__queue_remove(item);
-+  while (!QUEUE_EMPTY(&split_head)) {
-+    item = QUEUE_HEAD(&split_head);
-+    QUEUE_REMOVE(item);
- 
--    s = uv__queue_data(item, uv__cf_loop_signal_t, member);
-+    s = QUEUE_DATA(item, uv__cf_loop_signal_t, member);
- 
-     /* This was a termination signal */
-     if (s->handle == NULL)
-@@ -778,7 +778,7 @@ int uv__cf_loop_signal(uv_loop_t* loop,
-   item->type = type;
- 
-   uv_mutex_lock(&loop->cf_mutex);
--  uv__queue_insert_tail(&loop->cf_signals, &item->member);
-+  QUEUE_INSERT_TAIL(&loop->cf_signals, &item->member);
- 
-   state = loop->cf_state;
-   assert(state != NULL);
-@@ -807,7 +807,7 @@ int uv__fsevents_init(uv_fs_event_t* handle) {
-   handle->realpath_len = strlen(handle->realpath);
- 
-   /* Initialize event queue */
--  uv__queue_init(&handle->cf_events);
-+  QUEUE_INIT(&handle->cf_events);
-   handle->cf_error = 0;
- 
-   /*
-@@ -832,7 +832,7 @@ int uv__fsevents_init(uv_fs_event_t* handle) {
-   /* Insert handle into the list */
-   state = handle->loop->cf_state;
-   uv_mutex_lock(&state->fsevent_mutex);
--  uv__queue_insert_tail(&state->fsevent_handles, &handle->cf_member);
-+  QUEUE_INSERT_TAIL(&state->fsevent_handles, &handle->cf_member);
-   state->fsevent_handle_count++;
-   state->fsevent_need_reschedule = 1;
-   uv_mutex_unlock(&state->fsevent_mutex);
-@@ -872,7 +872,7 @@ int uv__fsevents_close(uv_fs_event_t* handle) {
-   /* Remove handle from  the list */
-   state = handle->loop->cf_state;
-   uv_mutex_lock(&state->fsevent_mutex);
--  uv__queue_remove(&handle->cf_member);
-+  QUEUE_REMOVE(&handle->cf_member);
-   state->fsevent_handle_count--;
-   state->fsevent_need_reschedule = 1;
-   uv_mutex_unlock(&state->fsevent_mutex);
-diff --git a/deps/uv/src/unix/internal.h b/deps/uv/src/unix/internal.h
-index fe588513603..6c5822e6a0d 100644
---- a/deps/uv/src/unix/internal.h
-+++ b/deps/uv/src/unix/internal.h
-@@ -335,30 +335,20 @@ int uv__iou_fs_close(uv_loop_t* loop, uv_fs_t* req);
- int uv__iou_fs_fsync_or_fdatasync(uv_loop_t* loop,
-                                   uv_fs_t* req,
-                                   uint32_t fsync_flags);
--int uv__iou_fs_link(uv_loop_t* loop, uv_fs_t* req);
--int uv__iou_fs_mkdir(uv_loop_t* loop, uv_fs_t* req);
- int uv__iou_fs_open(uv_loop_t* loop, uv_fs_t* req);
- int uv__iou_fs_read_or_write(uv_loop_t* loop,
-                              uv_fs_t* req,
-                              int is_read);
--int uv__iou_fs_rename(uv_loop_t* loop, uv_fs_t* req);
- int uv__iou_fs_statx(uv_loop_t* loop,
-                      uv_fs_t* req,
-                      int is_fstat,
-                      int is_lstat);
--int uv__iou_fs_symlink(uv_loop_t* loop, uv_fs_t* req);
--int uv__iou_fs_unlink(uv_loop_t* loop, uv_fs_t* req);
- #else
- #define uv__iou_fs_close(loop, req) 0
- #define uv__iou_fs_fsync_or_fdatasync(loop, req, fsync_flags) 0
--#define uv__iou_fs_link(loop, req) 0
--#define uv__iou_fs_mkdir(loop, req) 0
- #define uv__iou_fs_open(loop, req) 0
- #define uv__iou_fs_read_or_write(loop, req, is_read) 0
--#define uv__iou_fs_rename(loop, req) 0
- #define uv__iou_fs_statx(loop, req, is_fstat, is_lstat) 0
--#define uv__iou_fs_symlink(loop, req) 0
--#define uv__iou_fs_unlink(loop, req) 0
- #endif
- 
- #if defined(__APPLE__)
-@@ -439,7 +429,6 @@ int uv__statx(int dirfd,
-               struct uv__statx* statxbuf);
- void uv__statx_to_stat(const struct uv__statx* statxbuf, uv_stat_t* buf);
- ssize_t uv__getrandom(void* buf, size_t buflen, unsigned flags);
--unsigned uv__kernel_version(void);
- #endif
- 
- typedef int (*uv__peersockfunc)(int, struct sockaddr*, socklen_t*);
-diff --git a/deps/uv/src/unix/kqueue.c b/deps/uv/src/unix/kqueue.c
-index b78242d3be4..82916d65933 100644
---- a/deps/uv/src/unix/kqueue.c
-+++ b/deps/uv/src/unix/kqueue.c
-@@ -133,7 +133,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-   struct timespec spec;
-   unsigned int nevents;
-   unsigned int revents;
--  struct uv__queue* q;
-+  QUEUE* q;
-   uv__io_t* w;
-   uv_process_t* process;
-   sigset_t* pset;
-@@ -152,19 +152,19 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-   int reset_timeout;
- 
-   if (loop->nfds == 0) {
--    assert(uv__queue_empty(&loop->watcher_queue));
-+    assert(QUEUE_EMPTY(&loop->watcher_queue));
-     return;
-   }
- 
-   lfields = uv__get_internal_fields(loop);
-   nevents = 0;
- 
--  while (!uv__queue_empty(&loop->watcher_queue)) {
--    q = uv__queue_head(&loop->watcher_queue);
--    uv__queue_remove(q);
--    uv__queue_init(q);
-+  while (!QUEUE_EMPTY(&loop->watcher_queue)) {
-+    q = QUEUE_HEAD(&loop->watcher_queue);
-+    QUEUE_REMOVE(q);
-+    QUEUE_INIT(q);
- 
--    w = uv__queue_data(q, uv__io_t, watcher_queue);
-+    w = QUEUE_DATA(q, uv__io_t, watcher_queue);
-     assert(w->pevents != 0);
-     assert(w->fd >= 0);
-     assert(w->fd < (int) loop->nwatchers);
-@@ -307,8 +307,8 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
- 
-       /* Handle kevent NOTE_EXIT results */
-       if (ev->filter == EVFILT_PROC) {
--        uv__queue_foreach(q, &loop->process_handles) {
--          process = uv__queue_data(q, uv_process_t, queue);
-+        QUEUE_FOREACH(q, &loop->process_handles) {
-+          process = QUEUE_DATA(q, uv_process_t, queue);
-           if (process->pid == fd) {
-             process->flags |= UV_HANDLE_REAP;
-             loop->flags |= UV_LOOP_REAP_CHILDREN;
-diff --git a/deps/uv/src/unix/linux.c b/deps/uv/src/unix/linux.c
-index 48b9c2c43e1..5f84ad0eea3 100644
---- a/deps/uv/src/unix/linux.c
-+++ b/deps/uv/src/unix/linux.c
-@@ -48,7 +48,6 @@
- #include <sys/sysinfo.h>
- #include <sys/sysmacros.h>
- #include <sys/types.h>
--#include <sys/utsname.h>
- #include <time.h>
- #include <unistd.h>
- 
-@@ -151,11 +150,6 @@ enum {
-   UV__IORING_OP_CLOSE = 19,
-   UV__IORING_OP_STATX = 21,
-   UV__IORING_OP_EPOLL_CTL = 29,
--  UV__IORING_OP_RENAMEAT = 35,
--  UV__IORING_OP_UNLINKAT = 36,
--  UV__IORING_OP_MKDIRAT = 37,
--  UV__IORING_OP_SYMLINKAT = 38,
--  UV__IORING_OP_LINKAT = 39,
- };
- 
- enum {
-@@ -168,10 +162,6 @@ enum {
-   UV__IORING_SQ_CQ_OVERFLOW = 2u,
- };
- 
--enum {
--  UV__MKDIRAT_SYMLINKAT_LINKAT = 1u,
--};
--
- struct uv__io_cqring_offsets {
-   uint32_t head;
-   uint32_t tail;
-@@ -267,7 +257,7 @@ STATIC_ASSERT(EPOLL_CTL_MOD < 4);
- 
- struct watcher_list {
-   RB_ENTRY(watcher_list) entry;
--  struct uv__queue watchers;
-+  QUEUE watchers;
-   int iterating;
-   char* path;
-   int wd;
-@@ -310,31 +300,6 @@ static struct watcher_root* uv__inotify_watchers(uv_loop_t* loop) {
- }
- 
- 
--unsigned uv__kernel_version(void) {
--  static _Atomic unsigned cached_version;
--  struct utsname u;
--  unsigned version;
--  unsigned major;
--  unsigned minor;
--  unsigned patch;
--
--  version = atomic_load_explicit(&cached_version, memory_order_relaxed);
--  if (version != 0)
--    return version;
--
--  if (-1 == uname(&u))
--    return 0;
--
--  if (3 != sscanf(u.release, "%u.%u.%u", &major, &minor, &patch))
--    return 0;
--
--  version = major * 65536 + minor * 256 + patch;
--  atomic_store_explicit(&cached_version, version, memory_order_relaxed);
--
--  return version;
--}
--
--
- ssize_t
- uv__fs_copy_file_range(int fd_in,
-                        off_t* off_in,
-@@ -420,9 +385,6 @@ int uv__io_uring_register(int fd, unsigned opcode, void* arg, unsigned nargs) {
- 
- 
- static int uv__use_io_uring(void) {
--#if defined(__ANDROID_API__)
--  return 0;  /* Possibly available but blocked by seccomp. */
--#else
-   /* Ternary: unknown=0, yes=1, no=-1 */
-   static _Atomic int use_io_uring;
-   char* val;
-@@ -437,7 +399,6 @@ static int uv__use_io_uring(void) {
-   }
- 
-   return use > 0;
--#endif
- }
- 
- 
-@@ -542,10 +503,6 @@ static void uv__iou_init(int epollfd,
-   iou->sqelen = sqelen;
-   iou->ringfd = ringfd;
-   iou->in_flight = 0;
--  iou->flags = 0;
--
--  if (uv__kernel_version() >= /* 5.15.0 */ 0x050F00)
--    iou->flags |= UV__MKDIRAT_SYMLINKAT_LINKAT;
- 
-   for (i = 0; i <= iou->sqmask; i++)
-     iou->sqarray[i] = i;  /* Slot -> sqe identity mapping. */
-@@ -727,7 +684,7 @@ static struct uv__io_uring_sqe* uv__iou_get_sqe(struct uv__iou* iou,
-   req->work_req.loop = loop;
-   req->work_req.work = NULL;
-   req->work_req.done = NULL;
--  uv__queue_init(&req->work_req.wq);
-+  QUEUE_INIT(&req->work_req.wq);
- 
-   uv__req_register(loop, req);
-   iou->in_flight++;
-@@ -757,17 +714,6 @@ int uv__iou_fs_close(uv_loop_t* loop, uv_fs_t* req) {
-   struct uv__io_uring_sqe* sqe;
-   struct uv__iou* iou;
- 
--  /* Work around a poorly understood bug in older kernels where closing a file
--   * descriptor pointing to /foo/bar results in ETXTBSY errors when trying to
--   * execve("/foo/bar") later on. The bug seems to have been fixed somewhere
--   * between 5.15.85 and 5.15.90. I couldn't pinpoint the responsible commit
--   * but good candidates are the several data race fixes. Interestingly, it
--   * seems to manifest only when running under Docker so the possibility of
--   * a Docker bug can't be completely ruled out either. Yay, computers.
--   */
--  if (uv__kernel_version() < /* 5.15.90 */ 0x050F5A)
--    return 0;
--
-   iou = &uv__get_internal_fields(loop)->iou;
- 
-   sqe = uv__iou_get_sqe(iou, loop, req);
-@@ -808,55 +754,6 @@ int uv__iou_fs_fsync_or_fdatasync(uv_loop_t* loop,
- }
- 
- 
--int uv__iou_fs_link(uv_loop_t* loop, uv_fs_t* req) {
--  struct uv__io_uring_sqe* sqe;
--  struct uv__iou* iou;
--
--  iou = &uv__get_internal_fields(loop)->iou;
--
--  if (!(iou->flags & UV__MKDIRAT_SYMLINKAT_LINKAT))
--    return 0;
--
--  sqe = uv__iou_get_sqe(iou, loop, req);
--  if (sqe == NULL)
--    return 0;
--
--  sqe->addr = (uintptr_t) req->path;
--  sqe->fd = AT_FDCWD;
--  sqe->addr2 = (uintptr_t) req->new_path;
--  sqe->len = AT_FDCWD;
--  sqe->opcode = UV__IORING_OP_LINKAT;
--
--  uv__iou_submit(iou);
--
--  return 1;
--}
--
--
--int uv__iou_fs_mkdir(uv_loop_t* loop, uv_fs_t* req) {
--  struct uv__io_uring_sqe* sqe;
--  struct uv__iou* iou;
--
--  iou = &uv__get_internal_fields(loop)->iou;
--
--  if (!(iou->flags & UV__MKDIRAT_SYMLINKAT_LINKAT))
--    return 0;
--
--  sqe = uv__iou_get_sqe(iou, loop, req);
--  if (sqe == NULL)
--    return 0;
--
--  sqe->addr = (uintptr_t) req->path;
--  sqe->fd = AT_FDCWD;
--  sqe->len = req->mode;
--  sqe->opcode = UV__IORING_OP_MKDIRAT;
--
--  uv__iou_submit(iou);
--
--  return 1;
--}
--
--
- int uv__iou_fs_open(uv_loop_t* loop, uv_fs_t* req) {
-   struct uv__io_uring_sqe* sqe;
-   struct uv__iou* iou;
-@@ -879,86 +776,16 @@ int uv__iou_fs_open(uv_loop_t* loop, uv_fs_t* req) {
- }
- 
- 
--int uv__iou_fs_rename(uv_loop_t* loop, uv_fs_t* req) {
--  struct uv__io_uring_sqe* sqe;
--  struct uv__iou* iou;
--
--  iou = &uv__get_internal_fields(loop)->iou;
--
--  sqe = uv__iou_get_sqe(iou, loop, req);
--  if (sqe == NULL)
--    return 0;
--
--  sqe->addr = (uintptr_t) req->path;
--  sqe->fd = AT_FDCWD;
--  sqe->addr2 = (uintptr_t) req->new_path;
--  sqe->len = AT_FDCWD;
--  sqe->opcode = UV__IORING_OP_RENAMEAT;
--
--  uv__iou_submit(iou);
--
--  return 1;
--}
--
--
--int uv__iou_fs_symlink(uv_loop_t* loop, uv_fs_t* req) {
--  struct uv__io_uring_sqe* sqe;
--  struct uv__iou* iou;
--
--  iou = &uv__get_internal_fields(loop)->iou;
--
--  if (!(iou->flags & UV__MKDIRAT_SYMLINKAT_LINKAT))
--    return 0;
--
--  sqe = uv__iou_get_sqe(iou, loop, req);
--  if (sqe == NULL)
--    return 0;
--
--  sqe->addr = (uintptr_t) req->path;
--  sqe->fd = AT_FDCWD;
--  sqe->addr2 = (uintptr_t) req->new_path;
--  sqe->opcode = UV__IORING_OP_SYMLINKAT;
--
--  uv__iou_submit(iou);
--
--  return 1;
--}
--
--
--int uv__iou_fs_unlink(uv_loop_t* loop, uv_fs_t* req) {
--  struct uv__io_uring_sqe* sqe;
--  struct uv__iou* iou;
--
--  iou = &uv__get_internal_fields(loop)->iou;
--
--  sqe = uv__iou_get_sqe(iou, loop, req);
--  if (sqe == NULL)
--    return 0;
--
--  sqe->addr = (uintptr_t) req->path;
--  sqe->fd = AT_FDCWD;
--  sqe->opcode = UV__IORING_OP_UNLINKAT;
--
--  uv__iou_submit(iou);
--
--  return 1;
--}
--
--
- int uv__iou_fs_read_or_write(uv_loop_t* loop,
-                              uv_fs_t* req,
-                              int is_read) {
-   struct uv__io_uring_sqe* sqe;
-   struct uv__iou* iou;
- 
--  /* If iovcnt is greater than IOV_MAX, cap it to IOV_MAX on reads and fallback
--   * to the threadpool on writes */
--  if (req->nbufs > IOV_MAX) {
--    if (is_read)
--      req->nbufs = IOV_MAX;
--    else
--      return 0;
--  }
-+  /* For the moment, if iovcnt is greater than IOV_MAX, fallback to the
-+   * threadpool. In the future we might take advantage of IOSQE_IO_LINK. */
-+  if (req->nbufs > IOV_MAX)
-+    return 0;
- 
-   iou = &uv__get_internal_fields(loop)->iou;
- 
-@@ -1265,7 +1092,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-   struct uv__iou* ctl;
-   struct uv__iou* iou;
-   int real_timeout;
--  struct uv__queue* q;
-+  QUEUE* q;
-   uv__io_t* w;
-   sigset_t* sigmask;
-   sigset_t sigset;
-@@ -1311,11 +1138,11 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
- 
-   memset(&e, 0, sizeof(e));
- 
--  while (!uv__queue_empty(&loop->watcher_queue)) {
--    q = uv__queue_head(&loop->watcher_queue);
--    w = uv__queue_data(q, uv__io_t, watcher_queue);
--    uv__queue_remove(q);
--    uv__queue_init(q);
-+  while (!QUEUE_EMPTY(&loop->watcher_queue)) {
-+    q = QUEUE_HEAD(&loop->watcher_queue);
-+    w = QUEUE_DATA(q, uv__io_t, watcher_queue);
-+    QUEUE_REMOVE(q);
-+    QUEUE_INIT(q);
- 
-     op = EPOLL_CTL_MOD;
-     if (w->events == 0)
-@@ -2272,8 +2099,8 @@ static int uv__inotify_fork(uv_loop_t* loop, struct watcher_list* root) {
-   struct watcher_list* tmp_watcher_list_iter;
-   struct watcher_list* watcher_list;
-   struct watcher_list tmp_watcher_list;
--  struct uv__queue queue;
--  struct uv__queue* q;
-+  QUEUE queue;
-+  QUEUE* q;
-   uv_fs_event_t* handle;
-   char* tmp_path;
- 
-@@ -2285,41 +2112,41 @@ static int uv__inotify_fork(uv_loop_t* loop, struct watcher_list* root) {
-    */
-   loop->inotify_watchers = root;
- 
--  uv__queue_init(&tmp_watcher_list.watchers);
-+  QUEUE_INIT(&tmp_watcher_list.watchers);
-   /* Note that the queue we use is shared with the start and stop()
--   * functions, making uv__queue_foreach unsafe to use. So we use the
--   * uv__queue_move trick to safely iterate. Also don't free the watcher
-+   * functions, making QUEUE_FOREACH unsafe to use. So we use the
-+   * QUEUE_MOVE trick to safely iterate. Also don't free the watcher
-    * list until we're done iterating. c.f. uv__inotify_read.
-    */
-   RB_FOREACH_SAFE(watcher_list, watcher_root,
-                   uv__inotify_watchers(loop), tmp_watcher_list_iter) {
-     watcher_list->iterating = 1;
--    uv__queue_move(&watcher_list->watchers, &queue);
--    while (!uv__queue_empty(&queue)) {
--      q = uv__queue_head(&queue);
--      handle = uv__queue_data(q, uv_fs_event_t, watchers);
-+    QUEUE_MOVE(&watcher_list->watchers, &queue);
-+    while (!QUEUE_EMPTY(&queue)) {
-+      q = QUEUE_HEAD(&queue);
-+      handle = QUEUE_DATA(q, uv_fs_event_t, watchers);
-       /* It's critical to keep a copy of path here, because it
-        * will be set to NULL by stop() and then deallocated by
-        * maybe_free_watcher_list
-        */
-       tmp_path = uv__strdup(handle->path);
-       assert(tmp_path != NULL);
--      uv__queue_remove(q);
--      uv__queue_insert_tail(&watcher_list->watchers, q);
-+      QUEUE_REMOVE(q);
-+      QUEUE_INSERT_TAIL(&watcher_list->watchers, q);
-       uv_fs_event_stop(handle);
- 
--      uv__queue_insert_tail(&tmp_watcher_list.watchers, &handle->watchers);
-+      QUEUE_INSERT_TAIL(&tmp_watcher_list.watchers, &handle->watchers);
-       handle->path = tmp_path;
-     }
-     watcher_list->iterating = 0;
-     maybe_free_watcher_list(watcher_list, loop);
-   }
- 
--  uv__queue_move(&tmp_watcher_list.watchers, &queue);
--  while (!uv__queue_empty(&queue)) {
--      q = uv__queue_head(&queue);
--      uv__queue_remove(q);
--      handle = uv__queue_data(q, uv_fs_event_t, watchers);
-+  QUEUE_MOVE(&tmp_watcher_list.watchers, &queue);
-+  while (!QUEUE_EMPTY(&queue)) {
-+      q = QUEUE_HEAD(&queue);
-+      QUEUE_REMOVE(q);
-+      handle = QUEUE_DATA(q, uv_fs_event_t, watchers);
-       tmp_path = handle->path;
-       handle->path = NULL;
-       err = uv_fs_event_start(handle, handle->cb, tmp_path, 0);
-@@ -2341,7 +2168,7 @@ static struct watcher_list* find_watcher(uv_loop_t* loop, int wd) {
- 
- static void maybe_free_watcher_list(struct watcher_list* w, uv_loop_t* loop) {
-   /* if the watcher_list->watchers is being iterated over, we can't free it. */
--  if ((!w->iterating) && uv__queue_empty(&w->watchers)) {
-+  if ((!w->iterating) && QUEUE_EMPTY(&w->watchers)) {
-     /* No watchers left for this path. Clean up. */
-     RB_REMOVE(watcher_root, uv__inotify_watchers(loop), w);
-     inotify_rm_watch(loop->inotify_fd, w->wd);
-@@ -2356,8 +2183,8 @@ static void uv__inotify_read(uv_loop_t* loop,
-   const struct inotify_event* e;
-   struct watcher_list* w;
-   uv_fs_event_t* h;
--  struct uv__queue queue;
--  struct uv__queue* q;
-+  QUEUE queue;
-+  QUEUE* q;
-   const char* path;
-   ssize_t size;
-   const char *p;
-@@ -2400,7 +2227,7 @@ static void uv__inotify_read(uv_loop_t* loop,
-        * What can go wrong?
-        * A callback could call uv_fs_event_stop()
-        * and the queue can change under our feet.
--       * So, we use uv__queue_move() trick to safely iterate over the queue.
-+       * So, we use QUEUE_MOVE() trick to safely iterate over the queue.
-        * And we don't free the watcher_list until we're done iterating.
-        *
-        * First,
-@@ -2408,13 +2235,13 @@ static void uv__inotify_read(uv_loop_t* loop,
-        * not to free watcher_list.
-        */
-       w->iterating = 1;
--      uv__queue_move(&w->watchers, &queue);
--      while (!uv__queue_empty(&queue)) {
--        q = uv__queue_head(&queue);
--        h = uv__queue_data(q, uv_fs_event_t, watchers);
-+      QUEUE_MOVE(&w->watchers, &queue);
-+      while (!QUEUE_EMPTY(&queue)) {
-+        q = QUEUE_HEAD(&queue);
-+        h = QUEUE_DATA(q, uv_fs_event_t, watchers);
- 
--        uv__queue_remove(q);
--        uv__queue_insert_tail(&w->watchers, q);
-+        QUEUE_REMOVE(q);
-+        QUEUE_INSERT_TAIL(&w->watchers, q);
- 
-         h->cb(h, path, events, 0);
-       }
-@@ -2476,13 +2303,13 @@ int uv_fs_event_start(uv_fs_event_t* handle,
- 
-   w->wd = wd;
-   w->path = memcpy(w + 1, path, len);
--  uv__queue_init(&w->watchers);
-+  QUEUE_INIT(&w->watchers);
-   w->iterating = 0;
-   RB_INSERT(watcher_root, uv__inotify_watchers(loop), w);
- 
- no_insert:
-   uv__handle_start(handle);
--  uv__queue_insert_tail(&w->watchers, &handle->watchers);
-+  QUEUE_INSERT_TAIL(&w->watchers, &handle->watchers);
-   handle->path = w->path;
-   handle->cb = cb;
-   handle->wd = wd;
-@@ -2503,7 +2330,7 @@ int uv_fs_event_stop(uv_fs_event_t* handle) {
-   handle->wd = -1;
-   handle->path = NULL;
-   uv__handle_stop(handle);
--  uv__queue_remove(&handle->watchers);
-+  QUEUE_REMOVE(&handle->watchers);
- 
-   maybe_free_watcher_list(w, handle->loop);
- 
-diff --git a/deps/uv/src/unix/loop-watcher.c b/deps/uv/src/unix/loop-watcher.c
-index 2db8b515df7..b8c1c2a7102 100644
---- a/deps/uv/src/unix/loop-watcher.c
-+++ b/deps/uv/src/unix/loop-watcher.c
-@@ -32,7 +32,7 @@
-   int uv_##name##_start(uv_##name##_t* handle, uv_##name##_cb cb) {           \
-     if (uv__is_active(handle)) return 0;                                      \
-     if (cb == NULL) return UV_EINVAL;                                         \
--    uv__queue_insert_head(&handle->loop->name##_handles, &handle->queue);     \
-+    QUEUE_INSERT_HEAD(&handle->loop->name##_handles, &handle->queue);         \
-     handle->name##_cb = cb;                                                   \
-     uv__handle_start(handle);                                                 \
-     return 0;                                                                 \
-@@ -40,21 +40,21 @@
-                                                                               \
-   int uv_##name##_stop(uv_##name##_t* handle) {                               \
-     if (!uv__is_active(handle)) return 0;                                     \
--    uv__queue_remove(&handle->queue);                                         \
-+    QUEUE_REMOVE(&handle->queue);                                             \
-     uv__handle_stop(handle);                                                  \
-     return 0;                                                                 \
-   }                                                                           \
-                                                                               \
-   void uv__run_##name(uv_loop_t* loop) {                                      \
-     uv_##name##_t* h;                                                         \
--    struct uv__queue queue;                                                   \
--    struct uv__queue* q;                                                      \
--    uv__queue_move(&loop->name##_handles, &queue);                            \
--    while (!uv__queue_empty(&queue)) {                                        \
--      q = uv__queue_head(&queue);                                             \
--      h = uv__queue_data(q, uv_##name##_t, queue);                            \
--      uv__queue_remove(q);                                                    \
--      uv__queue_insert_tail(&loop->name##_handles, q);                        \
-+    QUEUE queue;                                                              \
-+    QUEUE* q;                                                                 \
-+    QUEUE_MOVE(&loop->name##_handles, &queue);                                \
-+    while (!QUEUE_EMPTY(&queue)) {                                            \
-+      q = QUEUE_HEAD(&queue);                                                 \
-+      h = QUEUE_DATA(q, uv_##name##_t, queue);                                \
-+      QUEUE_REMOVE(q);                                                        \
-+      QUEUE_INSERT_TAIL(&loop->name##_handles, q);                            \
-       h->name##_cb(h);                                                        \
-     }                                                                         \
-   }                                                                           \
-diff --git a/deps/uv/src/unix/loop.c b/deps/uv/src/unix/loop.c
-index a9468e8e19c..90a51b339de 100644
---- a/deps/uv/src/unix/loop.c
-+++ b/deps/uv/src/unix/loop.c
-@@ -50,20 +50,20 @@ int uv_loop_init(uv_loop_t* loop) {
-          sizeof(lfields->loop_metrics.metrics));
- 
-   heap_init((struct heap*) &loop->timer_heap);
--  uv__queue_init(&loop->wq);
--  uv__queue_init(&loop->idle_handles);
--  uv__queue_init(&loop->async_handles);
--  uv__queue_init(&loop->check_handles);
--  uv__queue_init(&loop->prepare_handles);
--  uv__queue_init(&loop->handle_queue);
-+  QUEUE_INIT(&loop->wq);
-+  QUEUE_INIT(&loop->idle_handles);
-+  QUEUE_INIT(&loop->async_handles);
-+  QUEUE_INIT(&loop->check_handles);
-+  QUEUE_INIT(&loop->prepare_handles);
-+  QUEUE_INIT(&loop->handle_queue);
- 
-   loop->active_handles = 0;
-   loop->active_reqs.count = 0;
-   loop->nfds = 0;
-   loop->watchers = NULL;
-   loop->nwatchers = 0;
--  uv__queue_init(&loop->pending_queue);
--  uv__queue_init(&loop->watcher_queue);
-+  QUEUE_INIT(&loop->pending_queue);
-+  QUEUE_INIT(&loop->watcher_queue);
- 
-   loop->closing_handles = NULL;
-   uv__update_time(loop);
-@@ -85,7 +85,7 @@ int uv_loop_init(uv_loop_t* loop) {
-   err = uv__process_init(loop);
-   if (err)
-     goto fail_signal_init;
--  uv__queue_init(&loop->process_handles);
-+  QUEUE_INIT(&loop->process_handles);
- 
-   err = uv_rwlock_init(&loop->cloexec_lock);
-   if (err)
-@@ -152,9 +152,9 @@ int uv_loop_fork(uv_loop_t* loop) {
-     if (w == NULL)
-       continue;
- 
--    if (w->pevents != 0 && uv__queue_empty(&w->watcher_queue)) {
-+    if (w->pevents != 0 && QUEUE_EMPTY(&w->watcher_queue)) {
-       w->events = 0; /* Force re-registration in uv__io_poll. */
--      uv__queue_insert_tail(&loop->watcher_queue, &w->watcher_queue);
-+      QUEUE_INSERT_TAIL(&loop->watcher_queue, &w->watcher_queue);
-     }
-   }
- 
-@@ -180,7 +180,7 @@ void uv__loop_close(uv_loop_t* loop) {
-   }
- 
-   uv_mutex_lock(&loop->wq_mutex);
--  assert(uv__queue_empty(&loop->wq) && "thread pool work queue not empty!");
-+  assert(QUEUE_EMPTY(&loop->wq) && "thread pool work queue not empty!");
-   assert(!uv__has_active_reqs(loop));
-   uv_mutex_unlock(&loop->wq_mutex);
-   uv_mutex_destroy(&loop->wq_mutex);
-@@ -192,8 +192,8 @@ void uv__loop_close(uv_loop_t* loop) {
-   uv_rwlock_destroy(&loop->cloexec_lock);
- 
- #if 0
--  assert(uv__queue_empty(&loop->pending_queue));
--  assert(uv__queue_empty(&loop->watcher_queue));
-+  assert(QUEUE_EMPTY(&loop->pending_queue));
-+  assert(QUEUE_EMPTY(&loop->watcher_queue));
-   assert(loop->nfds == 0);
- #endif
- 
-diff --git a/deps/uv/src/unix/os390-syscalls.c b/deps/uv/src/unix/os390-syscalls.c
-index 7f90c270906..5861aaaa202 100644
---- a/deps/uv/src/unix/os390-syscalls.c
-+++ b/deps/uv/src/unix/os390-syscalls.c
-@@ -27,7 +27,7 @@
- #include <termios.h>
- #include <sys/msg.h>
- 
--static struct uv__queue global_epoll_queue;
-+static QUEUE global_epoll_queue;
- static uv_mutex_t global_epoll_lock;
- static uv_once_t once = UV_ONCE_INIT;
- 
-@@ -178,18 +178,18 @@ static void after_fork(void) {
- 
- 
- static void child_fork(void) {
--  struct uv__queue* q;
-+  QUEUE* q;
-   uv_once_t child_once = UV_ONCE_INIT;
- 
-   /* reset once */
-   memcpy(&once, &child_once, sizeof(child_once));
- 
-   /* reset epoll list */
--  while (!uv__queue_empty(&global_epoll_queue)) {
-+  while (!QUEUE_EMPTY(&global_epoll_queue)) {
-     uv__os390_epoll* lst;
--    q = uv__queue_head(&global_epoll_queue);
--    uv__queue_remove(q);
--    lst = uv__queue_data(q, uv__os390_epoll, member);
-+    q = QUEUE_HEAD(&global_epoll_queue);
-+    QUEUE_REMOVE(q);
-+    lst = QUEUE_DATA(q, uv__os390_epoll, member);
-     uv__free(lst->items);
-     lst->items = NULL;
-     lst->size = 0;
-@@ -201,7 +201,7 @@ static void child_fork(void) {
- 
- 
- static void epoll_init(void) {
--  uv__queue_init(&global_epoll_queue);
-+  QUEUE_INIT(&global_epoll_queue);
-   if (uv_mutex_init(&global_epoll_lock))
-     abort();
- 
-@@ -225,7 +225,7 @@ uv__os390_epoll* epoll_create1(int flags) {
-     lst->items[lst->size - 1].revents = 0;
-     uv_once(&once, epoll_init);
-     uv_mutex_lock(&global_epoll_lock);
--    uv__queue_insert_tail(&global_epoll_queue, &lst->member);
-+    QUEUE_INSERT_TAIL(&global_epoll_queue, &lst->member);
-     uv_mutex_unlock(&global_epoll_lock);
-   }
- 
-@@ -352,14 +352,14 @@ int epoll_wait(uv__os390_epoll* lst, struct epoll_event* events,
- 
- 
- int epoll_file_close(int fd) {
--  struct uv__queue* q;
-+  QUEUE* q;
- 
-   uv_once(&once, epoll_init);
-   uv_mutex_lock(&global_epoll_lock);
--  uv__queue_foreach(q, &global_epoll_queue) {
-+  QUEUE_FOREACH(q, &global_epoll_queue) {
-     uv__os390_epoll* lst;
- 
--    lst = uv__queue_data(q, uv__os390_epoll, member);
-+    lst = QUEUE_DATA(q, uv__os390_epoll, member);
-     if (fd < lst->size && lst->items != NULL && lst->items[fd].fd != -1)
-       lst->items[fd].fd = -1;
-   }
-@@ -371,7 +371,7 @@ int epoll_file_close(int fd) {
- void epoll_queue_close(uv__os390_epoll* lst) {
-   /* Remove epoll instance from global queue */
-   uv_mutex_lock(&global_epoll_lock);
--  uv__queue_remove(&lst->member);
-+  QUEUE_REMOVE(&lst->member);
-   uv_mutex_unlock(&global_epoll_lock);
- 
-   /* Free resources */
-diff --git a/deps/uv/src/unix/os390-syscalls.h b/deps/uv/src/unix/os390-syscalls.h
-index d5f3bcf8b1c..9f504171d85 100644
---- a/deps/uv/src/unix/os390-syscalls.h
-+++ b/deps/uv/src/unix/os390-syscalls.h
-@@ -45,7 +45,7 @@ struct epoll_event {
- };
- 
- typedef struct {
--  struct uv__queue member;
-+  QUEUE member;
-   struct pollfd* items;
-   unsigned long size;
-   int msg_queue;
-diff --git a/deps/uv/src/unix/os390.c b/deps/uv/src/unix/os390.c
-index bbd37692d1d..a87c2d77faf 100644
---- a/deps/uv/src/unix/os390.c
-+++ b/deps/uv/src/unix/os390.c
-@@ -815,7 +815,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-   uv__os390_epoll* ep;
-   int have_signals;
-   int real_timeout;
--  struct uv__queue* q;
-+  QUEUE* q;
-   uv__io_t* w;
-   uint64_t base;
-   int count;
-@@ -827,19 +827,19 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-   int reset_timeout;
- 
-   if (loop->nfds == 0) {
--    assert(uv__queue_empty(&loop->watcher_queue));
-+    assert(QUEUE_EMPTY(&loop->watcher_queue));
-     return;
-   }
- 
-   lfields = uv__get_internal_fields(loop);
- 
--  while (!uv__queue_empty(&loop->watcher_queue)) {
-+  while (!QUEUE_EMPTY(&loop->watcher_queue)) {
-     uv_stream_t* stream;
- 
--    q = uv__queue_head(&loop->watcher_queue);
--    uv__queue_remove(q);
--    uv__queue_init(q);
--    w = uv__queue_data(q, uv__io_t, watcher_queue);
-+    q = QUEUE_HEAD(&loop->watcher_queue);
-+    QUEUE_REMOVE(q);
-+    QUEUE_INIT(q);
-+    w = QUEUE_DATA(q, uv__io_t, watcher_queue);
- 
-     assert(w->pevents != 0);
-     assert(w->fd >= 0);
-diff --git a/deps/uv/src/unix/pipe.c b/deps/uv/src/unix/pipe.c
-index d332f351830..ce91ac49b30 100644
---- a/deps/uv/src/unix/pipe.c
-+++ b/deps/uv/src/unix/pipe.c
-@@ -297,7 +297,7 @@ out:
-   uv__req_init(handle->loop, req, UV_CONNECT);
-   req->handle = (uv_stream_t*)handle;
-   req->cb = cb;
--  uv__queue_init(&req->queue);
-+  QUEUE_INIT(&req->queue);
- 
-   /* Force callback to run on next tick in case of error. */
-   if (err)
-diff --git a/deps/uv/src/unix/posix-poll.c b/deps/uv/src/unix/posix-poll.c
-index 2e016c2fbae..7e7de86845d 100644
---- a/deps/uv/src/unix/posix-poll.c
-+++ b/deps/uv/src/unix/posix-poll.c
-@@ -137,7 +137,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-   sigset_t set;
-   uint64_t time_base;
-   uint64_t time_diff;
--  struct uv__queue* q;
-+  QUEUE* q;
-   uv__io_t* w;
-   size_t i;
-   unsigned int nevents;
-@@ -149,19 +149,19 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-   int reset_timeout;
- 
-   if (loop->nfds == 0) {
--    assert(uv__queue_empty(&loop->watcher_queue));
-+    assert(QUEUE_EMPTY(&loop->watcher_queue));
-     return;
-   }
- 
-   lfields = uv__get_internal_fields(loop);
- 
-   /* Take queued watchers and add their fds to our poll fds array.  */
--  while (!uv__queue_empty(&loop->watcher_queue)) {
--    q = uv__queue_head(&loop->watcher_queue);
--    uv__queue_remove(q);
--    uv__queue_init(q);
-+  while (!QUEUE_EMPTY(&loop->watcher_queue)) {
-+    q = QUEUE_HEAD(&loop->watcher_queue);
-+    QUEUE_REMOVE(q);
-+    QUEUE_INIT(q);
- 
--    w = uv__queue_data(q, uv__io_t, watcher_queue);
-+    w = QUEUE_DATA(q, uv__io_t, watcher_queue);
-     assert(w->pevents != 0);
-     assert(w->fd >= 0);
-     assert(w->fd < (int) loop->nwatchers);
-diff --git a/deps/uv/src/unix/process.c b/deps/uv/src/unix/process.c
-index dd58c18d9b9..bbf367b57d3 100644
---- a/deps/uv/src/unix/process.c
-+++ b/deps/uv/src/unix/process.c
-@@ -108,17 +108,17 @@ void uv__wait_children(uv_loop_t* loop) {
-   int status;
-   int options;
-   pid_t pid;
--  struct uv__queue pending;
--  struct uv__queue* q;
--  struct uv__queue* h;
-+  QUEUE pending;
-+  QUEUE* q;
-+  QUEUE* h;
- 
--  uv__queue_init(&pending);
-+  QUEUE_INIT(&pending);
- 
-   h = &loop->process_handles;
--  q = uv__queue_head(h);
-+  q = QUEUE_HEAD(h);
-   while (q != h) {
--    process = uv__queue_data(q, uv_process_t, queue);
--    q = uv__queue_next(q);
-+    process = QUEUE_DATA(q, uv_process_t, queue);
-+    q = QUEUE_NEXT(q);
- 
- #ifndef UV_USE_SIGCHLD
-     if ((process->flags & UV_HANDLE_REAP) == 0)
-@@ -149,18 +149,18 @@ void uv__wait_children(uv_loop_t* loop) {
- 
-     assert(pid == process->pid);
-     process->status = status;
--    uv__queue_remove(&process->queue);
--    uv__queue_insert_tail(&pending, &process->queue);
-+    QUEUE_REMOVE(&process->queue);
-+    QUEUE_INSERT_TAIL(&pending, &process->queue);
-   }
- 
-   h = &pending;
--  q = uv__queue_head(h);
-+  q = QUEUE_HEAD(h);
-   while (q != h) {
--    process = uv__queue_data(q, uv_process_t, queue);
--    q = uv__queue_next(q);
-+    process = QUEUE_DATA(q, uv_process_t, queue);
-+    q = QUEUE_NEXT(q);
- 
--    uv__queue_remove(&process->queue);
--    uv__queue_init(&process->queue);
-+    QUEUE_REMOVE(&process->queue);
-+    QUEUE_INIT(&process->queue);
-     uv__handle_stop(process);
- 
-     if (process->exit_cb == NULL)
-@@ -176,7 +176,7 @@ void uv__wait_children(uv_loop_t* loop) {
- 
-     process->exit_cb(process, exit_status, term_signal);
-   }
--  assert(uv__queue_empty(&pending));
-+  assert(QUEUE_EMPTY(&pending));
- }
- 
- /*
-@@ -978,7 +978,7 @@ int uv_spawn(uv_loop_t* loop,
-                               UV_PROCESS_WINDOWS_VERBATIM_ARGUMENTS)));
- 
-   uv__handle_init(loop, (uv_handle_t*)process, UV_PROCESS);
--  uv__queue_init(&process->queue);
-+  QUEUE_INIT(&process->queue);
-   process->status = 0;
- 
-   stdio_count = options->stdio_count;
-@@ -1041,7 +1041,7 @@ int uv_spawn(uv_loop_t* loop,
- 
-     process->pid = pid;
-     process->exit_cb = options->exit_cb;
--    uv__queue_insert_tail(&loop->process_handles, &process->queue);
-+    QUEUE_INSERT_TAIL(&loop->process_handles, &process->queue);
-     uv__handle_start(process);
-   }
- 
-@@ -1103,10 +1103,10 @@ int uv_kill(int pid, int signum) {
- 
- 
- void uv__process_close(uv_process_t* handle) {
--  uv__queue_remove(&handle->queue);
-+  QUEUE_REMOVE(&handle->queue);
-   uv__handle_stop(handle);
- #ifdef UV_USE_SIGCHLD
--  if (uv__queue_empty(&handle->loop->process_handles))
-+  if (QUEUE_EMPTY(&handle->loop->process_handles))
-     uv_signal_stop(&handle->loop->child_watcher);
- #endif
- }
-diff --git a/deps/uv/src/unix/signal.c b/deps/uv/src/unix/signal.c
-index 63aba5a60e0..bb70523f561 100644
---- a/deps/uv/src/unix/signal.c
-+++ b/deps/uv/src/unix/signal.c
-@@ -291,16 +291,16 @@ int uv__signal_loop_fork(uv_loop_t* loop) {
- 
- 
- void uv__signal_loop_cleanup(uv_loop_t* loop) {
--  struct uv__queue* q;
-+  QUEUE* q;
- 
-   /* Stop all the signal watchers that are still attached to this loop. This
-    * ensures that the (shared) signal tree doesn't contain any invalid entries
-    * entries, and that signal handlers are removed when appropriate.
--   * It's safe to use uv__queue_foreach here because the handles and the handle
-+   * It's safe to use QUEUE_FOREACH here because the handles and the handle
-    * queue are not modified by uv__signal_stop().
-    */
--  uv__queue_foreach(q, &loop->handle_queue) {
--    uv_handle_t* handle = uv__queue_data(q, uv_handle_t, handle_queue);
-+  QUEUE_FOREACH(q, &loop->handle_queue) {
-+    uv_handle_t* handle = QUEUE_DATA(q, uv_handle_t, handle_queue);
- 
-     if (handle->type == UV_SIGNAL)
-       uv__signal_stop((uv_signal_t*) handle);
-diff --git a/deps/uv/src/unix/stream.c b/deps/uv/src/unix/stream.c
-index 28c4d5463c4..03f92b5045a 100644
---- a/deps/uv/src/unix/stream.c
-+++ b/deps/uv/src/unix/stream.c
-@@ -94,8 +94,8 @@ void uv__stream_init(uv_loop_t* loop,
-   stream->accepted_fd = -1;
-   stream->queued_fds = NULL;
-   stream->delayed_error = 0;
--  uv__queue_init(&stream->write_queue);
--  uv__queue_init(&stream->write_completed_queue);
-+  QUEUE_INIT(&stream->write_queue);
-+  QUEUE_INIT(&stream->write_completed_queue);
-   stream->write_queue_size = 0;
- 
-   if (loop->emfile_fd == -1) {
-@@ -439,15 +439,15 @@ int uv__stream_open(uv_stream_t* stream, int fd, int flags) {
- 
- void uv__stream_flush_write_queue(uv_stream_t* stream, int error) {
-   uv_write_t* req;
--  struct uv__queue* q;
--  while (!uv__queue_empty(&stream->write_queue)) {
--    q = uv__queue_head(&stream->write_queue);
--    uv__queue_remove(q);
-+  QUEUE* q;
-+  while (!QUEUE_EMPTY(&stream->write_queue)) {
-+    q = QUEUE_HEAD(&stream->write_queue);
-+    QUEUE_REMOVE(q);
- 
--    req = uv__queue_data(q, uv_write_t, queue);
-+    req = QUEUE_DATA(q, uv_write_t, queue);
-     req->error = error;
- 
--    uv__queue_insert_tail(&stream->write_completed_queue, &req->queue);
-+    QUEUE_INSERT_TAIL(&stream->write_completed_queue, &req->queue);
-   }
- }
- 
-@@ -627,7 +627,7 @@ static void uv__drain(uv_stream_t* stream) {
-   uv_shutdown_t* req;
-   int err;
- 
--  assert(uv__queue_empty(&stream->write_queue));
-+  assert(QUEUE_EMPTY(&stream->write_queue));
-   if (!(stream->flags & UV_HANDLE_CLOSING)) {
-     uv__io_stop(stream->loop, &stream->io_watcher, POLLOUT);
-     uv__stream_osx_interrupt_select(stream);
-@@ -714,7 +714,7 @@ static void uv__write_req_finish(uv_write_t* req) {
-   uv_stream_t* stream = req->handle;
- 
-   /* Pop the req off tcp->write_queue. */
--  uv__queue_remove(&req->queue);
-+  QUEUE_REMOVE(&req->queue);
- 
-   /* Only free when there was no error. On error, we touch up write_queue_size
-    * right before making the callback. The reason we don't do that right away
-@@ -731,7 +731,7 @@ static void uv__write_req_finish(uv_write_t* req) {
-   /* Add it to the write_completed_queue where it will have its
-    * callback called in the near future.
-    */
--  uv__queue_insert_tail(&stream->write_completed_queue, &req->queue);
-+  QUEUE_INSERT_TAIL(&stream->write_completed_queue, &req->queue);
-   uv__io_feed(stream->loop, &stream->io_watcher);
- }
- 
-@@ -837,7 +837,7 @@ static int uv__try_write(uv_stream_t* stream,
- }
- 
- static void uv__write(uv_stream_t* stream) {
--  struct uv__queue* q;
-+  QUEUE* q;
-   uv_write_t* req;
-   ssize_t n;
-   int count;
-@@ -851,11 +851,11 @@ static void uv__write(uv_stream_t* stream) {
-   count = 32;
- 
-   for (;;) {
--    if (uv__queue_empty(&stream->write_queue))
-+    if (QUEUE_EMPTY(&stream->write_queue))
-       return;
- 
--    q = uv__queue_head(&stream->write_queue);
--    req = uv__queue_data(q, uv_write_t, queue);
-+    q = QUEUE_HEAD(&stream->write_queue);
-+    req = QUEUE_DATA(q, uv_write_t, queue);
-     assert(req->handle == stream);
- 
-     n = uv__try_write(stream,
-@@ -899,19 +899,19 @@ error:
- 
- static void uv__write_callbacks(uv_stream_t* stream) {
-   uv_write_t* req;
--  struct uv__queue* q;
--  struct uv__queue pq;
-+  QUEUE* q;
-+  QUEUE pq;
- 
--  if (uv__queue_empty(&stream->write_completed_queue))
-+  if (QUEUE_EMPTY(&stream->write_completed_queue))
-     return;
- 
--  uv__queue_move(&stream->write_completed_queue, &pq);
-+  QUEUE_MOVE(&stream->write_completed_queue, &pq);
- 
--  while (!uv__queue_empty(&pq)) {
-+  while (!QUEUE_EMPTY(&pq)) {
-     /* Pop a req off write_completed_queue. */
--    q = uv__queue_head(&pq);
--    req = uv__queue_data(q, uv_write_t, queue);
--    uv__queue_remove(q);
-+    q = QUEUE_HEAD(&pq);
-+    req = QUEUE_DATA(q, uv_write_t, queue);
-+    QUEUE_REMOVE(q);
-     uv__req_unregister(stream->loop, req);
- 
-     if (req->bufs != NULL) {
-@@ -1174,7 +1174,7 @@ int uv_shutdown(uv_shutdown_t* req, uv_stream_t* stream, uv_shutdown_cb cb) {
-   stream->shutdown_req = req;
-   stream->flags &= ~UV_HANDLE_WRITABLE;
- 
--  if (uv__queue_empty(&stream->write_queue))
-+  if (QUEUE_EMPTY(&stream->write_queue))
-     uv__io_feed(stream->loop, &stream->io_watcher);
- 
-   return 0;
-@@ -1227,7 +1227,7 @@ static void uv__stream_io(uv_loop_t* loop, uv__io_t* w, unsigned int events) {
-     uv__write_callbacks(stream);
- 
-     /* Write queue drained. */
--    if (uv__queue_empty(&stream->write_queue))
-+    if (QUEUE_EMPTY(&stream->write_queue))
-       uv__drain(stream);
-   }
- }
-@@ -1270,7 +1270,7 @@ static void uv__stream_connect(uv_stream_t* stream) {
-   stream->connect_req = NULL;
-   uv__req_unregister(stream->loop, req);
- 
--  if (error < 0 || uv__queue_empty(&stream->write_queue)) {
-+  if (error < 0 || QUEUE_EMPTY(&stream->write_queue)) {
-     uv__io_stop(stream->loop, &stream->io_watcher, POLLOUT);
-   }
- 
-@@ -1352,7 +1352,7 @@ int uv_write2(uv_write_t* req,
-   req->handle = stream;
-   req->error = 0;
-   req->send_handle = send_handle;
--  uv__queue_init(&req->queue);
-+  QUEUE_INIT(&req->queue);
- 
-   req->bufs = req->bufsml;
-   if (nbufs > ARRAY_SIZE(req->bufsml))
-@@ -1367,7 +1367,7 @@ int uv_write2(uv_write_t* req,
-   stream->write_queue_size += uv__count_bufs(bufs, nbufs);
- 
-   /* Append the request to write_queue. */
--  uv__queue_insert_tail(&stream->write_queue, &req->queue);
-+  QUEUE_INSERT_TAIL(&stream->write_queue, &req->queue);
- 
-   /* If the queue was empty when this function began, we should attempt to
-    * do the write immediately. Otherwise start the write_watcher and wait
-diff --git a/deps/uv/src/unix/sunos.c b/deps/uv/src/unix/sunos.c
-index 2d6bae79604..75b6fbad493 100644
---- a/deps/uv/src/unix/sunos.c
-+++ b/deps/uv/src/unix/sunos.c
-@@ -148,7 +148,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-   struct port_event events[1024];
-   struct port_event* pe;
-   struct timespec spec;
--  struct uv__queue* q;
-+  QUEUE* q;
-   uv__io_t* w;
-   sigset_t* pset;
-   sigset_t set;
-@@ -166,16 +166,16 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-   int reset_timeout;
- 
-   if (loop->nfds == 0) {
--    assert(uv__queue_empty(&loop->watcher_queue));
-+    assert(QUEUE_EMPTY(&loop->watcher_queue));
-     return;
-   }
- 
--  while (!uv__queue_empty(&loop->watcher_queue)) {
--    q = uv__queue_head(&loop->watcher_queue);
--    uv__queue_remove(q);
--    uv__queue_init(q);
-+  while (!QUEUE_EMPTY(&loop->watcher_queue)) {
-+    q = QUEUE_HEAD(&loop->watcher_queue);
-+    QUEUE_REMOVE(q);
-+    QUEUE_INIT(q);
- 
--    w = uv__queue_data(q, uv__io_t, watcher_queue);
-+    w = QUEUE_DATA(q, uv__io_t, watcher_queue);
-     assert(w->pevents != 0);
- 
-     if (port_associate(loop->backend_fd,
-@@ -316,8 +316,8 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-         continue;  /* Disabled by callback. */
- 
-       /* Events Ports operates in oneshot mode, rearm timer on next run. */
--      if (w->pevents != 0 && uv__queue_empty(&w->watcher_queue))
--        uv__queue_insert_tail(&loop->watcher_queue, &w->watcher_queue);
-+      if (w->pevents != 0 && QUEUE_EMPTY(&w->watcher_queue))
-+        QUEUE_INSERT_TAIL(&loop->watcher_queue, &w->watcher_queue);
-     }
- 
-     uv__metrics_inc_events(loop, nevents);
-diff --git a/deps/uv/src/unix/tcp.c b/deps/uv/src/unix/tcp.c
-index d6c848f4610..ab4e06c2f67 100644
---- a/deps/uv/src/unix/tcp.c
-+++ b/deps/uv/src/unix/tcp.c
-@@ -124,7 +124,7 @@ int uv_tcp_init_ex(uv_loop_t* loop, uv_tcp_t* tcp, unsigned int flags) {
-   if (domain != AF_UNSPEC) {
-     err = new_socket(tcp, domain, 0);
-     if (err) {
--      uv__queue_remove(&tcp->handle_queue);
-+      QUEUE_REMOVE(&tcp->handle_queue);
-       if (tcp->io_watcher.fd != -1)
-         uv__close(tcp->io_watcher.fd);
-       tcp->io_watcher.fd = -1;
-@@ -252,7 +252,7 @@ out:
-   uv__req_init(handle->loop, req, UV_CONNECT);
-   req->cb = cb;
-   req->handle = (uv_stream_t*) handle;
--  uv__queue_init(&req->queue);
-+  QUEUE_INIT(&req->queue);
-   handle->connect_req = req;
- 
-   uv__io_start(handle->loop, &handle->io_watcher, POLLOUT);
-diff --git a/deps/uv/src/unix/tty.c b/deps/uv/src/unix/tty.c
-index d099bdb3b67..7a5390c1a8b 100644
---- a/deps/uv/src/unix/tty.c
-+++ b/deps/uv/src/unix/tty.c
-@@ -222,7 +222,7 @@ skip:
-     int rc = r;
-     if (newfd != -1)
-       uv__close(newfd);
--    uv__queue_remove(&tty->handle_queue);
-+    QUEUE_REMOVE(&tty->handle_queue);
-     do
-       r = fcntl(fd, F_SETFL, saved_flags);
-     while (r == -1 && errno == EINTR);
-diff --git a/deps/uv/src/unix/udp.c b/deps/uv/src/unix/udp.c
-index c2814512a5f..f556808fbae 100644
---- a/deps/uv/src/unix/udp.c
-+++ b/deps/uv/src/unix/udp.c
-@@ -62,18 +62,18 @@ void uv__udp_close(uv_udp_t* handle) {
- 
- void uv__udp_finish_close(uv_udp_t* handle) {
-   uv_udp_send_t* req;
--  struct uv__queue* q;
-+  QUEUE* q;
- 
-   assert(!uv__io_active(&handle->io_watcher, POLLIN | POLLOUT));
-   assert(handle->io_watcher.fd == -1);
- 
--  while (!uv__queue_empty(&handle->write_queue)) {
--    q = uv__queue_head(&handle->write_queue);
--    uv__queue_remove(q);
-+  while (!QUEUE_EMPTY(&handle->write_queue)) {
-+    q = QUEUE_HEAD(&handle->write_queue);
-+    QUEUE_REMOVE(q);
- 
--    req = uv__queue_data(q, uv_udp_send_t, queue);
-+    req = QUEUE_DATA(q, uv_udp_send_t, queue);
-     req->status = UV_ECANCELED;
--    uv__queue_insert_tail(&handle->write_completed_queue, &req->queue);
-+    QUEUE_INSERT_TAIL(&handle->write_completed_queue, &req->queue);
-   }
- 
-   uv__udp_run_completed(handle);
-@@ -90,16 +90,16 @@ void uv__udp_finish_close(uv_udp_t* handle) {
- 
- static void uv__udp_run_completed(uv_udp_t* handle) {
-   uv_udp_send_t* req;
--  struct uv__queue* q;
-+  QUEUE* q;
- 
-   assert(!(handle->flags & UV_HANDLE_UDP_PROCESSING));
-   handle->flags |= UV_HANDLE_UDP_PROCESSING;
- 
--  while (!uv__queue_empty(&handle->write_completed_queue)) {
--    q = uv__queue_head(&handle->write_completed_queue);
--    uv__queue_remove(q);
-+  while (!QUEUE_EMPTY(&handle->write_completed_queue)) {
-+    q = QUEUE_HEAD(&handle->write_completed_queue);
-+    QUEUE_REMOVE(q);
- 
--    req = uv__queue_data(q, uv_udp_send_t, queue);
-+    req = QUEUE_DATA(q, uv_udp_send_t, queue);
-     uv__req_unregister(handle->loop, req);
- 
-     handle->send_queue_size -= uv__count_bufs(req->bufs, req->nbufs);
-@@ -121,7 +121,7 @@ static void uv__udp_run_completed(uv_udp_t* handle) {
-       req->send_cb(req, req->status);
-   }
- 
--  if (uv__queue_empty(&handle->write_queue)) {
-+  if (QUEUE_EMPTY(&handle->write_queue)) {
-     /* Pending queue and completion queue empty, stop watcher. */
-     uv__io_stop(handle->loop, &handle->io_watcher, POLLOUT);
-     if (!uv__io_active(&handle->io_watcher, POLLIN))
-@@ -280,20 +280,20 @@ static void uv__udp_sendmsg(uv_udp_t* handle) {
-   uv_udp_send_t* req;
-   struct mmsghdr h[20];
-   struct mmsghdr* p;
--  struct uv__queue* q;
-+  QUEUE* q;
-   ssize_t npkts;
-   size_t pkts;
-   size_t i;
- 
--  if (uv__queue_empty(&handle->write_queue))
-+  if (QUEUE_EMPTY(&handle->write_queue))
-     return;
- 
- write_queue_drain:
--  for (pkts = 0, q = uv__queue_head(&handle->write_queue);
-+  for (pkts = 0, q = QUEUE_HEAD(&handle->write_queue);
-        pkts < ARRAY_SIZE(h) && q != &handle->write_queue;
--       ++pkts, q = uv__queue_head(q)) {
-+       ++pkts, q = QUEUE_HEAD(q)) {
-     assert(q != NULL);
--    req = uv__queue_data(q, uv_udp_send_t, queue);
-+    req = QUEUE_DATA(q, uv_udp_send_t, queue);
-     assert(req != NULL);
- 
-     p = &h[pkts];
-@@ -325,16 +325,16 @@ write_queue_drain:
-   if (npkts < 1) {
-     if (errno == EAGAIN || errno == EWOULDBLOCK || errno == ENOBUFS)
-       return;
--    for (i = 0, q = uv__queue_head(&handle->write_queue);
-+    for (i = 0, q = QUEUE_HEAD(&handle->write_queue);
-          i < pkts && q != &handle->write_queue;
--         ++i, q = uv__queue_head(&handle->write_queue)) {
-+         ++i, q = QUEUE_HEAD(&handle->write_queue)) {
-       assert(q != NULL);
--      req = uv__queue_data(q, uv_udp_send_t, queue);
-+      req = QUEUE_DATA(q, uv_udp_send_t, queue);
-       assert(req != NULL);
- 
-       req->status = UV__ERR(errno);
--      uv__queue_remove(&req->queue);
--      uv__queue_insert_tail(&handle->write_completed_queue, &req->queue);
-+      QUEUE_REMOVE(&req->queue);
-+      QUEUE_INSERT_TAIL(&handle->write_completed_queue, &req->queue);
-     }
-     uv__io_feed(handle->loop, &handle->io_watcher);
-     return;
-@@ -343,11 +343,11 @@ write_queue_drain:
-   /* Safety: npkts known to be >0 below. Hence cast from ssize_t
-    * to size_t safe.
-    */
--  for (i = 0, q = uv__queue_head(&handle->write_queue);
-+  for (i = 0, q = QUEUE_HEAD(&handle->write_queue);
-        i < (size_t)npkts && q != &handle->write_queue;
--       ++i, q = uv__queue_head(&handle->write_queue)) {
-+       ++i, q = QUEUE_HEAD(&handle->write_queue)) {
-     assert(q != NULL);
--    req = uv__queue_data(q, uv_udp_send_t, queue);
-+    req = QUEUE_DATA(q, uv_udp_send_t, queue);
-     assert(req != NULL);
- 
-     req->status = req->bufs[0].len;
-@@ -357,25 +357,25 @@ write_queue_drain:
-      * why we don't handle partial writes. Just pop the request
-      * off the write queue and onto the completed queue, done.
-      */
--    uv__queue_remove(&req->queue);
--    uv__queue_insert_tail(&handle->write_completed_queue, &req->queue);
-+    QUEUE_REMOVE(&req->queue);
-+    QUEUE_INSERT_TAIL(&handle->write_completed_queue, &req->queue);
-   }
- 
-   /* couldn't batch everything, continue sending (jump to avoid stack growth) */
--  if (!uv__queue_empty(&handle->write_queue))
-+  if (!QUEUE_EMPTY(&handle->write_queue))
-     goto write_queue_drain;
-   uv__io_feed(handle->loop, &handle->io_watcher);
- #else  /* __linux__ || ____FreeBSD__ */
-   uv_udp_send_t* req;
-   struct msghdr h;
--  struct uv__queue* q;
-+  QUEUE* q;
-   ssize_t size;
- 
--  while (!uv__queue_empty(&handle->write_queue)) {
--    q = uv__queue_head(&handle->write_queue);
-+  while (!QUEUE_EMPTY(&handle->write_queue)) {
-+    q = QUEUE_HEAD(&handle->write_queue);
-     assert(q != NULL);
- 
--    req = uv__queue_data(q, uv_udp_send_t, queue);
-+    req = QUEUE_DATA(q, uv_udp_send_t, queue);
-     assert(req != NULL);
- 
-     memset(&h, 0, sizeof h);
-@@ -414,8 +414,8 @@ write_queue_drain:
-      * why we don't handle partial writes. Just pop the request
-      * off the write queue and onto the completed queue, done.
-      */
--    uv__queue_remove(&req->queue);
--    uv__queue_insert_tail(&handle->write_completed_queue, &req->queue);
-+    QUEUE_REMOVE(&req->queue);
-+    QUEUE_INSERT_TAIL(&handle->write_completed_queue, &req->queue);
-     uv__io_feed(handle->loop, &handle->io_watcher);
-   }
- #endif  /* __linux__ || ____FreeBSD__ */
-@@ -729,7 +729,7 @@ int uv__udp_send(uv_udp_send_t* req,
-   memcpy(req->bufs, bufs, nbufs * sizeof(bufs[0]));
-   handle->send_queue_size += uv__count_bufs(req->bufs, req->nbufs);
-   handle->send_queue_count++;
--  uv__queue_insert_tail(&handle->write_queue, &req->queue);
-+  QUEUE_INSERT_TAIL(&handle->write_queue, &req->queue);
-   uv__handle_start(handle);
- 
-   if (empty_queue && !(handle->flags & UV_HANDLE_UDP_PROCESSING)) {
-@@ -739,7 +739,7 @@ int uv__udp_send(uv_udp_send_t* req,
-      * away. In such cases the `io_watcher` has to be queued for asynchronous
-      * write.
-      */
--    if (!uv__queue_empty(&handle->write_queue))
-+    if (!QUEUE_EMPTY(&handle->write_queue))
-       uv__io_start(handle->loop, &handle->io_watcher, POLLOUT);
-   } else {
-     uv__io_start(handle->loop, &handle->io_watcher, POLLOUT);
-@@ -1007,8 +1007,8 @@ int uv__udp_init_ex(uv_loop_t* loop,
-   handle->send_queue_size = 0;
-   handle->send_queue_count = 0;
-   uv__io_init(&handle->io_watcher, uv__udp_io, fd);
--  uv__queue_init(&handle->write_queue);
--  uv__queue_init(&handle->write_completed_queue);
-+  QUEUE_INIT(&handle->write_queue);
-+  QUEUE_INIT(&handle->write_completed_queue);
- 
-   return 0;
- }
-diff --git a/deps/uv/src/uv-common.c b/deps/uv/src/uv-common.c
-index 916f3f4e006..cec771fab21 100644
---- a/deps/uv/src/uv-common.c
-+++ b/deps/uv/src/uv-common.c
-@@ -533,17 +533,17 @@ int uv_udp_recv_stop(uv_udp_t* handle) {
- 
- 
- void uv_walk(uv_loop_t* loop, uv_walk_cb walk_cb, void* arg) {
--  struct uv__queue queue;
--  struct uv__queue* q;
-+  QUEUE queue;
-+  QUEUE* q;
-   uv_handle_t* h;
- 
--  uv__queue_move(&loop->handle_queue, &queue);
--  while (!uv__queue_empty(&queue)) {
--    q = uv__queue_head(&queue);
--    h = uv__queue_data(q, uv_handle_t, handle_queue);
-+  QUEUE_MOVE(&loop->handle_queue, &queue);
-+  while (!QUEUE_EMPTY(&queue)) {
-+    q = QUEUE_HEAD(&queue);
-+    h = QUEUE_DATA(q, uv_handle_t, handle_queue);
- 
--    uv__queue_remove(q);
--    uv__queue_insert_tail(&loop->handle_queue, q);
-+    QUEUE_REMOVE(q);
-+    QUEUE_INSERT_TAIL(&loop->handle_queue, q);
- 
-     if (h->flags & UV_HANDLE_INTERNAL) continue;
-     walk_cb(h, arg);
-@@ -553,14 +553,14 @@ void uv_walk(uv_loop_t* loop, uv_walk_cb walk_cb, void* arg) {
- 
- static void uv__print_handles(uv_loop_t* loop, int only_active, FILE* stream) {
-   const char* type;
--  struct uv__queue* q;
-+  QUEUE* q;
-   uv_handle_t* h;
- 
-   if (loop == NULL)
-     loop = uv_default_loop();
- 
--  uv__queue_foreach(q, &loop->handle_queue) {
--    h = uv__queue_data(q, uv_handle_t, handle_queue);
-+  QUEUE_FOREACH(q, &loop->handle_queue) {
-+    h = QUEUE_DATA(q, uv_handle_t, handle_queue);
- 
-     if (only_active && !uv__is_active(h))
-       continue;
-@@ -846,7 +846,7 @@ uv_loop_t* uv_loop_new(void) {
- 
- 
- int uv_loop_close(uv_loop_t* loop) {
--  struct uv__queue* q;
-+  QUEUE* q;
-   uv_handle_t* h;
- #ifndef NDEBUG
-   void* saved_data;
-@@ -855,8 +855,8 @@ int uv_loop_close(uv_loop_t* loop) {
-   if (uv__has_active_reqs(loop))
-     return UV_EBUSY;
- 
--  uv__queue_foreach(q, &loop->handle_queue) {
--    h = uv__queue_data(q, uv_handle_t, handle_queue);
-+  QUEUE_FOREACH(q, &loop->handle_queue) {
-+    h = QUEUE_DATA(q, uv_handle_t, handle_queue);
-     if (!(h->flags & UV_HANDLE_INTERNAL))
-       return UV_EBUSY;
-   }
-diff --git a/deps/uv/src/uv-common.h b/deps/uv/src/uv-common.h
-index cd57e5a3515..decde5362c8 100644
---- a/deps/uv/src/uv-common.h
-+++ b/deps/uv/src/uv-common.h
-@@ -323,7 +323,7 @@ void uv__threadpool_cleanup(void);
-     (h)->loop = (loop_);                                                      \
-     (h)->type = (type_);                                                      \
-     (h)->flags = UV_HANDLE_REF;  /* Ref the loop when active. */              \
--    uv__queue_insert_tail(&(loop_)->handle_queue, &(h)->handle_queue);        \
-+    QUEUE_INSERT_TAIL(&(loop_)->handle_queue, &(h)->handle_queue);            \
-     uv__handle_platform_init(h);                                              \
-   }                                                                           \
-   while (0)
-@@ -415,7 +415,6 @@ struct uv__iou {
-   size_t sqelen;
-   int ringfd;
-   uint32_t in_flight;
--  uint32_t flags;
- };
- #endif  /* __linux__ */
- 
-diff --git a/deps/uv/src/win/core.c b/deps/uv/src/win/core.c
-index e9885a0f1ff..9a3be58849a 100644
---- a/deps/uv/src/win/core.c
-+++ b/deps/uv/src/win/core.c
-@@ -255,8 +255,8 @@ int uv_loop_init(uv_loop_t* loop) {
-   loop->time = 0;
-   uv_update_time(loop);
- 
--  uv__queue_init(&loop->wq);
--  uv__queue_init(&loop->handle_queue);
-+  QUEUE_INIT(&loop->wq);
-+  QUEUE_INIT(&loop->handle_queue);
-   loop->active_reqs.count = 0;
-   loop->active_handles = 0;
- 
-@@ -358,7 +358,7 @@ void uv__loop_close(uv_loop_t* loop) {
-   }
- 
-   uv_mutex_lock(&loop->wq_mutex);
--  assert(uv__queue_empty(&loop->wq) && "thread pool work queue not empty!");
-+  assert(QUEUE_EMPTY(&loop->wq) && "thread pool work queue not empty!");
-   assert(!uv__has_active_reqs(loop));
-   uv_mutex_unlock(&loop->wq_mutex);
-   uv_mutex_destroy(&loop->wq_mutex);
-diff --git a/deps/uv/src/win/fs.c b/deps/uv/src/win/fs.c
-index fc209c54f47..deb9438d689 100644
---- a/deps/uv/src/win/fs.c
-+++ b/deps/uv/src/win/fs.c
-@@ -144,97 +144,26 @@ void uv__fs_init(void) {
- }
- 
- 
--static int32_t fs__decode_wtf8_char(const char** input) {
--  uint32_t code_point;
--  uint8_t b1;
--  uint8_t b2;
--  uint8_t b3;
--  uint8_t b4;
--
--  b1 = **input;
--  if (b1 <= 0x7F)
--    return b1; /* ASCII code point */
--  if (b1 < 0xC2)
--    return -1; /* invalid: continuation byte */
--  code_point = b1;
--
--  b2 = *++*input;
--  if ((b2 & 0xC0) != 0x80)
--    return -1; /* invalid: not a continuation byte */
--  code_point = (code_point << 6) | (b2 & 0x3F);
--  if (b1 <= 0xDF)
--    return 0x7FF & code_point; /* two-byte character */
--
--  b3 = *++*input;
--  if ((b3 & 0xC0) != 0x80)
--    return -1; /* invalid: not a continuation byte */
--  code_point = (code_point << 6) | (b3 & 0x3F);
--  if (b1 <= 0xEF)
--    return 0xFFFF & code_point; /* three-byte character */
--
--  b4 = *++*input;
--  if ((b4 & 0xC0) != 0x80)
--    return -1; /* invalid: not a continuation byte */
--  code_point = (code_point << 6) | (b4 & 0x3F);
--  if (b1 <= 0xF4)
--    if (code_point <= 0x10FFFF)
--      return code_point; /* four-byte character */
--
--  /* code point too large */
--  return -1;
--}
--
--
--static ssize_t fs__get_length_wtf8(const char* source_ptr) {
--  size_t w_target_len = 0;
--  int32_t code_point;
--
--  do {
--    code_point = fs__decode_wtf8_char(&source_ptr);
--    if (code_point < 0)
--      return -1;
--    if (code_point > 0xFFFF)
--      w_target_len++;
--    w_target_len++;
--  } while (*source_ptr++);
--  return w_target_len;
--}
--
--
--static void fs__wtf8_to_wide(const char* source_ptr, WCHAR* w_target) {
--  int32_t code_point;
--
--  do {
--    code_point = fs__decode_wtf8_char(&source_ptr);
--    /* fs__get_length_wtf8 should have been called and checked first. */
--    assert(code_point >= 0);
--    if (code_point > 0x10000) {
--      assert(code_point < 0x10FFFF);
--      *w_target++ = (((code_point - 0x10000) >> 10) + 0xD800);
--      *w_target++ = ((code_point - 0x10000) & 0x3FF) + 0xDC00;
--    } else {
--      *w_target++ = code_point;
--    }
--  } while (*source_ptr++);
--}
--
--
- INLINE static int fs__capture_path(uv_fs_t* req, const char* path,
-     const char* new_path, const int copy_path) {
--  WCHAR* buf;
--  WCHAR* pos;
--  size_t buf_sz = 0;
--  size_t path_len = 0;
--  ssize_t pathw_len = 0;
--  ssize_t new_pathw_len = 0;
-+  char* buf;
-+  char* pos;
-+  ssize_t buf_sz = 0, path_len = 0, pathw_len = 0, new_pathw_len = 0;
- 
-   /* new_path can only be set if path is also set. */
-   assert(new_path == NULL || path != NULL);
- 
-   if (path != NULL) {
--    pathw_len = fs__get_length_wtf8(path);
--    if (pathw_len < 0)
--      return ERROR_INVALID_NAME;
-+    pathw_len = MultiByteToWideChar(CP_UTF8,
-+                                    0,
-+                                    path,
-+                                    -1,
-+                                    NULL,
-+                                    0);
-+    if (pathw_len == 0) {
-+      return GetLastError();
-+    }
-+
-     buf_sz += pathw_len * sizeof(WCHAR);
-   }
- 
-@@ -244,9 +173,16 @@ INLINE static int fs__capture_path(uv_fs_t* req, const char* path,
-   }
- 
-   if (new_path != NULL) {
--    new_pathw_len = fs__get_length_wtf8(new_path);
--    if (new_pathw_len < 0)
--      return ERROR_INVALID_NAME;
-+    new_pathw_len = MultiByteToWideChar(CP_UTF8,
-+                                        0,
-+                                        new_path,
-+                                        -1,
-+                                        NULL,
-+                                        0);
-+    if (new_pathw_len == 0) {
-+      return GetLastError();
-+    }
-+
-     buf_sz += new_pathw_len * sizeof(WCHAR);
-   }
- 
-@@ -258,7 +194,7 @@ INLINE static int fs__capture_path(uv_fs_t* req, const char* path,
-     return 0;
-   }
- 
--  buf = uv__malloc(buf_sz);
-+  buf = (char*) uv__malloc(buf_sz);
-   if (buf == NULL) {
-     return ERROR_OUTOFMEMORY;
-   }
-@@ -266,17 +202,29 @@ INLINE static int fs__capture_path(uv_fs_t* req, const char* path,
-   pos = buf;
- 
-   if (path != NULL) {
--    fs__wtf8_to_wide(path, pos);
--    req->file.pathw = pos;
--    pos += pathw_len;
-+    DWORD r = MultiByteToWideChar(CP_UTF8,
-+                                  0,
-+                                  path,
-+                                  -1,
-+                                  (WCHAR*) pos,
-+                                  pathw_len);
-+    assert(r == (DWORD) pathw_len);
-+    req->file.pathw = (WCHAR*) pos;
-+    pos += r * sizeof(WCHAR);
-   } else {
-     req->file.pathw = NULL;
-   }
- 
-   if (new_path != NULL) {
--    fs__wtf8_to_wide(new_path, pos);
--    req->fs.info.new_pathw = pos;
--    pos += new_pathw_len;
-+    DWORD r = MultiByteToWideChar(CP_UTF8,
-+                                  0,
-+                                  new_path,
-+                                  -1,
-+                                  (WCHAR*) pos,
-+                                  new_pathw_len);
-+    assert(r == (DWORD) new_pathw_len);
-+    req->fs.info.new_pathw = (WCHAR*) pos;
-+    pos += r * sizeof(WCHAR);
-   } else {
-     req->fs.info.new_pathw = NULL;
-   }
-@@ -284,8 +232,8 @@ INLINE static int fs__capture_path(uv_fs_t* req, const char* path,
-   req->path = path;
-   if (path != NULL && copy_path) {
-     memcpy(pos, path, path_len);
--    assert(path_len == buf_sz - (pos - buf) * sizeof(WCHAR));
--    req->path = (char*) pos;
-+    assert(path_len == buf_sz - (pos - buf));
-+    req->path = pos;
-   }
- 
-   req->flags |= UV_FS_FREE_PATHS;
-@@ -311,115 +259,57 @@ INLINE static void uv__fs_req_init(uv_loop_t* loop, uv_fs_t* req,
- }
- 
- 
--static int32_t fs__get_surrogate_value(const WCHAR* w_source_ptr,
--                                       size_t w_source_len) {
--  WCHAR u;
--  WCHAR next;
--
--  u = w_source_ptr[0];
--  if (u >= 0xD800 && u <= 0xDBFF && w_source_len > 1) {
--    next = w_source_ptr[1];
--    if (next >= 0xDC00 && next <= 0xDFFF)
--      return 0x10000 + ((u - 0xD800) << 10) + (next - 0xDC00);
--  }
--  return u;
--}
--
--
--static size_t fs__get_length_wide(const WCHAR* w_source_ptr,
--                                  size_t w_source_len) {
--  size_t target_len;
--  int32_t code_point;
-+static int fs__wide_to_utf8(WCHAR* w_source_ptr,
-+                               DWORD w_source_len,
-+                               char** target_ptr,
-+                               uint64_t* target_len_ptr) {
-+  int r;
-+  int target_len;
-+  char* target;
-+  target_len = WideCharToMultiByte(CP_UTF8,
-+                                   0,
-+                                   w_source_ptr,
-+                                   w_source_len,
-+                                   NULL,
-+                                   0,
-+                                   NULL,
-+                                   NULL);
- 
--  target_len = 0;
--  for (; w_source_len; w_source_len--, w_source_ptr++) {
--    code_point = fs__get_surrogate_value(w_source_ptr, w_source_len);
--    /* Can be invalid UTF-8 but must be valid WTF-8. */
--    assert(code_point >= 0);
--    if (code_point < 0x80)
--      target_len += 1;
--    else if (code_point < 0x800)
--      target_len += 2;
--    else if (code_point < 0x10000)
--      target_len += 3;
--    else {
--      target_len += 4;
--      w_source_ptr++;
--      w_source_len--;
--    }
-+  if (target_len == 0) {
-+    return -1;
-   }
--  return target_len;
--}
- 
--
--static int fs__wide_to_wtf8(WCHAR* w_source_ptr,
--                            size_t w_source_len,
--                            char** target_ptr,
--                            size_t* target_len_ptr) {
--  size_t target_len;
--  char* target;
--  int32_t code_point;
--
--  /* If *target_ptr is provided, then *target_len_ptr must be its length
--   * (excluding space for null), otherwise we will compute the target_len_ptr
--   * length and may return a new allocation in *target_ptr if target_ptr is
--   * provided. */
--  if (target_ptr == NULL || *target_ptr == NULL) {
--    target_len = fs__get_length_wide(w_source_ptr, w_source_len);
--    if (target_len_ptr != NULL)
--      *target_len_ptr = target_len;
--  } else {
--    target_len = *target_len_ptr;
-+  if (target_len_ptr != NULL) {
-+    *target_len_ptr = target_len;
-   }
- 
--  if (target_ptr == NULL)
-+  if (target_ptr == NULL) {
-     return 0;
--
--  if (*target_ptr == NULL) {
--    target = uv__malloc(target_len + 1);
--    if (target == NULL) {
--      SetLastError(ERROR_OUTOFMEMORY);
--      return -1;
--    }
--    *target_ptr = target;
--  } else {
--    target = *target_ptr;
--  }
--
--  for (; w_source_len; w_source_len--, w_source_ptr++) {
--    code_point = fs__get_surrogate_value(w_source_ptr, w_source_len);
--    /* Can be invalid UTF-8 but must be valid WTF-8. */
--    assert(code_point >= 0);
--
--    if (code_point < 0x80) {
--      *target++ = code_point;
--    } else if (code_point < 0x800) {
--      *target++ = 0xC0 | (code_point >> 6);
--      *target++ = 0x80 | (code_point & 0x3F);
--    } else if (code_point < 0x10000) {
--      *target++ = 0xE0 | (code_point >> 12);
--      *target++ = 0x80 | ((code_point >> 6) & 0x3F);
--      *target++ = 0x80 | (code_point & 0x3F);
--    } else {
--      *target++ = 0xF0 | (code_point >> 18);
--      *target++ = 0x80 | ((code_point >> 12) & 0x3F);
--      *target++ = 0x80 | ((code_point >> 6) & 0x3F);
--      *target++ = 0x80 | (code_point & 0x3F);
--      w_source_ptr++;
--      w_source_len--;
--    }
-   }
--  assert((size_t) (target - *target_ptr) == target_len);
- 
--  *target++ = '\0';
-+  target = uv__malloc(target_len + 1);
-+  if (target == NULL) {
-+    SetLastError(ERROR_OUTOFMEMORY);
-+    return -1;
-+  }
- 
-+  r = WideCharToMultiByte(CP_UTF8,
-+                          0,
-+                          w_source_ptr,
-+                          w_source_len,
-+                          target,
-+                          target_len,
-+                          NULL,
-+                          NULL);
-+  assert(r == target_len);
-+  target[target_len] = '\0';
-+  *target_ptr = target;
-   return 0;
- }
- 
- 
--INLINE static int fs__readlink_handle(HANDLE handle,
--                                      char** target_ptr,
--                                      size_t* target_len_ptr) {
-+INLINE static int fs__readlink_handle(HANDLE handle, char** target_ptr,
-+    uint64_t* target_len_ptr) {
-   char buffer[MAXIMUM_REPARSE_DATA_BUFFER_SIZE];
-   REPARSE_DATA_BUFFER* reparse_data = (REPARSE_DATA_BUFFER*) buffer;
-   WCHAR* w_target;
-@@ -549,8 +439,7 @@ INLINE static int fs__readlink_handle(HANDLE handle,
-     return -1;
-   }
- 
--  assert(target_ptr == NULL || *target_ptr == NULL);
--  return fs__wide_to_wtf8(w_target, w_target_len, target_ptr, target_len_ptr);
-+  return fs__wide_to_utf8(w_target, w_target_len, target_ptr, target_len_ptr);
- }
- 
- 
-@@ -1540,8 +1429,7 @@ void fs__scandir(uv_fs_t* req) {
-       uv__dirent_t* dirent;
- 
-       size_t wchar_len;
--      size_t wtf8_len;
--      char* wtf8;
-+      size_t utf8_len;
- 
-       /* Obtain a pointer to the current directory entry. */
-       position += next_entry_offset;
-@@ -1568,8 +1456,11 @@ void fs__scandir(uv_fs_t* req) {
-           info->FileName[1] == L'.')
-         continue;
- 
--      /* Compute the space required to store the filename as WTF-8. */
--      wtf8_len = fs__get_length_wide(&info->FileName[0], wchar_len);
-+      /* Compute the space required to store the filename as UTF-8. */
-+      utf8_len = WideCharToMultiByte(
-+          CP_UTF8, 0, &info->FileName[0], wchar_len, NULL, 0, NULL, NULL);
-+      if (utf8_len == 0)
-+        goto win32_error;
- 
-       /* Resize the dirent array if needed. */
-       if (dirents_used >= dirents_size) {
-@@ -1589,17 +1480,26 @@ void fs__scandir(uv_fs_t* req) {
-        * includes room for the first character of the filename, but `utf8_len`
-        * doesn't count the NULL terminator at this point.
-        */
--      dirent = uv__malloc(sizeof *dirent + wtf8_len);
-+      dirent = uv__malloc(sizeof *dirent + utf8_len);
-       if (dirent == NULL)
-         goto out_of_memory_error;
- 
-       dirents[dirents_used++] = dirent;
- 
-       /* Convert file name to UTF-8. */
--      wtf8 = &dirent->d_name[0];
--      if (fs__wide_to_wtf8(&info->FileName[0], wchar_len, &wtf8, &wtf8_len) == -1)
-+      if (WideCharToMultiByte(CP_UTF8,
-+                              0,
-+                              &info->FileName[0],
-+                              wchar_len,
-+                              &dirent->d_name[0],
-+                              utf8_len,
-+                              NULL,
-+                              NULL) == 0)
-         goto win32_error;
- 
-+      /* Add a null terminator to the filename. */
-+      dirent->d_name[utf8_len] = '\0';
-+
-       /* Fill out the type field. */
-       if (info->FileAttributes & FILE_ATTRIBUTE_DEVICE)
-         dirent->d_type = UV__DT_CHAR;
-@@ -1808,7 +1708,6 @@ void fs__closedir(uv_fs_t* req) {
- 
- INLINE static int fs__stat_handle(HANDLE handle, uv_stat_t* statbuf,
-     int do_lstat) {
--  size_t target_length = 0;
-   FILE_FS_DEVICE_INFORMATION device_info;
-   FILE_ALL_INFORMATION file_info;
-   FILE_FS_VOLUME_INFORMATION volume_info;
-@@ -1904,10 +1803,9 @@ INLINE static int fs__stat_handle(HANDLE handle, uv_stat_t* statbuf,
-      * to be treated as a regular file. The higher level lstat function will
-      * detect this failure and retry without do_lstat if appropriate.
-      */
--    if (fs__readlink_handle(handle, NULL, &target_length) != 0)
-+    if (fs__readlink_handle(handle, NULL, &statbuf->st_size) != 0)
-       return -1;
-     statbuf->st_mode |= S_IFLNK;
--    statbuf->st_size = target_length;
-   }
- 
-   if (statbuf->st_mode == 0) {
-@@ -2763,7 +2661,6 @@ static void fs__readlink(uv_fs_t* req) {
-     return;
-   }
- 
--  assert(req->ptr == NULL);
-   if (fs__readlink_handle(handle, (char**) &req->ptr, NULL) != 0) {
-     DWORD error = GetLastError();
-     SET_REQ_WIN32_ERROR(req, error);
-@@ -2823,8 +2720,7 @@ static ssize_t fs__realpath_handle(HANDLE handle, char** realpath_ptr) {
-     return -1;
-   }
- 
--  assert(*realpath_ptr == NULL);
--  r = fs__wide_to_wtf8(w_realpath_ptr, w_realpath_len, realpath_ptr, NULL);
-+  r = fs__wide_to_utf8(w_realpath_ptr, w_realpath_len, realpath_ptr, NULL);
-   uv__free(w_realpath_buf);
-   return r;
- }
-@@ -2844,7 +2740,6 @@ static void fs__realpath(uv_fs_t* req) {
-     return;
-   }
- 
--  assert(req->ptr == NULL);
-   if (fs__realpath_handle(handle, (char**) &req->ptr) == -1) {
-     CloseHandle(handle);
-     SET_REQ_WIN32_ERROR(req, GetLastError());
-diff --git a/deps/uv/src/win/handle-inl.h b/deps/uv/src/win/handle-inl.h
-index 4722e85790a..5c843c241ef 100644
---- a/deps/uv/src/win/handle-inl.h
-+++ b/deps/uv/src/win/handle-inl.h
-@@ -75,7 +75,7 @@
- 
- #define uv__handle_close(handle)                                        \
-   do {                                                                  \
--    uv__queue_remove(&(handle)->handle_queue);                          \
-+    QUEUE_REMOVE(&(handle)->handle_queue);                              \
-     uv__active_handle_rm((uv_handle_t*) (handle));                      \
-                                                                         \
-     (handle)->flags |= UV_HANDLE_CLOSED;                                \
-diff --git a/deps/uv/src/win/pipe.c b/deps/uv/src/win/pipe.c
-index f0cac382256..5e4276387ac 100644
---- a/deps/uv/src/win/pipe.c
-+++ b/deps/uv/src/win/pipe.c
-@@ -55,7 +55,7 @@ static const int pipe_prefix_len = sizeof(pipe_prefix) - 1;
- typedef struct {
-   uv__ipc_socket_xfer_type_t xfer_type;
-   uv__ipc_socket_xfer_info_t xfer_info;
--  struct uv__queue member;
-+  QUEUE member;
- } uv__ipc_xfer_queue_item_t;
- 
- /* IPC frame header flags. */
-@@ -111,7 +111,7 @@ int uv_pipe_init(uv_loop_t* loop, uv_pipe_t* handle, int ipc) {
-   handle->name = NULL;
-   handle->pipe.conn.ipc_remote_pid = 0;
-   handle->pipe.conn.ipc_data_frame.payload_remaining = 0;
--  uv__queue_init(&handle->pipe.conn.ipc_xfer_queue);
-+  QUEUE_INIT(&handle->pipe.conn.ipc_xfer_queue);
-   handle->pipe.conn.ipc_xfer_queue_length = 0;
-   handle->ipc = ipc;
-   handle->pipe.conn.non_overlapped_writes_tail = NULL;
-@@ -637,13 +637,13 @@ void uv__pipe_endgame(uv_loop_t* loop, uv_pipe_t* handle) {
- 
-   if (handle->flags & UV_HANDLE_CONNECTION) {
-     /* Free pending sockets */
--    while (!uv__queue_empty(&handle->pipe.conn.ipc_xfer_queue)) {
--      struct uv__queue* q;
-+    while (!QUEUE_EMPTY(&handle->pipe.conn.ipc_xfer_queue)) {
-+      QUEUE* q;
-       SOCKET socket;
- 
--      q = uv__queue_head(&handle->pipe.conn.ipc_xfer_queue);
--      uv__queue_remove(q);
--      xfer_queue_item = uv__queue_data(q, uv__ipc_xfer_queue_item_t, member);
-+      q = QUEUE_HEAD(&handle->pipe.conn.ipc_xfer_queue);
-+      QUEUE_REMOVE(q);
-+      xfer_queue_item = QUEUE_DATA(q, uv__ipc_xfer_queue_item_t, member);
- 
-       /* Materialize socket and close it */
-       socket = WSASocketW(FROM_PROTOCOL_INFO,
-@@ -1124,20 +1124,20 @@ int uv__pipe_accept(uv_pipe_t* server, uv_stream_t* client) {
-   uv_loop_t* loop = server->loop;
-   uv_pipe_t* pipe_client;
-   uv_pipe_accept_t* req;
--  struct uv__queue* q;
-+  QUEUE* q;
-   uv__ipc_xfer_queue_item_t* item;
-   int err;
- 
-   if (server->ipc) {
--    if (uv__queue_empty(&server->pipe.conn.ipc_xfer_queue)) {
-+    if (QUEUE_EMPTY(&server->pipe.conn.ipc_xfer_queue)) {
-       /* No valid pending sockets. */
-       return WSAEWOULDBLOCK;
-     }
- 
--    q = uv__queue_head(&server->pipe.conn.ipc_xfer_queue);
--    uv__queue_remove(q);
-+    q = QUEUE_HEAD(&server->pipe.conn.ipc_xfer_queue);
-+    QUEUE_REMOVE(q);
-     server->pipe.conn.ipc_xfer_queue_length--;
--    item = uv__queue_data(q, uv__ipc_xfer_queue_item_t, member);
-+    item = QUEUE_DATA(q, uv__ipc_xfer_queue_item_t, member);
- 
-     err = uv__tcp_xfer_import(
-         (uv_tcp_t*) client, item->xfer_type, &item->xfer_info);
-@@ -1891,7 +1891,7 @@ static void uv__pipe_queue_ipc_xfer_info(
-   item->xfer_type = xfer_type;
-   item->xfer_info = *xfer_info;
- 
--  uv__queue_insert_tail(&handle->pipe.conn.ipc_xfer_queue, &item->member);
-+  QUEUE_INSERT_TAIL(&handle->pipe.conn.ipc_xfer_queue, &item->member);
-   handle->pipe.conn.ipc_xfer_queue_length++;
- }
- 
-diff --git a/deps/uv/src/win/tcp.c b/deps/uv/src/win/tcp.c
-index 187f36e2a61..6b282e0b501 100644
---- a/deps/uv/src/win/tcp.c
-+++ b/deps/uv/src/win/tcp.c
-@@ -175,14 +175,14 @@ int uv_tcp_init_ex(uv_loop_t* loop, uv_tcp_t* handle, unsigned int flags) {
-     sock = socket(domain, SOCK_STREAM, 0);
-     if (sock == INVALID_SOCKET) {
-       err = WSAGetLastError();
--      uv__queue_remove(&handle->handle_queue);
-+      QUEUE_REMOVE(&handle->handle_queue);
-       return uv_translate_sys_error(err);
-     }
- 
-     err = uv__tcp_set_socket(handle->loop, handle, sock, domain, 0);
-     if (err) {
-       closesocket(sock);
--      uv__queue_remove(&handle->handle_queue);
-+      QUEUE_REMOVE(&handle->handle_queue);
-       return uv_translate_sys_error(err);
-     }
- 
-diff --git a/deps/uv/src/win/udp.c b/deps/uv/src/win/udp.c
-index eab53842d4f..8a982d1907d 100644
---- a/deps/uv/src/win/udp.c
-+++ b/deps/uv/src/win/udp.c
-@@ -146,14 +146,14 @@ int uv__udp_init_ex(uv_loop_t* loop,
-     sock = socket(domain, SOCK_DGRAM, 0);
-     if (sock == INVALID_SOCKET) {
-       err = WSAGetLastError();
--      uv__queue_remove(&handle->handle_queue);
-+      QUEUE_REMOVE(&handle->handle_queue);
-       return uv_translate_sys_error(err);
-     }
- 
-     err = uv__udp_set_socket(handle->loop, handle, sock, domain);
-     if (err) {
-       closesocket(sock);
--      uv__queue_remove(&handle->handle_queue);
-+      QUEUE_REMOVE(&handle->handle_queue);
-       return uv_translate_sys_error(err);
-     }
-   }
diff --git a/meta-oe/recipes-devtools/nodejs/nodejs/0001-deps-disable-io_uring-support-in-libuv.patch b/meta-oe/recipes-devtools/nodejs/nodejs/0001-deps-disable-io_uring-support-in-libuv.patch
new file mode 100644
index 000000000..f11cd08cb
--- /dev/null
+++ b/meta-oe/recipes-devtools/nodejs/nodejs/0001-deps-disable-io_uring-support-in-libuv.patch
@@ -0,0 +1,35 @@ 
+From 9838be9c710ab4249df86726fa390232a3b6a6e7 Mon Sep 17 00:00:00 2001
+From: Changqing Li <changqing.li@windriver.com>
+Date: Fri, 1 Mar 2024 15:46:11 +0800
+Subject: [PATCH] deps: disable io_uring support in libuv
+
+Refer [1], Pseudo fails to intercept some of the syscalls when io_uring
+enabled. Refer [2], always disable io_uring support in libuv to fix
+issue in [1].
+
+[1] https://git.openembedded.org/meta-openembedded/commit/?id=d08453978c31ee41d28206c6ff198d7d9d701d88
+[2] https://github.com/nodejs/node/commit/686da19abb
+
+Upstream-Status: Inappropriate [oe specific]
+
+Signed-off-by: Changqing Li <changqing.li@windriver.com>
+---
+ deps/uv/src/unix/linux.c | 2 +-
+ 1 file changed, 1 insertion(+), 1 deletion(-)
+
+diff --git a/deps/uv/src/unix/linux.c b/deps/uv/src/unix/linux.c
+index 0c997185..7508409d 100644
+--- a/deps/uv/src/unix/linux.c
++++ b/deps/uv/src/unix/linux.c
+@@ -433,7 +433,7 @@ static int uv__use_io_uring(void) {
+   if (use == 0) {
+     /* Disable io_uring by default due to CVE-2024-22017. */
+     val = getenv("UV_USE_IO_URING");
+-    use = val != NULL && atoi(val) ? 1 : -1;
++    use = 0;
+     atomic_store_explicit(&use_io_uring, use, memory_order_relaxed);
+   }
+ 
+-- 
+2.25.1
+
diff --git a/meta-oe/recipes-devtools/nodejs/nodejs/0002-Revert-io_uring-changes-from-libuv-1.45.0.patch b/meta-oe/recipes-devtools/nodejs/nodejs/0002-Revert-io_uring-changes-from-libuv-1.45.0.patch
deleted file mode 100644
index 77cd53b75..000000000
--- a/meta-oe/recipes-devtools/nodejs/nodejs/0002-Revert-io_uring-changes-from-libuv-1.45.0.patch
+++ /dev/null
@@ -1,1803 +0,0 @@ 
-From 6d2ef4c8ba2304ee4941a6719b3ad9bd63e415a9 Mon Sep 17 00:00:00 2001
-From: Martin Jansa <martin.jansa@gmail.com>
-Date: Wed, 18 Oct 2023 21:09:44 +0200
-Subject: [PATCH] Revert io_uring changes from libuv-1.45.0
-
-This reverts https://github.com/libuv/libuv/pull/3952/commits/26c79a942b92573a1388c0ee8a6ad4397f009318
-
-Included in nodejs-20.3.0 with the libuv upgrade to 1.45.0 in:
-https://github.com/nodejs/node/commit/bfcb3d1d9a876f399013d326bd65804f9eda77e4
-
-Reverted libuv commits:
-Revert "linux: fix WRITEV with lots of bufs using io_uring (#4004)"
-This reverts commit ef6a9a624df0a00687037474025a3608472f722a.
-Revert "linux: work around EOWNERDEAD io_uring kernel bug (#4002)"
-This reverts commit d23a20f62cc50b9fd7694992263f1d296d8f5cb4.
-Revert "unix: handle CQ overflow in iou ring (#3991)"
-This reverts commit 30fc896cc1b5822e9f1eb462587fe4b368a6215c.
-Revert "unix: constrained_memory should return UINT64_MAX (#3753)"
-This reverts commit 6ad347fae4520f39520d34bd7c7f5ddafab13a69.
-Revert "linux: use io_uring to batch epoll_ctl calls (#3979)"
-This reverts commit 6e073ef5daf93b708a654008959b823b58029e88.
-Revert "linux: fix logic bug in sqe ring space check (#3980)"
-This reverts commit f27208224084fc972b9d2802486d97ef31b51a39.
-Revert "src: fix events/events_waiting metrics counter (#3957)"
-This reverts commit e02642cf3b768b2c58a41f97fa38507e032ae415.
-Revert "linux: remove bug workaround for obsolete kernels (#3965)"
-This reverts commit 1c935a34454167b23f8eef7f0f63d7119f0de747.
-Revert "linux: add IORING_OP_CLOSE support (#3964)"
-This reverts commit dfae365f844e127621128a76bce7165e3f99a8d9.
-Revert "linux: add IORING_OP_OPENAT support (#3963)"
-This reverts commit 5ca5e475bb1711e65323ef1594a31818e5a1a9eb.
-Revert "linux: fix academic valgrind warning (#3960)"
-This reverts commit a7ff759ca1deacb2e0e6ae3c2d3dce91cc637dfe.
-Revert "linux: introduce io_uring support (#3952)"
-This reverts commit d2c31f429b87b476a7f1344d145dad4752a406d4.
-
-Dropped deps/uv/docs deps/uv/test changes as these dirs aren't included
-in nodejs tarballs.
-
-Signed-off-by: Martin Jansa <martin.jansa@gmail.com>
----
-Upstream-Status: Inappropriate [OE specific]
-
- deps/uv/docs/src/fs.rst               |   6 -
- deps/uv/docs/src/misc.rst             |   5 +-
- deps/uv/src/threadpool.c              |  21 -
- deps/uv/src/unix/aix.c                |  11 +-
- deps/uv/src/unix/fs.c                 |  57 +-
- deps/uv/src/unix/internal.h           |  23 -
- deps/uv/src/unix/kqueue.c             |  10 +-
- deps/uv/src/unix/linux.c              | 998 +++-----------------------
- deps/uv/src/unix/os390.c              |  11 +-
- deps/uv/src/unix/posix-poll.c         |  11 +-
- deps/uv/src/uv-common.h               |  28 -
- deps/uv/src/win/core.c                |  20 +-
- deps/uv/test/test-fs.c                |  33 -
- deps/uv/test/test-list.h              |   4 -
- deps/uv/test/test-metrics.c           | 151 ----
- deps/uv/test/test-threadpool-cancel.c |  34 +-
- 16 files changed, 138 insertions(+), 1285 deletions(-)
-
-diff --git a/deps/uv/src/threadpool.c b/deps/uv/src/threadpool.c
-index 51962bf0021..a3da53026f9 100644
---- a/deps/uv/src/threadpool.c
-+++ b/deps/uv/src/threadpool.c
-@@ -275,13 +275,9 @@ void uv__work_submit(uv_loop_t* loop,
- }
- 
- 
--/* TODO(bnoordhuis) teach libuv how to cancel file operations
-- * that go through io_uring instead of the thread pool.
-- */
- static int uv__work_cancel(uv_loop_t* loop, uv_req_t* req, struct uv__work* w) {
-   int cancelled;
- 
--  uv_once(&once, init_once);  /* Ensure |mutex| is initialized. */
-   uv_mutex_lock(&mutex);
-   uv_mutex_lock(&w->loop->wq_mutex);
- 
-@@ -311,15 +307,12 @@ void uv__work_done(uv_async_t* handle) {
-   QUEUE* q;
-   QUEUE wq;
-   int err;
--  int nevents;
- 
-   loop = container_of(handle, uv_loop_t, wq_async);
-   uv_mutex_lock(&loop->wq_mutex);
-   QUEUE_MOVE(&loop->wq, &wq);
-   uv_mutex_unlock(&loop->wq_mutex);
- 
--  nevents = 0;
--
-   while (!QUEUE_EMPTY(&wq)) {
-     q = QUEUE_HEAD(&wq);
-     QUEUE_REMOVE(q);
-@@ -327,20 +320,6 @@ void uv__work_done(uv_async_t* handle) {
-     w = container_of(q, struct uv__work, wq);
-     err = (w->work == uv__cancelled) ? UV_ECANCELED : 0;
-     w->done(w, err);
--    nevents++;
--  }
--
--  /* This check accomplishes 2 things:
--   * 1. Even if the queue was empty, the call to uv__work_done() should count
--   *    as an event. Which will have been added by the event loop when
--   *    calling this callback.
--   * 2. Prevents accidental wrap around in case nevents == 0 events == 0.
--   */
--  if (nevents > 1) {
--    /* Subtract 1 to counter the call to uv__work_done(). */
--    uv__metrics_inc_events(loop, nevents - 1);
--    if (uv__get_internal_fields(loop)->current_timeout == 0)
--      uv__metrics_inc_events_waiting(loop, nevents - 1);
-   }
- }
- 
-diff --git a/deps/uv/src/unix/aix.c b/deps/uv/src/unix/aix.c
-index f1afbed49ec..b855282ebc8 100644
---- a/deps/uv/src/unix/aix.c
-+++ b/deps/uv/src/unix/aix.c
-@@ -131,7 +131,6 @@ int uv__io_check_fd(uv_loop_t* loop, int fd) {
- 
- 
- void uv__io_poll(uv_loop_t* loop, int timeout) {
--  uv__loop_internal_fields_t* lfields;
-   struct pollfd events[1024];
-   struct pollfd pqry;
-   struct pollfd* pe;
-@@ -155,8 +154,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-     return;
-   }
- 
--  lfields = uv__get_internal_fields(loop);
--
-   while (!QUEUE_EMPTY(&loop->watcher_queue)) {
-     q = QUEUE_HEAD(&loop->watcher_queue);
-     QUEUE_REMOVE(q);
-@@ -220,7 +217,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-   base = loop->time;
-   count = 48; /* Benchmarks suggest this gives the best throughput. */
- 
--  if (lfields->flags & UV_METRICS_IDLE_TIME) {
-+  if (uv__get_internal_fields(loop)->flags & UV_METRICS_IDLE_TIME) {
-     reset_timeout = 1;
-     user_timeout = timeout;
-     timeout = 0;
-@@ -235,12 +232,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-     if (timeout != 0)
-       uv__metrics_set_provider_entry_time(loop);
- 
--    /* Store the current timeout in a location that's globally accessible so
--     * other locations like uv__work_done() can determine whether the queue
--     * of events in the callback were waiting when poll was called.
--     */
--    lfields->current_timeout = timeout;
--
-     nfds = pollset_poll(loop->backend_fd,
-                         events,
-                         ARRAY_SIZE(events),
-diff --git a/deps/uv/src/unix/fs.c b/deps/uv/src/unix/fs.c
-index 00d385c24b7..c696a8d5500 100644
---- a/deps/uv/src/unix/fs.c
-+++ b/deps/uv/src/unix/fs.c
-@@ -61,11 +61,10 @@
- #endif
- 
- #if defined(__linux__)
--# include <sys/sendfile.h>
--# include <sys/utsname.h>
-+# include "sys/utsname.h"
- #endif
- 
--#if defined(__sun)
-+#if defined(__linux__) || defined(__sun)
- # include <sys/sendfile.h>
- # include <sys/sysmacros.h>
- #endif
-@@ -1554,7 +1553,26 @@ static int uv__fs_statx(int fd,
-     return UV_ENOSYS;
-   }
- 
--  uv__statx_to_stat(&statxbuf, buf);
-+  buf->st_dev = makedev(statxbuf.stx_dev_major, statxbuf.stx_dev_minor);
-+  buf->st_mode = statxbuf.stx_mode;
-+  buf->st_nlink = statxbuf.stx_nlink;
-+  buf->st_uid = statxbuf.stx_uid;
-+  buf->st_gid = statxbuf.stx_gid;
-+  buf->st_rdev = makedev(statxbuf.stx_rdev_major, statxbuf.stx_rdev_minor);
-+  buf->st_ino = statxbuf.stx_ino;
-+  buf->st_size = statxbuf.stx_size;
-+  buf->st_blksize = statxbuf.stx_blksize;
-+  buf->st_blocks = statxbuf.stx_blocks;
-+  buf->st_atim.tv_sec = statxbuf.stx_atime.tv_sec;
-+  buf->st_atim.tv_nsec = statxbuf.stx_atime.tv_nsec;
-+  buf->st_mtim.tv_sec = statxbuf.stx_mtime.tv_sec;
-+  buf->st_mtim.tv_nsec = statxbuf.stx_mtime.tv_nsec;
-+  buf->st_ctim.tv_sec = statxbuf.stx_ctime.tv_sec;
-+  buf->st_ctim.tv_nsec = statxbuf.stx_ctime.tv_nsec;
-+  buf->st_birthtim.tv_sec = statxbuf.stx_btime.tv_sec;
-+  buf->st_birthtim.tv_nsec = statxbuf.stx_btime.tv_nsec;
-+  buf->st_flags = 0;
-+  buf->st_gen = 0;
- 
-   return 0;
- #else
-@@ -1798,9 +1816,6 @@ int uv_fs_chown(uv_loop_t* loop,
- int uv_fs_close(uv_loop_t* loop, uv_fs_t* req, uv_file file, uv_fs_cb cb) {
-   INIT(CLOSE);
-   req->file = file;
--  if (cb != NULL)
--    if (uv__iou_fs_close(loop, req))
--      return 0;
-   POST;
- }
- 
-@@ -1848,9 +1863,6 @@ int uv_fs_lchown(uv_loop_t* loop,
- int uv_fs_fdatasync(uv_loop_t* loop, uv_fs_t* req, uv_file file, uv_fs_cb cb) {
-   INIT(FDATASYNC);
-   req->file = file;
--  if (cb != NULL)
--    if (uv__iou_fs_fsync_or_fdatasync(loop, req, /* IORING_FSYNC_DATASYNC */ 1))
--      return 0;
-   POST;
- }
- 
-@@ -1858,9 +1870,6 @@ int uv_fs_fdatasync(uv_loop_t* loop, uv_fs_t* req, uv_file file, uv_fs_cb cb) {
- int uv_fs_fstat(uv_loop_t* loop, uv_fs_t* req, uv_file file, uv_fs_cb cb) {
-   INIT(FSTAT);
-   req->file = file;
--  if (cb != NULL)
--    if (uv__iou_fs_statx(loop, req, /* is_fstat */ 1, /* is_lstat */ 0))
--      return 0;
-   POST;
- }
- 
-@@ -1868,9 +1877,6 @@ int uv_fs_fstat(uv_loop_t* loop, uv_fs_t* req, uv_file file, uv_fs_cb cb) {
- int uv_fs_fsync(uv_loop_t* loop, uv_fs_t* req, uv_file file, uv_fs_cb cb) {
-   INIT(FSYNC);
-   req->file = file;
--  if (cb != NULL)
--    if (uv__iou_fs_fsync_or_fdatasync(loop, req, /* no flags */ 0))
--      return 0;
-   POST;
- }
- 
-@@ -1917,9 +1923,6 @@ int uv_fs_lutime(uv_loop_t* loop,
- int uv_fs_lstat(uv_loop_t* loop, uv_fs_t* req, const char* path, uv_fs_cb cb) {
-   INIT(LSTAT);
-   PATH;
--  if (cb != NULL)
--    if (uv__iou_fs_statx(loop, req, /* is_fstat */ 0, /* is_lstat */ 1))
--      return 0;
-   POST;
- }
- 
-@@ -1981,9 +1984,6 @@ int uv_fs_open(uv_loop_t* loop,
-   PATH;
-   req->flags = flags;
-   req->mode = mode;
--  if (cb != NULL)
--    if (uv__iou_fs_open(loop, req))
--      return 0;
-   POST;
- }
- 
-@@ -2012,11 +2012,6 @@ int uv_fs_read(uv_loop_t* loop, uv_fs_t* req,
-   memcpy(req->bufs, bufs, nbufs * sizeof(*bufs));
- 
-   req->off = off;
--
--  if (cb != NULL)
--    if (uv__iou_fs_read_or_write(loop, req, /* is_read */ 1))
--      return 0;
--
-   POST;
- }
- 
-@@ -2124,9 +2119,6 @@ int uv_fs_sendfile(uv_loop_t* loop,
- int uv_fs_stat(uv_loop_t* loop, uv_fs_t* req, const char* path, uv_fs_cb cb) {
-   INIT(STAT);
-   PATH;
--  if (cb != NULL)
--    if (uv__iou_fs_statx(loop, req, /* is_fstat */ 0, /* is_lstat */ 0))
--      return 0;
-   POST;
- }
- 
-@@ -2190,11 +2182,6 @@ int uv_fs_write(uv_loop_t* loop,
-   memcpy(req->bufs, bufs, nbufs * sizeof(*bufs));
- 
-   req->off = off;
--
--  if (cb != NULL)
--    if (uv__iou_fs_read_or_write(loop, req, /* is_read */ 0))
--      return 0;
--
-   POST;
- }
- 
-diff --git a/deps/uv/src/unix/internal.h b/deps/uv/src/unix/internal.h
-index 6c5822e6a0d..d439ae6dd3d 100644
---- a/deps/uv/src/unix/internal.h
-+++ b/deps/uv/src/unix/internal.h
-@@ -329,28 +329,6 @@ int uv__random_getentropy(void* buf, size_t buflen);
- int uv__random_readpath(const char* path, void* buf, size_t buflen);
- int uv__random_sysctl(void* buf, size_t buflen);
- 
--/* io_uring */
--#ifdef __linux__
--int uv__iou_fs_close(uv_loop_t* loop, uv_fs_t* req);
--int uv__iou_fs_fsync_or_fdatasync(uv_loop_t* loop,
--                                  uv_fs_t* req,
--                                  uint32_t fsync_flags);
--int uv__iou_fs_open(uv_loop_t* loop, uv_fs_t* req);
--int uv__iou_fs_read_or_write(uv_loop_t* loop,
--                             uv_fs_t* req,
--                             int is_read);
--int uv__iou_fs_statx(uv_loop_t* loop,
--                     uv_fs_t* req,
--                     int is_fstat,
--                     int is_lstat);
--#else
--#define uv__iou_fs_close(loop, req) 0
--#define uv__iou_fs_fsync_or_fdatasync(loop, req, fsync_flags) 0
--#define uv__iou_fs_open(loop, req) 0
--#define uv__iou_fs_read_or_write(loop, req, is_read) 0
--#define uv__iou_fs_statx(loop, req, is_fstat, is_lstat) 0
--#endif
--
- #if defined(__APPLE__)
- int uv___stream_fd(const uv_stream_t* handle);
- #define uv__stream_fd(handle) (uv___stream_fd((const uv_stream_t*) (handle)))
-@@ -427,7 +405,6 @@ int uv__statx(int dirfd,
-               int flags,
-               unsigned int mask,
-               struct uv__statx* statxbuf);
--void uv__statx_to_stat(const struct uv__statx* statxbuf, uv_stat_t* buf);
- ssize_t uv__getrandom(void* buf, size_t buflen, unsigned flags);
- #endif
- 
-diff --git a/deps/uv/src/unix/kqueue.c b/deps/uv/src/unix/kqueue.c
-index 82916d65933..deb486bae7a 100644
---- a/deps/uv/src/unix/kqueue.c
-+++ b/deps/uv/src/unix/kqueue.c
-@@ -127,7 +127,6 @@ static void uv__kqueue_delete(int kqfd, const struct kevent *ev) {
- 
- 
- void uv__io_poll(uv_loop_t* loop, int timeout) {
--  uv__loop_internal_fields_t* lfields;
-   struct kevent events[1024];
-   struct kevent* ev;
-   struct timespec spec;
-@@ -156,7 +155,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-     return;
-   }
- 
--  lfields = uv__get_internal_fields(loop);
-   nevents = 0;
- 
-   while (!QUEUE_EMPTY(&loop->watcher_queue)) {
-@@ -224,7 +222,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-   base = loop->time;
-   count = 48; /* Benchmarks suggest this gives the best throughput. */
- 
--  if (lfields->flags & UV_METRICS_IDLE_TIME) {
-+  if (uv__get_internal_fields(loop)->flags & UV_METRICS_IDLE_TIME) {
-     reset_timeout = 1;
-     user_timeout = timeout;
-     timeout = 0;
-@@ -247,12 +245,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-     if (pset != NULL)
-       pthread_sigmask(SIG_BLOCK, pset, NULL);
- 
--    /* Store the current timeout in a location that's globally accessible so
--     * other locations like uv__work_done() can determine whether the queue
--     * of events in the callback were waiting when poll was called.
--     */
--    lfields->current_timeout = timeout;
--
-     nfds = kevent(loop->backend_fd,
-                   events,
-                   nevents,
-diff --git a/deps/uv/src/unix/linux.c b/deps/uv/src/unix/linux.c
-index 5f84ad0eea3..343e37f2527 100644
---- a/deps/uv/src/unix/linux.c
-+++ b/deps/uv/src/unix/linux.c
-@@ -27,8 +27,6 @@
- #include "internal.h"
- 
- #include <inttypes.h>
--#include <stdatomic.h>
--#include <stddef.h>  /* offsetof */
- #include <stdint.h>
- #include <stdio.h>
- #include <stdlib.h>
-@@ -40,29 +38,15 @@
- #include <net/if.h>
- #include <sys/epoll.h>
- #include <sys/inotify.h>
--#include <sys/mman.h>
- #include <sys/param.h>
- #include <sys/prctl.h>
- #include <sys/stat.h>
- #include <sys/syscall.h>
- #include <sys/sysinfo.h>
--#include <sys/sysmacros.h>
- #include <sys/types.h>
- #include <time.h>
- #include <unistd.h>
- 
--#ifndef __NR_io_uring_setup
--# define __NR_io_uring_setup 425
--#endif
--
--#ifndef __NR_io_uring_enter
--# define __NR_io_uring_enter 426
--#endif
--
--#ifndef __NR_io_uring_register
--# define __NR_io_uring_register 427
--#endif
--
- #ifndef __NR_copy_file_range
- # if defined(__x86_64__)
- #  define __NR_copy_file_range 326
-@@ -132,129 +116,6 @@
- # include <netpacket/packet.h>
- #endif /* HAVE_IFADDRS_H */
- 
--enum {
--  UV__IORING_SETUP_SQPOLL = 2u,
--};
--
--enum {
--  UV__IORING_FEAT_SINGLE_MMAP = 1u,
--  UV__IORING_FEAT_NODROP = 2u,
--  UV__IORING_FEAT_RSRC_TAGS = 1024u,  /* linux v5.13 */
--};
--
--enum {
--  UV__IORING_OP_READV = 1,
--  UV__IORING_OP_WRITEV = 2,
--  UV__IORING_OP_FSYNC = 3,
--  UV__IORING_OP_OPENAT = 18,
--  UV__IORING_OP_CLOSE = 19,
--  UV__IORING_OP_STATX = 21,
--  UV__IORING_OP_EPOLL_CTL = 29,
--};
--
--enum {
--  UV__IORING_ENTER_GETEVENTS = 1u,
--  UV__IORING_ENTER_SQ_WAKEUP = 2u,
--};
--
--enum {
--  UV__IORING_SQ_NEED_WAKEUP = 1u,
--  UV__IORING_SQ_CQ_OVERFLOW = 2u,
--};
--
--struct uv__io_cqring_offsets {
--  uint32_t head;
--  uint32_t tail;
--  uint32_t ring_mask;
--  uint32_t ring_entries;
--  uint32_t overflow;
--  uint32_t cqes;
--  uint64_t reserved0;
--  uint64_t reserved1;
--};
--
--STATIC_ASSERT(40 == sizeof(struct uv__io_cqring_offsets));
--
--struct uv__io_sqring_offsets {
--  uint32_t head;
--  uint32_t tail;
--  uint32_t ring_mask;
--  uint32_t ring_entries;
--  uint32_t flags;
--  uint32_t dropped;
--  uint32_t array;
--  uint32_t reserved0;
--  uint64_t reserved1;
--};
--
--STATIC_ASSERT(40 == sizeof(struct uv__io_sqring_offsets));
--
--struct uv__io_uring_cqe {
--  uint64_t user_data;
--  int32_t res;
--  uint32_t flags;
--};
--
--STATIC_ASSERT(16 == sizeof(struct uv__io_uring_cqe));
--
--struct uv__io_uring_sqe {
--  uint8_t opcode;
--  uint8_t flags;
--  uint16_t ioprio;
--  int32_t fd;
--  union {
--    uint64_t off;
--    uint64_t addr2;
--  };
--  union {
--    uint64_t addr;
--  };
--  uint32_t len;
--  union {
--    uint32_t rw_flags;
--    uint32_t fsync_flags;
--    uint32_t open_flags;
--    uint32_t statx_flags;
--  };
--  uint64_t user_data;
--  union {
--    uint16_t buf_index;
--    uint64_t pad[3];
--  };
--};
--
--STATIC_ASSERT(64 == sizeof(struct uv__io_uring_sqe));
--STATIC_ASSERT(0 == offsetof(struct uv__io_uring_sqe, opcode));
--STATIC_ASSERT(1 == offsetof(struct uv__io_uring_sqe, flags));
--STATIC_ASSERT(2 == offsetof(struct uv__io_uring_sqe, ioprio));
--STATIC_ASSERT(4 == offsetof(struct uv__io_uring_sqe, fd));
--STATIC_ASSERT(8 == offsetof(struct uv__io_uring_sqe, off));
--STATIC_ASSERT(16 == offsetof(struct uv__io_uring_sqe, addr));
--STATIC_ASSERT(24 == offsetof(struct uv__io_uring_sqe, len));
--STATIC_ASSERT(28 == offsetof(struct uv__io_uring_sqe, rw_flags));
--STATIC_ASSERT(32 == offsetof(struct uv__io_uring_sqe, user_data));
--STATIC_ASSERT(40 == offsetof(struct uv__io_uring_sqe, buf_index));
--
--struct uv__io_uring_params {
--  uint32_t sq_entries;
--  uint32_t cq_entries;
--  uint32_t flags;
--  uint32_t sq_thread_cpu;
--  uint32_t sq_thread_idle;
--  uint32_t features;
--  uint32_t reserved[4];
--  struct uv__io_sqring_offsets sq_off;  /* 40 bytes */
--  struct uv__io_cqring_offsets cq_off;  /* 40 bytes */
--};
--
--STATIC_ASSERT(40 + 40 + 40 == sizeof(struct uv__io_uring_params));
--STATIC_ASSERT(40 == offsetof(struct uv__io_uring_params, sq_off));
--STATIC_ASSERT(80 == offsetof(struct uv__io_uring_params, cq_off));
--
--STATIC_ASSERT(EPOLL_CTL_ADD < 4);
--STATIC_ASSERT(EPOLL_CTL_DEL < 4);
--STATIC_ASSERT(EPOLL_CTL_MOD < 4);
--
- struct watcher_list {
-   RB_ENTRY(watcher_list) entry;
-   QUEUE watchers;
-@@ -276,17 +137,6 @@ static int compare_watchers(const struct watcher_list* a,
- static void maybe_free_watcher_list(struct watcher_list* w,
-                                     uv_loop_t* loop);
- 
--static void uv__epoll_ctl_flush(int epollfd,
--                                struct uv__iou* ctl,
--                                struct epoll_event (*events)[256]);
--
--static void uv__epoll_ctl_prep(int epollfd,
--                               struct uv__iou* ctl,
--                               struct epoll_event (*events)[256],
--                               int op,
--                               int fd,
--                               struct epoll_event* e);
--
- RB_GENERATE_STATIC(watcher_root, watcher_list, entry, compare_watchers)
- 
- 
-@@ -356,187 +206,7 @@ ssize_t uv__getrandom(void* buf, size_t buflen, unsigned flags) {
- }
- 
- 
--int uv__io_uring_setup(int entries, struct uv__io_uring_params* params) {
--  return syscall(__NR_io_uring_setup, entries, params);
--}
--
--
--int uv__io_uring_enter(int fd,
--                       unsigned to_submit,
--                       unsigned min_complete,
--                       unsigned flags) {
--  /* io_uring_enter used to take a sigset_t but it's unused
--   * in newer kernels unless IORING_ENTER_EXT_ARG is set,
--   * in which case it takes a struct io_uring_getevents_arg.
--   */
--  return syscall(__NR_io_uring_enter,
--                 fd,
--                 to_submit,
--                 min_complete,
--                 flags,
--                 NULL,
--                 0L);
--}
--
--
--int uv__io_uring_register(int fd, unsigned opcode, void* arg, unsigned nargs) {
--  return syscall(__NR_io_uring_register, fd, opcode, arg, nargs);
--}
--
--
--static int uv__use_io_uring(void) {
--  /* Ternary: unknown=0, yes=1, no=-1 */
--  static _Atomic int use_io_uring;
--  char* val;
--  int use;
--
--  use = atomic_load_explicit(&use_io_uring, memory_order_relaxed);
--
--  if (use == 0) {
--    val = getenv("UV_USE_IO_URING");
--    use = val == NULL || atoi(val) ? 1 : -1;
--    atomic_store_explicit(&use_io_uring, use, memory_order_relaxed);
--  }
--
--  return use > 0;
--}
--
--
--static void uv__iou_init(int epollfd,
--                         struct uv__iou* iou,
--                         uint32_t entries,
--                         uint32_t flags) {
--  struct uv__io_uring_params params;
--  struct epoll_event e;
--  size_t cqlen;
--  size_t sqlen;
--  size_t maxlen;
--  size_t sqelen;
--  uint32_t i;
--  char* sq;
--  char* sqe;
--  int ringfd;
--
--  sq = MAP_FAILED;
--  sqe = MAP_FAILED;
--
--  if (!uv__use_io_uring())
--    return;
--
--  /* SQPOLL required CAP_SYS_NICE until linux v5.12 relaxed that requirement.
--   * Mostly academic because we check for a v5.13 kernel afterwards anyway.
--   */
--  memset(&params, 0, sizeof(params));
--  params.flags = flags;
--
--  if (flags & UV__IORING_SETUP_SQPOLL)
--    params.sq_thread_idle = 10;  /* milliseconds */
--
--  /* Kernel returns a file descriptor with O_CLOEXEC flag set. */
--  ringfd = uv__io_uring_setup(entries, &params);
--  if (ringfd == -1)
--    return;
--
--  /* IORING_FEAT_RSRC_TAGS is used to detect linux v5.13 but what we're
--   * actually detecting is whether IORING_OP_STATX works with SQPOLL.
--   */
--  if (!(params.features & UV__IORING_FEAT_RSRC_TAGS))
--    goto fail;
--
--  /* Implied by IORING_FEAT_RSRC_TAGS but checked explicitly anyway. */
--  if (!(params.features & UV__IORING_FEAT_SINGLE_MMAP))
--    goto fail;
--
--  /* Implied by IORING_FEAT_RSRC_TAGS but checked explicitly anyway. */
--  if (!(params.features & UV__IORING_FEAT_NODROP))
--    goto fail;
--
--  sqlen = params.sq_off.array + params.sq_entries * sizeof(uint32_t);
--  cqlen =
--      params.cq_off.cqes + params.cq_entries * sizeof(struct uv__io_uring_cqe);
--  maxlen = sqlen < cqlen ? cqlen : sqlen;
--  sqelen = params.sq_entries * sizeof(struct uv__io_uring_sqe);
--
--  sq = mmap(0,
--            maxlen,
--            PROT_READ | PROT_WRITE,
--            MAP_SHARED | MAP_POPULATE,
--            ringfd,
--            0);  /* IORING_OFF_SQ_RING */
--
--  sqe = mmap(0,
--             sqelen,
--             PROT_READ | PROT_WRITE,
--             MAP_SHARED | MAP_POPULATE,
--             ringfd,
--             0x10000000ull);  /* IORING_OFF_SQES */
--
--  if (sq == MAP_FAILED || sqe == MAP_FAILED)
--    goto fail;
--
--  if (flags & UV__IORING_SETUP_SQPOLL) {
--    /* Only interested in completion events. To get notified when
--     * the kernel pulls items from the submission ring, add POLLOUT.
--     */
--    memset(&e, 0, sizeof(e));
--    e.events = POLLIN;
--    e.data.fd = ringfd;
--
--    if (epoll_ctl(epollfd, EPOLL_CTL_ADD, ringfd, &e))
--      goto fail;
--  }
--
--  iou->sqhead = (uint32_t*) (sq + params.sq_off.head);
--  iou->sqtail = (uint32_t*) (sq + params.sq_off.tail);
--  iou->sqmask = *(uint32_t*) (sq + params.sq_off.ring_mask);
--  iou->sqarray = (uint32_t*) (sq + params.sq_off.array);
--  iou->sqflags = (uint32_t*) (sq + params.sq_off.flags);
--  iou->cqhead = (uint32_t*) (sq + params.cq_off.head);
--  iou->cqtail = (uint32_t*) (sq + params.cq_off.tail);
--  iou->cqmask = *(uint32_t*) (sq + params.cq_off.ring_mask);
--  iou->sq = sq;
--  iou->cqe = sq + params.cq_off.cqes;
--  iou->sqe = sqe;
--  iou->sqlen = sqlen;
--  iou->cqlen = cqlen;
--  iou->maxlen = maxlen;
--  iou->sqelen = sqelen;
--  iou->ringfd = ringfd;
--  iou->in_flight = 0;
--
--  for (i = 0; i <= iou->sqmask; i++)
--    iou->sqarray[i] = i;  /* Slot -> sqe identity mapping. */
--
--  return;
--
--fail:
--  if (sq != MAP_FAILED)
--    munmap(sq, maxlen);
--
--  if (sqe != MAP_FAILED)
--    munmap(sqe, sqelen);
--
--  uv__close(ringfd);
--}
--
--
--static void uv__iou_delete(struct uv__iou* iou) {
--  if (iou->ringfd != -1) {
--    munmap(iou->sq, iou->maxlen);
--    munmap(iou->sqe, iou->sqelen);
--    uv__close(iou->ringfd);
--    iou->ringfd = -1;
--  }
--}
--
--
- int uv__platform_loop_init(uv_loop_t* loop) {
--  uv__loop_internal_fields_t* lfields;
--
--  lfields = uv__get_internal_fields(loop);
--  lfields->ctl.ringfd = -1;
--  lfields->iou.ringfd = -1;
--
-   loop->inotify_watchers = NULL;
-   loop->inotify_fd = -1;
-   loop->backend_fd = epoll_create1(O_CLOEXEC);
-@@ -544,9 +214,6 @@ int uv__platform_loop_init(uv_loop_t* loop) {
-   if (loop->backend_fd == -1)
-     return UV__ERR(errno);
- 
--  uv__iou_init(loop->backend_fd, &lfields->iou, 64, UV__IORING_SETUP_SQPOLL);
--  uv__iou_init(loop->backend_fd, &lfields->ctl, 256, 0);
--
-   return 0;
- }
- 
-@@ -559,8 +226,6 @@ int uv__io_fork(uv_loop_t* loop) {
- 
-   uv__close(loop->backend_fd);
-   loop->backend_fd = -1;
--
--  /* TODO(bnoordhuis) Loses items from the submission and completion rings. */
-   uv__platform_loop_delete(loop);
- 
-   err = uv__platform_loop_init(loop);
-@@ -572,62 +237,42 @@ int uv__io_fork(uv_loop_t* loop) {
- 
- 
- void uv__platform_loop_delete(uv_loop_t* loop) {
--  uv__loop_internal_fields_t* lfields;
--
--  lfields = uv__get_internal_fields(loop);
--  uv__iou_delete(&lfields->ctl);
--  uv__iou_delete(&lfields->iou);
--
--  if (loop->inotify_fd != -1) {
--    uv__io_stop(loop, &loop->inotify_read_watcher, POLLIN);
--    uv__close(loop->inotify_fd);
--    loop->inotify_fd = -1;
--  }
-+  if (loop->inotify_fd == -1) return;
-+  uv__io_stop(loop, &loop->inotify_read_watcher, POLLIN);
-+  uv__close(loop->inotify_fd);
-+  loop->inotify_fd = -1;
- }
- 
- 
--struct uv__invalidate {
--  struct epoll_event (*prep)[256];
--  struct epoll_event* events;
--  int nfds;
--};
--
--
- void uv__platform_invalidate_fd(uv_loop_t* loop, int fd) {
--  uv__loop_internal_fields_t* lfields;
--  struct uv__invalidate* inv;
-+  struct epoll_event* events;
-   struct epoll_event dummy;
--  int i;
-+  uintptr_t i;
-+  uintptr_t nfds;
- 
--  lfields = uv__get_internal_fields(loop);
--  inv = lfields->inv;
-+  assert(loop->watchers != NULL);
-+  assert(fd >= 0);
- 
--  /* Invalidate events with same file descriptor */
--  if (inv != NULL)
--    for (i = 0; i < inv->nfds; i++)
--      if (inv->events[i].data.fd == fd)
--        inv->events[i].data.fd = -1;
-+  events = (struct epoll_event*) loop->watchers[loop->nwatchers];
-+  nfds = (uintptr_t) loop->watchers[loop->nwatchers + 1];
-+  if (events != NULL)
-+    /* Invalidate events with same file descriptor */
-+    for (i = 0; i < nfds; i++)
-+      if (events[i].data.fd == fd)
-+        events[i].data.fd = -1;
- 
-   /* Remove the file descriptor from the epoll.
-    * This avoids a problem where the same file description remains open
-    * in another process, causing repeated junk epoll events.
-    *
-    * We pass in a dummy epoll_event, to work around a bug in old kernels.
--   *
--   * Work around a bug in kernels 3.10 to 3.19 where passing a struct that
--   * has the EPOLLWAKEUP flag set generates spurious audit syslog warnings.
-    */
--  memset(&dummy, 0, sizeof(dummy));
--
--  if (inv == NULL) {
-+  if (loop->backend_fd >= 0) {
-+    /* Work around a bug in kernels 3.10 to 3.19 where passing a struct that
-+     * has the EPOLLWAKEUP flag set generates spurious audit syslog warnings.
-+     */
-+    memset(&dummy, 0, sizeof(dummy));
-     epoll_ctl(loop->backend_fd, EPOLL_CTL_DEL, fd, &dummy);
--  } else {
--    uv__epoll_ctl_prep(loop->backend_fd,
--                       &lfields->ctl,
--                       inv->prep,
--                       EPOLL_CTL_DEL,
--                       fd,
--                       &dummy);
-   }
- }
- 
-@@ -653,454 +298,27 @@ int uv__io_check_fd(uv_loop_t* loop, int fd) {
- }
- 
- 
--/* Caller must initialize SQE and call uv__iou_submit(). */
--static struct uv__io_uring_sqe* uv__iou_get_sqe(struct uv__iou* iou,
--                                                uv_loop_t* loop,
--                                                uv_fs_t* req) {
--  struct uv__io_uring_sqe* sqe;
--  uint32_t head;
--  uint32_t tail;
--  uint32_t mask;
--  uint32_t slot;
--
--  if (iou->ringfd == -1)
--    return NULL;
--
--  head = atomic_load_explicit((_Atomic uint32_t*) iou->sqhead,
--                              memory_order_acquire);
--  tail = *iou->sqtail;
--  mask = iou->sqmask;
--
--  if ((head & mask) == ((tail + 1) & mask))
--    return NULL;  /* No room in ring buffer. TODO(bnoordhuis) maybe flush it? */
--
--  slot = tail & mask;
--  sqe = iou->sqe;
--  sqe = &sqe[slot];
--  memset(sqe, 0, sizeof(*sqe));
--  sqe->user_data = (uintptr_t) req;
--
--  /* Pacify uv_cancel(). */
--  req->work_req.loop = loop;
--  req->work_req.work = NULL;
--  req->work_req.done = NULL;
--  QUEUE_INIT(&req->work_req.wq);
--
--  uv__req_register(loop, req);
--  iou->in_flight++;
--
--  return sqe;
--}
--
--
--static void uv__iou_submit(struct uv__iou* iou) {
--  uint32_t flags;
--
--  atomic_store_explicit((_Atomic uint32_t*) iou->sqtail,
--                        *iou->sqtail + 1,
--                        memory_order_release);
--
--  flags = atomic_load_explicit((_Atomic uint32_t*) iou->sqflags,
--                               memory_order_acquire);
--
--  if (flags & UV__IORING_SQ_NEED_WAKEUP)
--    if (uv__io_uring_enter(iou->ringfd, 0, 0, UV__IORING_ENTER_SQ_WAKEUP))
--      if (errno != EOWNERDEAD)  /* Kernel bug. Harmless, ignore. */
--        perror("libuv: io_uring_enter(wakeup)");  /* Can't happen. */
--}
--
--
--int uv__iou_fs_close(uv_loop_t* loop, uv_fs_t* req) {
--  struct uv__io_uring_sqe* sqe;
--  struct uv__iou* iou;
--
--  iou = &uv__get_internal_fields(loop)->iou;
--
--  sqe = uv__iou_get_sqe(iou, loop, req);
--  if (sqe == NULL)
--    return 0;
--
--  sqe->fd = req->file;
--  sqe->opcode = UV__IORING_OP_CLOSE;
--
--  uv__iou_submit(iou);
--
--  return 1;
--}
--
--
--int uv__iou_fs_fsync_or_fdatasync(uv_loop_t* loop,
--                                  uv_fs_t* req,
--                                  uint32_t fsync_flags) {
--  struct uv__io_uring_sqe* sqe;
--  struct uv__iou* iou;
--
--  iou = &uv__get_internal_fields(loop)->iou;
--
--  sqe = uv__iou_get_sqe(iou, loop, req);
--  if (sqe == NULL)
--    return 0;
--
--  /* Little known fact: setting seq->off and seq->len turns
--   * it into an asynchronous sync_file_range() operation.
--   */
--  sqe->fd = req->file;
--  sqe->fsync_flags = fsync_flags;
--  sqe->opcode = UV__IORING_OP_FSYNC;
--
--  uv__iou_submit(iou);
--
--  return 1;
--}
--
--
--int uv__iou_fs_open(uv_loop_t* loop, uv_fs_t* req) {
--  struct uv__io_uring_sqe* sqe;
--  struct uv__iou* iou;
--
--  iou = &uv__get_internal_fields(loop)->iou;
--
--  sqe = uv__iou_get_sqe(iou, loop, req);
--  if (sqe == NULL)
--    return 0;
--
--  sqe->addr = (uintptr_t) req->path;
--  sqe->fd = AT_FDCWD;
--  sqe->len = req->mode;
--  sqe->opcode = UV__IORING_OP_OPENAT;
--  sqe->open_flags = req->flags | O_CLOEXEC;
--
--  uv__iou_submit(iou);
--
--  return 1;
--}
--
--
--int uv__iou_fs_read_or_write(uv_loop_t* loop,
--                             uv_fs_t* req,
--                             int is_read) {
--  struct uv__io_uring_sqe* sqe;
--  struct uv__iou* iou;
--
--  /* For the moment, if iovcnt is greater than IOV_MAX, fallback to the
--   * threadpool. In the future we might take advantage of IOSQE_IO_LINK. */
--  if (req->nbufs > IOV_MAX)
--    return 0;
--
--  iou = &uv__get_internal_fields(loop)->iou;
--
--  sqe = uv__iou_get_sqe(iou, loop, req);
--  if (sqe == NULL)
--    return 0;
--
--  sqe->addr = (uintptr_t) req->bufs;
--  sqe->fd = req->file;
--  sqe->len = req->nbufs;
--  sqe->off = req->off < 0 ? -1 : req->off;
--  sqe->opcode = is_read ? UV__IORING_OP_READV : UV__IORING_OP_WRITEV;
--
--  uv__iou_submit(iou);
--
--  return 1;
--}
--
--
--int uv__iou_fs_statx(uv_loop_t* loop,
--                     uv_fs_t* req,
--                     int is_fstat,
--                     int is_lstat) {
--  struct uv__io_uring_sqe* sqe;
--  struct uv__statx* statxbuf;
--  struct uv__iou* iou;
--
--  statxbuf = uv__malloc(sizeof(*statxbuf));
--  if (statxbuf == NULL)
--    return 0;
--
--  iou = &uv__get_internal_fields(loop)->iou;
--
--  sqe = uv__iou_get_sqe(iou, loop, req);
--  if (sqe == NULL) {
--    uv__free(statxbuf);
--    return 0;
--  }
--
--  req->ptr = statxbuf;
--
--  sqe->addr = (uintptr_t) req->path;
--  sqe->addr2 = (uintptr_t) statxbuf;
--  sqe->fd = AT_FDCWD;
--  sqe->len = 0xFFF; /* STATX_BASIC_STATS + STATX_BTIME */
--  sqe->opcode = UV__IORING_OP_STATX;
--
--  if (is_fstat) {
--    sqe->addr = (uintptr_t) "";
--    sqe->fd = req->file;
--    sqe->statx_flags |= 0x1000; /* AT_EMPTY_PATH */
--  }
--
--  if (is_lstat)
--    sqe->statx_flags |= AT_SYMLINK_NOFOLLOW;
--
--  uv__iou_submit(iou);
--
--  return 1;
--}
--
--
--void uv__statx_to_stat(const struct uv__statx* statxbuf, uv_stat_t* buf) {
--  buf->st_dev = makedev(statxbuf->stx_dev_major, statxbuf->stx_dev_minor);
--  buf->st_mode = statxbuf->stx_mode;
--  buf->st_nlink = statxbuf->stx_nlink;
--  buf->st_uid = statxbuf->stx_uid;
--  buf->st_gid = statxbuf->stx_gid;
--  buf->st_rdev = makedev(statxbuf->stx_rdev_major, statxbuf->stx_rdev_minor);
--  buf->st_ino = statxbuf->stx_ino;
--  buf->st_size = statxbuf->stx_size;
--  buf->st_blksize = statxbuf->stx_blksize;
--  buf->st_blocks = statxbuf->stx_blocks;
--  buf->st_atim.tv_sec = statxbuf->stx_atime.tv_sec;
--  buf->st_atim.tv_nsec = statxbuf->stx_atime.tv_nsec;
--  buf->st_mtim.tv_sec = statxbuf->stx_mtime.tv_sec;
--  buf->st_mtim.tv_nsec = statxbuf->stx_mtime.tv_nsec;
--  buf->st_ctim.tv_sec = statxbuf->stx_ctime.tv_sec;
--  buf->st_ctim.tv_nsec = statxbuf->stx_ctime.tv_nsec;
--  buf->st_birthtim.tv_sec = statxbuf->stx_btime.tv_sec;
--  buf->st_birthtim.tv_nsec = statxbuf->stx_btime.tv_nsec;
--  buf->st_flags = 0;
--  buf->st_gen = 0;
--}
--
--
--static void uv__iou_fs_statx_post(uv_fs_t* req) {
--  struct uv__statx* statxbuf;
--  uv_stat_t* buf;
--
--  buf = &req->statbuf;
--  statxbuf = req->ptr;
--  req->ptr = NULL;
--
--  if (req->result == 0) {
--    uv__msan_unpoison(statxbuf, sizeof(*statxbuf));
--    uv__statx_to_stat(statxbuf, buf);
--    req->ptr = buf;
--  }
--
--  uv__free(statxbuf);
--}
--
--
--static void uv__poll_io_uring(uv_loop_t* loop, struct uv__iou* iou) {
--  struct uv__io_uring_cqe* cqe;
--  struct uv__io_uring_cqe* e;
--  uv_fs_t* req;
--  uint32_t head;
--  uint32_t tail;
--  uint32_t mask;
--  uint32_t i;
--  uint32_t flags;
--  int nevents;
--  int rc;
--
--  head = *iou->cqhead;
--  tail = atomic_load_explicit((_Atomic uint32_t*) iou->cqtail,
--                              memory_order_acquire);
--  mask = iou->cqmask;
--  cqe = iou->cqe;
--  nevents = 0;
--
--  for (i = head; i != tail; i++) {
--    e = &cqe[i & mask];
--
--    req = (uv_fs_t*) (uintptr_t) e->user_data;
--    assert(req->type == UV_FS);
--
--    uv__req_unregister(loop, req);
--    iou->in_flight--;
--
--    /* io_uring stores error codes as negative numbers, same as libuv. */
--    req->result = e->res;
--
--    switch (req->fs_type) {
--      case UV_FS_FSTAT:
--      case UV_FS_LSTAT:
--      case UV_FS_STAT:
--        uv__iou_fs_statx_post(req);
--        break;
--      default:  /* Squelch -Wswitch warnings. */
--        break;
--    }
--
--    uv__metrics_update_idle_time(loop);
--    req->cb(req);
--    nevents++;
--  }
--
--  atomic_store_explicit((_Atomic uint32_t*) iou->cqhead,
--                        tail,
--                        memory_order_release);
--
--  /* Check whether CQE's overflowed, if so enter the kernel to make them
--   * available. Don't grab them immediately but in the next loop iteration to
--   * avoid loop starvation. */
--  flags = atomic_load_explicit((_Atomic uint32_t*) iou->sqflags,
--                               memory_order_acquire);
--
--  if (flags & UV__IORING_SQ_CQ_OVERFLOW) {
--    do
--      rc = uv__io_uring_enter(iou->ringfd, 0, 0, UV__IORING_ENTER_GETEVENTS);
--    while (rc == -1 && errno == EINTR);
--
--    if (rc < 0)
--      perror("libuv: io_uring_enter(getevents)");  /* Can't happen. */
--  }
--
--  uv__metrics_inc_events(loop, nevents);
--  if (uv__get_internal_fields(loop)->current_timeout == 0)
--    uv__metrics_inc_events_waiting(loop, nevents);
--}
--
--
--static void uv__epoll_ctl_prep(int epollfd,
--                               struct uv__iou* ctl,
--                               struct epoll_event (*events)[256],
--                               int op,
--                               int fd,
--                               struct epoll_event* e) {
--  struct uv__io_uring_sqe* sqe;
--  struct epoll_event* pe;
--  uint32_t mask;
--  uint32_t slot;
--
--  if (ctl->ringfd == -1) {
--    if (!epoll_ctl(epollfd, op, fd, e))
--      return;
--
--    if (op == EPOLL_CTL_DEL)
--      return;  /* Ignore errors, may be racing with another thread. */
--
--    if (op != EPOLL_CTL_ADD)
--      abort();
--
--    if (errno != EEXIST)
--      abort();
--
--    /* File descriptor that's been watched before, update event mask. */
--    if (!epoll_ctl(epollfd, EPOLL_CTL_MOD, fd, e))
--      return;
--
--    abort();
--  } else {
--    mask = ctl->sqmask;
--    slot = (*ctl->sqtail)++ & mask;
--
--    pe = &(*events)[slot];
--    *pe = *e;
--
--    sqe = ctl->sqe;
--    sqe = &sqe[slot];
--
--    memset(sqe, 0, sizeof(*sqe));
--    sqe->addr = (uintptr_t) pe;
--    sqe->fd = epollfd;
--    sqe->len = op;
--    sqe->off = fd;
--    sqe->opcode = UV__IORING_OP_EPOLL_CTL;
--    sqe->user_data = op | slot << 2 | (int64_t) fd << 32;
--
--    if ((*ctl->sqhead & mask) == (*ctl->sqtail & mask))
--      uv__epoll_ctl_flush(epollfd, ctl, events);
--  }
--}
--
--
--static void uv__epoll_ctl_flush(int epollfd,
--                                struct uv__iou* ctl,
--                                struct epoll_event (*events)[256]) {
--  struct epoll_event oldevents[256];
--  struct uv__io_uring_cqe* cqe;
--  uint32_t oldslot;
--  uint32_t slot;
--  uint32_t n;
--  int fd;
--  int op;
--  int rc;
--
--  STATIC_ASSERT(sizeof(oldevents) == sizeof(*events));
--  assert(ctl->ringfd != -1);
--  assert(*ctl->sqhead != *ctl->sqtail);
--
--  n = *ctl->sqtail - *ctl->sqhead;
--  do
--    rc = uv__io_uring_enter(ctl->ringfd, n, n, UV__IORING_ENTER_GETEVENTS);
--  while (rc == -1 && errno == EINTR);
--
--  if (rc < 0)
--    perror("libuv: io_uring_enter(getevents)");  /* Can't happen. */
--
--  if (rc != (int) n)
--    abort();
--
--  assert(*ctl->sqhead == *ctl->sqtail);
--
--  memcpy(oldevents, *events, sizeof(*events));
--
--  /* Failed submissions are either EPOLL_CTL_DEL commands for file descriptors
--   * that have been closed, or EPOLL_CTL_ADD commands for file descriptors
--   * that we are already watching. Ignore the former and retry the latter
--   * with EPOLL_CTL_MOD.
--   */
--  while (*ctl->cqhead != *ctl->cqtail) {
--    slot = (*ctl->cqhead)++ & ctl->cqmask;
--
--    cqe = ctl->cqe;
--    cqe = &cqe[slot];
--
--    if (cqe->res == 0)
--      continue;
--
--    fd = cqe->user_data >> 32;
--    op = 3 & cqe->user_data;
--    oldslot = 255 & (cqe->user_data >> 2);
--
--    if (op == EPOLL_CTL_DEL)
--      continue;
--
--    if (op != EPOLL_CTL_ADD)
--      abort();
--
--    if (cqe->res != -EEXIST)
--      abort();
--
--    uv__epoll_ctl_prep(epollfd,
--                       ctl,
--                       events,
--                       EPOLL_CTL_MOD,
--                       fd,
--                       &oldevents[oldslot]);
--  }
--}
--
--
- void uv__io_poll(uv_loop_t* loop, int timeout) {
--  uv__loop_internal_fields_t* lfields;
-+  /* A bug in kernels < 2.6.37 makes timeouts larger than ~30 minutes
-+   * effectively infinite on 32 bits architectures.  To avoid blocking
-+   * indefinitely, we cap the timeout and poll again if necessary.
-+   *
-+   * Note that "30 minutes" is a simplification because it depends on
-+   * the value of CONFIG_HZ.  The magic constant assumes CONFIG_HZ=1200,
-+   * that being the largest value I have seen in the wild (and only once.)
-+   */
-+  static const int max_safe_timeout = 1789569;
-   struct epoll_event events[1024];
--  struct epoll_event prep[256];
--  struct uv__invalidate inv;
-   struct epoll_event* pe;
-   struct epoll_event e;
--  struct uv__iou* ctl;
--  struct uv__iou* iou;
-   int real_timeout;
-   QUEUE* q;
-   uv__io_t* w;
-   sigset_t* sigmask;
-   sigset_t sigset;
-   uint64_t base;
--  int have_iou_events;
-   int have_signals;
-   int nevents;
--  int epollfd;
-   int count;
-   int nfds;
-   int fd;
-@@ -1109,9 +327,47 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-   int user_timeout;
-   int reset_timeout;
- 
--  lfields = uv__get_internal_fields(loop);
--  ctl = &lfields->ctl;
--  iou = &lfields->iou;
-+  if (loop->nfds == 0) {
-+    assert(QUEUE_EMPTY(&loop->watcher_queue));
-+    return;
-+  }
-+
-+  memset(&e, 0, sizeof(e));
-+
-+  while (!QUEUE_EMPTY(&loop->watcher_queue)) {
-+    q = QUEUE_HEAD(&loop->watcher_queue);
-+    QUEUE_REMOVE(q);
-+    QUEUE_INIT(q);
-+
-+    w = QUEUE_DATA(q, uv__io_t, watcher_queue);
-+    assert(w->pevents != 0);
-+    assert(w->fd >= 0);
-+    assert(w->fd < (int) loop->nwatchers);
-+
-+    e.events = w->pevents;
-+    e.data.fd = w->fd;
-+
-+    if (w->events == 0)
-+      op = EPOLL_CTL_ADD;
-+    else
-+      op = EPOLL_CTL_MOD;
-+
-+    /* XXX Future optimization: do EPOLL_CTL_MOD lazily if we stop watching
-+     * events, skip the syscall and squelch the events after epoll_wait().
-+     */
-+    if (epoll_ctl(loop->backend_fd, op, w->fd, &e)) {
-+      if (errno != EEXIST)
-+        abort();
-+
-+      assert(op == EPOLL_CTL_ADD);
-+
-+      /* We've reactivated a file descriptor that's been watched before. */
-+      if (epoll_ctl(loop->backend_fd, EPOLL_CTL_MOD, w->fd, &e))
-+        abort();
-+    }
-+
-+    w->events = w->pevents;
-+  }
- 
-   sigmask = NULL;
-   if (loop->flags & UV_LOOP_BLOCK_SIGPROF) {
-@@ -1125,7 +381,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-   count = 48; /* Benchmarks suggest this gives the best throughput. */
-   real_timeout = timeout;
- 
--  if (lfields->flags & UV_METRICS_IDLE_TIME) {
-+  if (uv__get_internal_fields(loop)->flags & UV_METRICS_IDLE_TIME) {
-     reset_timeout = 1;
-     user_timeout = timeout;
-     timeout = 0;
-@@ -1134,56 +390,24 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-     user_timeout = 0;
-   }
- 
--  epollfd = loop->backend_fd;
--
--  memset(&e, 0, sizeof(e));
--
--  while (!QUEUE_EMPTY(&loop->watcher_queue)) {
--    q = QUEUE_HEAD(&loop->watcher_queue);
--    w = QUEUE_DATA(q, uv__io_t, watcher_queue);
--    QUEUE_REMOVE(q);
--    QUEUE_INIT(q);
--
--    op = EPOLL_CTL_MOD;
--    if (w->events == 0)
--      op = EPOLL_CTL_ADD;
--
--    w->events = w->pevents;
--    e.events = w->pevents;
--    e.data.fd = w->fd;
--
--    uv__epoll_ctl_prep(epollfd, ctl, &prep, op, w->fd, &e);
--  }
--
--  inv.events = events;
--  inv.prep = &prep;
--  inv.nfds = -1;
--
-   for (;;) {
--    if (loop->nfds == 0)
--      if (iou->in_flight == 0)
--        break;
--
--    /* All event mask mutations should be visible to the kernel before
--     * we enter epoll_pwait().
--     */
--    if (ctl->ringfd != -1)
--      while (*ctl->sqhead != *ctl->sqtail)
--        uv__epoll_ctl_flush(epollfd, ctl, &prep);
--
-     /* Only need to set the provider_entry_time if timeout != 0. The function
-      * will return early if the loop isn't configured with UV_METRICS_IDLE_TIME.
-      */
-     if (timeout != 0)
-       uv__metrics_set_provider_entry_time(loop);
- 
--    /* Store the current timeout in a location that's globally accessible so
--     * other locations like uv__work_done() can determine whether the queue
--     * of events in the callback were waiting when poll was called.
-+    /* See the comment for max_safe_timeout for an explanation of why
-+     * this is necessary.  Executive summary: kernel bug workaround.
-      */
--    lfields->current_timeout = timeout;
-+    if (sizeof(int32_t) == sizeof(long) && timeout >= max_safe_timeout)
-+      timeout = max_safe_timeout;
- 
--    nfds = epoll_pwait(epollfd, events, ARRAY_SIZE(events), timeout, sigmask);
-+    nfds = epoll_pwait(loop->backend_fd,
-+                       events,
-+                       ARRAY_SIZE(events),
-+                       timeout,
-+                       sigmask);
- 
-     /* Update loop->time unconditionally. It's tempting to skip the update when
-      * timeout == 0 (i.e. non-blocking poll) but there is no guarantee that the
-@@ -1203,7 +427,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-         continue;
- 
-       if (timeout == 0)
--        break;
-+        return;
- 
-       /* We may have been inside the system call for longer than |timeout|
-        * milliseconds so we need to update the timestamp to avoid drift.
-@@ -1224,18 +448,27 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-         continue;
- 
-       if (timeout == 0)
--        break;
-+        return;
- 
-       /* Interrupted by a signal. Update timeout and poll again. */
-       goto update_timeout;
-     }
- 
--    have_iou_events = 0;
-     have_signals = 0;
-     nevents = 0;
- 
--    inv.nfds = nfds;
--    lfields->inv = &inv;
-+    {
-+      /* Squelch a -Waddress-of-packed-member warning with gcc >= 9. */
-+      union {
-+        struct epoll_event* events;
-+        uv__io_t* watchers;
-+      } x;
-+
-+      x.events = events;
-+      assert(loop->watchers != NULL);
-+      loop->watchers[loop->nwatchers] = x.watchers;
-+      loop->watchers[loop->nwatchers + 1] = (void*) (uintptr_t) nfds;
-+    }
- 
-     for (i = 0; i < nfds; i++) {
-       pe = events + i;
-@@ -1245,12 +478,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-       if (fd == -1)
-         continue;
- 
--      if (fd == iou->ringfd) {
--        uv__poll_io_uring(loop, iou);
--        have_iou_events = 1;
--        continue;
--      }
--
-       assert(fd >= 0);
-       assert((unsigned) fd < loop->nwatchers);
- 
-@@ -1262,7 +489,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-          * Ignore all errors because we may be racing with another thread
-          * when the file descriptor is closed.
-          */
--        uv__epoll_ctl_prep(epollfd, ctl, &prep, EPOLL_CTL_DEL, fd, pe);
-+        epoll_ctl(loop->backend_fd, EPOLL_CTL_DEL, fd, pe);
-         continue;
-       }
- 
-@@ -1319,13 +546,11 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-       loop->signal_io_watcher.cb(loop, &loop->signal_io_watcher, POLLIN);
-     }
- 
--    lfields->inv = NULL;
--
--    if (have_iou_events != 0)
--      break;  /* Event loop should cycle now so don't poll again. */
-+    loop->watchers[loop->nwatchers] = NULL;
-+    loop->watchers[loop->nwatchers + 1] = NULL;
- 
-     if (have_signals != 0)
--      break;  /* Event loop should cycle now so don't poll again. */
-+      return;  /* Event loop should cycle now so don't poll again. */
- 
-     if (nevents != 0) {
-       if (nfds == ARRAY_SIZE(events) && --count != 0) {
-@@ -1333,11 +558,11 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-         timeout = 0;
-         continue;
-       }
--      break;
-+      return;
-     }
- 
-     if (timeout == 0)
--      break;
-+      return;
- 
-     if (timeout == -1)
-       continue;
-@@ -1347,14 +572,10 @@ update_timeout:
- 
-     real_timeout -= (loop->time - base);
-     if (real_timeout <= 0)
--      break;
-+      return;
- 
-     timeout = real_timeout;
-   }
--
--  if (ctl->ringfd != -1)
--    while (*ctl->sqhead != *ctl->sqtail)
--      uv__epoll_ctl_flush(epollfd, ctl, &prep);
- }
- 
- uint64_t uv__hrtime(uv_clocktype_t type) {
-@@ -1867,7 +1088,7 @@ static uint64_t uv__read_uint64(const char* filename) {
-   if (0 == uv__slurp(filename, buf, sizeof(buf)))
-     if (1 != sscanf(buf, "%" PRIu64, &rc))
-       if (0 == strcmp(buf, "max\n"))
--        rc = UINT64_MAX;
-+        rc = ~0ull;
- 
-   return rc;
- }
-@@ -1903,7 +1124,6 @@ static void uv__get_cgroup1_memory_limits(char buf[static 1024], uint64_t* high,
-   char filename[4097];
-   char* p;
-   int n;
--  uint64_t cgroup1_max;
- 
-   /* Find out where the controller is mounted. */
-   p = uv__cgroup1_find_memory_controller(buf, &n);
-@@ -1920,22 +1140,12 @@ static void uv__get_cgroup1_memory_limits(char buf[static 1024], uint64_t* high,
-      * as indicated by uv__read_uint64 returning 0.
-      */
-      if (*high != 0 && *max != 0)
--       goto update_limits;
-+       return;
-   }
- 
-   /* Fall back to the limits of the global memory controller. */
-   *high = uv__read_uint64("/sys/fs/cgroup/memory/memory.soft_limit_in_bytes");
-   *max = uv__read_uint64("/sys/fs/cgroup/memory/memory.limit_in_bytes");
--
--  /* uv__read_uint64 detects cgroup2's "max", so we need to separately detect
--   * cgroup1's maximum value (which is derived from LONG_MAX and PAGE_SIZE).
--   */
--update_limits:
--  cgroup1_max = LONG_MAX & ~(sysconf(_SC_PAGESIZE) - 1);
--  if (*high == cgroup1_max)
--    *high = UINT64_MAX;
--  if (*max == cgroup1_max)
--    *max = UINT64_MAX;
- }
- 
- static void uv__get_cgroup2_memory_limits(char buf[static 1024], uint64_t* high,
-diff --git a/deps/uv/src/unix/os390.c b/deps/uv/src/unix/os390.c
-index a87c2d77faf..3954b2c2753 100644
---- a/deps/uv/src/unix/os390.c
-+++ b/deps/uv/src/unix/os390.c
-@@ -808,7 +808,6 @@ static int os390_message_queue_handler(uv__os390_epoll* ep) {
- 
- void uv__io_poll(uv_loop_t* loop, int timeout) {
-   static const int max_safe_timeout = 1789569;
--  uv__loop_internal_fields_t* lfields;
-   struct epoll_event events[1024];
-   struct epoll_event* pe;
-   struct epoll_event e;
-@@ -831,8 +830,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-     return;
-   }
- 
--  lfields = uv__get_internal_fields(loop);
--
-   while (!QUEUE_EMPTY(&loop->watcher_queue)) {
-     uv_stream_t* stream;
- 
-@@ -880,7 +877,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-   int nevents = 0;
-   have_signals = 0;
- 
--  if (lfields->flags & UV_METRICS_IDLE_TIME) {
-+  if (uv__get_internal_fields(loop)->flags & UV_METRICS_IDLE_TIME) {
-     reset_timeout = 1;
-     user_timeout = timeout;
-     timeout = 0;
-@@ -899,12 +896,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-     if (sizeof(int32_t) == sizeof(long) && timeout >= max_safe_timeout)
-       timeout = max_safe_timeout;
- 
--    /* Store the current timeout in a location that's globally accessible so
--     * other locations like uv__work_done() can determine whether the queue
--     * of events in the callback were waiting when poll was called.
--     */
--    lfields->current_timeout = timeout;
--
-     nfds = epoll_wait(loop->ep, events,
-                       ARRAY_SIZE(events), timeout);
- 
-diff --git a/deps/uv/src/unix/posix-poll.c b/deps/uv/src/unix/posix-poll.c
-index 7e7de86845d..711780ece8d 100644
---- a/deps/uv/src/unix/posix-poll.c
-+++ b/deps/uv/src/unix/posix-poll.c
-@@ -132,7 +132,6 @@ static void uv__pollfds_del(uv_loop_t* loop, int fd) {
- 
- 
- void uv__io_poll(uv_loop_t* loop, int timeout) {
--  uv__loop_internal_fields_t* lfields;
-   sigset_t* pset;
-   sigset_t set;
-   uint64_t time_base;
-@@ -153,8 +152,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-     return;
-   }
- 
--  lfields = uv__get_internal_fields(loop);
--
-   /* Take queued watchers and add their fds to our poll fds array.  */
-   while (!QUEUE_EMPTY(&loop->watcher_queue)) {
-     q = QUEUE_HEAD(&loop->watcher_queue);
-@@ -182,7 +179,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-   assert(timeout >= -1);
-   time_base = loop->time;
- 
--  if (lfields->flags & UV_METRICS_IDLE_TIME) {
-+  if (uv__get_internal_fields(loop)->flags & UV_METRICS_IDLE_TIME) {
-     reset_timeout = 1;
-     user_timeout = timeout;
-     timeout = 0;
-@@ -201,12 +198,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) {
-     if (timeout != 0)
-       uv__metrics_set_provider_entry_time(loop);
- 
--    /* Store the current timeout in a location that's globally accessible so
--     * other locations like uv__work_done() can determine whether the queue
--     * of events in the callback were waiting when poll was called.
--     */
--    lfields->current_timeout = timeout;
--
-     if (pset != NULL)
-       if (pthread_sigmask(SIG_BLOCK, pset, NULL))
-         abort();
-diff --git a/deps/uv/src/uv-common.h b/deps/uv/src/uv-common.h
-index decde5362c8..2720121addc 100644
---- a/deps/uv/src/uv-common.h
-+++ b/deps/uv/src/uv-common.h
-@@ -396,37 +396,9 @@ struct uv__loop_metrics_s {
- void uv__metrics_update_idle_time(uv_loop_t* loop);
- void uv__metrics_set_provider_entry_time(uv_loop_t* loop);
- 
--#ifdef __linux__
--struct uv__iou {
--  uint32_t* sqhead;
--  uint32_t* sqtail;
--  uint32_t* sqarray;
--  uint32_t sqmask;
--  uint32_t* sqflags;
--  uint32_t* cqhead;
--  uint32_t* cqtail;
--  uint32_t cqmask;
--  void* sq;   /* pointer to munmap() on event loop teardown */
--  void* cqe;  /* pointer to array of struct uv__io_uring_cqe */
--  void* sqe;  /* pointer to array of struct uv__io_uring_sqe */
--  size_t sqlen;
--  size_t cqlen;
--  size_t maxlen;
--  size_t sqelen;
--  int ringfd;
--  uint32_t in_flight;
--};
--#endif  /* __linux__ */
--
- struct uv__loop_internal_fields_s {
-   unsigned int flags;
-   uv__loop_metrics_t loop_metrics;
--  int current_timeout;
--#ifdef __linux__
--  struct uv__iou ctl;
--  struct uv__iou iou;
--  void* inv;  /* used by uv__platform_invalidate_fd() */
--#endif  /* __linux__ */
- };
- 
- #endif /* UV_COMMON_H_ */
-diff --git a/deps/uv/src/win/core.c b/deps/uv/src/win/core.c
-index 9a3be58849a..e4041ec86a6 100644
---- a/deps/uv/src/win/core.c
-+++ b/deps/uv/src/win/core.c
-@@ -424,7 +424,6 @@ int uv_backend_timeout(const uv_loop_t* loop) {
- 
- 
- static void uv__poll_wine(uv_loop_t* loop, DWORD timeout) {
--  uv__loop_internal_fields_t* lfields;
-   DWORD bytes;
-   ULONG_PTR key;
-   OVERLAPPED* overlapped;
-@@ -434,10 +433,9 @@ static void uv__poll_wine(uv_loop_t* loop, DWORD timeout) {
-   uint64_t user_timeout;
-   int reset_timeout;
- 
--  lfields = uv__get_internal_fields(loop);
-   timeout_time = loop->time + timeout;
- 
--  if (lfields->flags & UV_METRICS_IDLE_TIME) {
-+  if (uv__get_internal_fields(loop)->flags & UV_METRICS_IDLE_TIME) {
-     reset_timeout = 1;
-     user_timeout = timeout;
-     timeout = 0;
-@@ -452,12 +450,6 @@ static void uv__poll_wine(uv_loop_t* loop, DWORD timeout) {
-     if (timeout != 0)
-       uv__metrics_set_provider_entry_time(loop);
- 
--    /* Store the current timeout in a location that's globally accessible so
--     * other locations like uv__work_done() can determine whether the queue
--     * of events in the callback were waiting when poll was called.
--     */
--    lfields->current_timeout = timeout;
--
-     GetQueuedCompletionStatus(loop->iocp,
-                               &bytes,
-                               &key,
-@@ -515,7 +507,6 @@ static void uv__poll_wine(uv_loop_t* loop, DWORD timeout) {
- 
- 
- static void uv__poll(uv_loop_t* loop, DWORD timeout) {
--  uv__loop_internal_fields_t* lfields;
-   BOOL success;
-   uv_req_t* req;
-   OVERLAPPED_ENTRY overlappeds[128];
-@@ -527,10 +518,9 @@ static void uv__poll(uv_loop_t* loop, DWORD timeout) {
-   uint64_t actual_timeout;
-   int reset_timeout;
- 
--  lfields = uv__get_internal_fields(loop);
-   timeout_time = loop->time + timeout;
- 
--  if (lfields->flags & UV_METRICS_IDLE_TIME) {
-+  if (uv__get_internal_fields(loop)->flags & UV_METRICS_IDLE_TIME) {
-     reset_timeout = 1;
-     user_timeout = timeout;
-     timeout = 0;
-@@ -547,12 +537,6 @@ static void uv__poll(uv_loop_t* loop, DWORD timeout) {
-     if (timeout != 0)
-       uv__metrics_set_provider_entry_time(loop);
- 
--    /* Store the current timeout in a location that's globally accessible so
--     * other locations like uv__work_done() can determine whether the queue
--     * of events in the callback were waiting when poll was called.
--     */
--    lfields->current_timeout = timeout;
--
-     success = pGetQueuedCompletionStatusEx(loop->iocp,
-                                            overlappeds,
-                                            ARRAY_SIZE(overlappeds),
diff --git a/meta-oe/recipes-devtools/nodejs/nodejs_20.11.0.bb b/meta-oe/recipes-devtools/nodejs/nodejs_20.11.1.bb
similarity index 96%
rename from meta-oe/recipes-devtools/nodejs/nodejs_20.11.0.bb
rename to meta-oe/recipes-devtools/nodejs/nodejs_20.11.1.bb
index cc5135478..e2e2f5ae8 100644
--- a/meta-oe/recipes-devtools/nodejs/nodejs_20.11.0.bb
+++ b/meta-oe/recipes-devtools/nodejs/nodejs_20.11.1.bb
@@ -24,8 +24,7 @@  SRC_URI = "http://nodejs.org/dist/v${PV}/node-v${PV}.tar.xz \
            file://0004-v8-don-t-override-ARM-CFLAGS.patch \
            file://system-c-ares.patch \
            file://0001-liftoff-Correct-function-signatures.patch \
-           file://0001-Revert-io_uring-changes-from-libuv-1.46.0.patch \
-           file://0002-Revert-io_uring-changes-from-libuv-1.45.0.patch \
+           file://0001-deps-disable-io_uring-support-in-libuv.patch \
            file://run-ptest \
            "
 
@@ -38,7 +37,7 @@  SRC_URI:append:toolchain-clang:x86 = " \
 SRC_URI:append:toolchain-clang:powerpc64le = " \
            file://0001-ppc64-Do-not-use-mminimal-toc-with-clang.patch \
            "
-SRC_URI[sha256sum] = "31807ebeeeb049c53f1765e4a95aed69476a4b696dd100cb539ab668d7950b40"
+SRC_URI[sha256sum] = "77813edbf3f7f16d2d35d3353443dee4e61d5ee84d9e3138c7538a3c0ca5209e"
 
 S = "${WORKDIR}/node-v${PV}"