From patchwork Tue Mar 5 02:24:33 2024 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Changqing Li X-Patchwork-Id: 40464 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from aws-us-west-2-korg-lkml-1.web.codeaurora.org (localhost.localdomain [127.0.0.1]) by smtp.lore.kernel.org (Postfix) with ESMTP id C20A4C48BF6 for ; Tue, 5 Mar 2024 02:24:41 +0000 (UTC) Received: from mx0b-0064b401.pphosted.com (mx0b-0064b401.pphosted.com [205.220.178.238]) by mx.groups.io with SMTP id smtpd.web10.12103.1709605479811685178 for ; Mon, 04 Mar 2024 18:24:41 -0800 Authentication-Results: mx.groups.io; dkim=pass header.i=@windriver.com header.s=PPS06212021 header.b=rXjnNxjW; spf=permerror, err=parse error for token &{10 18 %{ir}.%{v}.%{d}.spf.has.pphosted.com}: invalid domain name (domain: windriver.com, ip: 205.220.178.238, mailfrom: prvs=37941debd1=changqing.li@windriver.com) Received: from pps.filterd (m0250812.ppops.net [127.0.0.1]) by mx0a-0064b401.pphosted.com (8.17.1.24/8.17.1.24) with ESMTP id 424MVoKG003131 for ; Tue, 5 Mar 2024 02:24:37 GMT DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=windriver.com; h=from:to:subject:date:message-id:mime-version :content-transfer-encoding:content-type; s=PPS06212021; bh=tIP2P zEaP1LIRffR3jrnkdWsTWukHONc71/dvvQeBNM=; b=rXjnNxjWUjEq/7VlHIcBD L2QGg2ZzuesQDQoFbOkHrd81ZmTOJfpoboUxnTqcyGq3Bc2voEqibnC5LD/IpSjO 1Go4TLCPnmXBltVIr4qAwbeg4JzbL2aIkqhVxBRwlQ+QPoWgbRipj/DS5pEbyodn 0uiXqcUsQlDtVmZTZIt30Tdss/xTdR2dmNOQKlUdcJjo1Y4smGHrqdJ4D28s+xH6 +FtnkVufuD4almuAR+AurFollAeQFX95ZptxoP2O2s8o2pqSP0tWPym0nW8Z1eLh h5SEnRDB1POPEnAmIrvfdS0Ukr986xskLnplSVpRhvayq6/+EoZHB286u8FuUkEp Q== Received: from ala-exchng01.corp.ad.wrs.com (ala-exchng01.wrs.com [147.11.82.252]) by mx0a-0064b401.pphosted.com (PPS) with ESMTPS id 3wku36aavc-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128 verify=NOT) for ; Tue, 05 Mar 2024 02:24:36 +0000 (GMT) Received: from ala-exchng01.corp.ad.wrs.com (147.11.82.252) by ala-exchng01.corp.ad.wrs.com (147.11.82.252) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256) id 15.1.2507.35; Mon, 4 Mar 2024 18:24:35 -0800 Received: from pek-lpg-core2.wrs.com (128.224.153.41) by ala-exchng01.corp.ad.wrs.com (147.11.82.252) with Microsoft SMTP Server id 15.1.2507.35 via Frontend Transport; Mon, 4 Mar 2024 18:24:34 -0800 From: To: Subject: [meta-oe][PATCH] nodejs: upgrade 20.11.0 -> 20.11.1 Date: Tue, 5 Mar 2024 10:24:33 +0800 Message-ID: <20240305022433.2792409-1-changqing.li@windriver.com> X-Mailer: git-send-email 2.25.1 MIME-Version: 1.0 X-Proofpoint-ORIG-GUID: QCMr0hBJpfyRifHa-ohZ2vgrKm60psSj X-Proofpoint-GUID: QCMr0hBJpfyRifHa-ohZ2vgrKm60psSj X-Proofpoint-Virus-Version: vendor=baseguard engine=ICAP:2.0.272,Aquarius:18.0.1011,Hydra:6.0.619,FMLib:17.11.176.26 definitions=2024-03-04_20,2024-03-04_01,2023-05-22_02 X-Proofpoint-Spam-Details: rule=outbound_notspam policy=outbound score=0 clxscore=1015 spamscore=0 phishscore=0 priorityscore=1501 malwarescore=0 bulkscore=0 adultscore=0 mlxlogscore=999 suspectscore=0 impostorscore=0 mlxscore=0 lowpriorityscore=0 classifier=spam adjust=0 reason=mlx scancount=1 engine=8.19.0-2402120000 definitions=main-2403050016 List-Id: X-Webhook-Received: from li982-79.members.linode.com [45.33.32.79] by aws-us-west-2-korg-lkml-1.web.codeaurora.org with HTTPS for ; Tue, 05 Mar 2024 02:24:41 -0000 X-Groupsio-URL: https://lists.openembedded.org/g/openembedded-devel/message/109110 From: Changqing Li Drop patches for revert io_uring support in libuv: 0001-Revert-io_uring-changes-from-libuv-1.46.0.patch 0002-Revert-io_uring-changes-from-libuv-1.45.0.patch Change to just always disable io_uring in libuv, in this way, we don't have to pick out io_uring related changes and revert them when internal libuv upgraded. Signed-off-by: Changqing Li --- ...t-io_uring-changes-from-libuv-1.46.0.patch | 3414 ----------------- ...ps-disable-io_uring-support-in-libuv.patch | 35 + ...t-io_uring-changes-from-libuv-1.45.0.patch | 1803 --------- .../{nodejs_20.11.0.bb => nodejs_20.11.1.bb} | 5 +- 4 files changed, 37 insertions(+), 5220 deletions(-) delete mode 100644 meta-oe/recipes-devtools/nodejs/nodejs/0001-Revert-io_uring-changes-from-libuv-1.46.0.patch create mode 100644 meta-oe/recipes-devtools/nodejs/nodejs/0001-deps-disable-io_uring-support-in-libuv.patch delete mode 100644 meta-oe/recipes-devtools/nodejs/nodejs/0002-Revert-io_uring-changes-from-libuv-1.45.0.patch rename meta-oe/recipes-devtools/nodejs/{nodejs_20.11.0.bb => nodejs_20.11.1.bb} (96%) diff --git a/meta-oe/recipes-devtools/nodejs/nodejs/0001-Revert-io_uring-changes-from-libuv-1.46.0.patch b/meta-oe/recipes-devtools/nodejs/nodejs/0001-Revert-io_uring-changes-from-libuv-1.46.0.patch deleted file mode 100644 index d98463d0d..000000000 --- a/meta-oe/recipes-devtools/nodejs/nodejs/0001-Revert-io_uring-changes-from-libuv-1.46.0.patch +++ /dev/null @@ -1,3414 +0,0 @@ -From 937dfa70104b7401a7570a98cf6e0a534e250557 Mon Sep 17 00:00:00 2001 -From: Martin Jansa -Date: Wed, 18 Oct 2023 22:43:55 +0200 -Subject: [PATCH] Revert io_uring changes from libuv-1.46.0 - -Included in nodejs-20.4.0 with the libuv upgrade to 1.46.0 in: -https://github.com/nodejs/node/commit/6199e1946c9abd59bd71a61870a4f6c85e100b18 - -Revert "linux: work around io_uring IORING_OP_CLOSE bug (#4059)" -This reverts commit 1752791c9ea89dbf54e2a20a9d9f899119a2d179. -Revert "linux: fs_read to use io_uring if iovcnt > IOV_MAX (#4023)" -This reverts commit e7b96331703e929e75d93c574573c9736e34b0c0. -Revert "linux: add some more iouring backed fs ops (#4012)" -This reverts commit 962b8e626ceaaf296eecab1d008e8b70dc6da5e7. -Revert "android: disable io_uring support (#4016)" -This reverts commit 281e6185cc9e77c582ff6ca9e4c00d57e6b90d95. -Revert "unix,win: replace QUEUE with struct uv__queue (#4022)" -This reverts commit 1b01b786c0180d29f07dccbb98001a2b3148828a. -Revert "fs: use WTF-8 on Windows (#2970)" -This reverts commit 8f32a14afaaa47514a7d28e1e069a8329e2dd939. - -Dropped deps/uv/docs deps/uv/test changes as these dirs aren't included -in nodejs tarballs. - -Signed-off-by: Martin Jansa ---- -Upstream-Status: Inappropriate [OE specific] - - deps/uv/include/uv.h | 10 +- - deps/uv/include/uv/darwin.h | 6 +- - deps/uv/include/uv/linux.h | 2 +- - deps/uv/include/uv/threadpool.h | 2 +- - deps/uv/include/uv/unix.h | 44 ++-- - deps/uv/include/uv/win.h | 4 +- - deps/uv/src/queue.h | 154 ++++++----- - deps/uv/src/threadpool.c | 72 +++--- - deps/uv/src/unix/aix.c | 14 +- - deps/uv/src/unix/async.c | 52 ++-- - deps/uv/src/unix/core.c | 53 ++-- - deps/uv/src/unix/fs.c | 41 +-- - deps/uv/src/unix/fsevents.c | 84 +++--- - deps/uv/src/unix/internal.h | 11 - - deps/uv/src/unix/kqueue.c | 18 +- - deps/uv/src/unix/linux.c | 257 +++---------------- - deps/uv/src/unix/loop-watcher.c | 20 +- - deps/uv/src/unix/loop.c | 28 +- - deps/uv/src/unix/os390-syscalls.c | 24 +- - deps/uv/src/unix/os390-syscalls.h | 2 +- - deps/uv/src/unix/os390.c | 14 +- - deps/uv/src/unix/pipe.c | 2 +- - deps/uv/src/unix/posix-poll.c | 14 +- - deps/uv/src/unix/process.c | 38 +-- - deps/uv/src/unix/signal.c | 8 +- - deps/uv/src/unix/stream.c | 56 ++-- - deps/uv/src/unix/sunos.c | 18 +- - deps/uv/src/unix/tcp.c | 4 +- - deps/uv/src/unix/tty.c | 2 +- - deps/uv/src/unix/udp.c | 76 +++--- - deps/uv/src/uv-common.c | 28 +- - deps/uv/src/uv-common.h | 3 +- - deps/uv/src/win/core.c | 6 +- - deps/uv/src/win/fs.c | 311 ++++++++--------------- - deps/uv/src/win/handle-inl.h | 2 +- - deps/uv/src/win/pipe.c | 26 +- - deps/uv/src/win/tcp.c | 4 +- - deps/uv/src/win/udp.c | 4 +- - deps/uv/test/test-fs.c | 53 ---- - deps/uv/test/test-list.h | 2 - - deps/uv/test/test-queue-foreach-delete.c | 5 +- - deps/uv/test/test-threadpool-cancel.c | 5 - - 42 files changed, 625 insertions(+), 954 deletions(-) - -diff --git a/deps/uv/include/uv.h b/deps/uv/include/uv.h -index 02397dd0fdd..1af95289776 100644 ---- a/deps/uv/include/uv.h -+++ b/deps/uv/include/uv.h -@@ -59,12 +59,6 @@ extern "C" { - #include - #include - --/* Internal type, do not use. */ --struct uv__queue { -- struct uv__queue* next; -- struct uv__queue* prev; --}; -- - #if defined(_WIN32) - # include "uv/win.h" - #else -@@ -466,7 +460,7 @@ struct uv_shutdown_s { - uv_handle_type type; \ - /* private */ \ - uv_close_cb close_cb; \ -- struct uv__queue handle_queue; \ -+ void* handle_queue[2]; \ - union { \ - int fd; \ - void* reserved[4]; \ -@@ -1870,7 +1864,7 @@ struct uv_loop_s { - void* data; - /* Loop reference counting. */ - unsigned int active_handles; -- struct uv__queue handle_queue; -+ void* handle_queue[2]; - union { - void* unused; - unsigned int count; -diff --git a/deps/uv/include/uv/darwin.h b/deps/uv/include/uv/darwin.h -index 06962bfda80..d226415820b 100644 ---- a/deps/uv/include/uv/darwin.h -+++ b/deps/uv/include/uv/darwin.h -@@ -40,7 +40,7 @@ - void* cf_state; \ - uv_mutex_t cf_mutex; \ - uv_sem_t cf_sem; \ -- struct uv__queue cf_signals; \ -+ void* cf_signals[2]; \ - - #define UV_PLATFORM_FS_EVENT_FIELDS \ - uv__io_t event_watcher; \ -@@ -48,8 +48,8 @@ - int realpath_len; \ - int cf_flags; \ - uv_async_t* cf_cb; \ -- struct uv__queue cf_events; \ -- struct uv__queue cf_member; \ -+ void* cf_events[2]; \ -+ void* cf_member[2]; \ - int cf_error; \ - uv_mutex_t cf_mutex; \ - -diff --git a/deps/uv/include/uv/linux.h b/deps/uv/include/uv/linux.h -index 9f22f8cf726..9b38405a190 100644 ---- a/deps/uv/include/uv/linux.h -+++ b/deps/uv/include/uv/linux.h -@@ -28,7 +28,7 @@ - int inotify_fd; \ - - #define UV_PLATFORM_FS_EVENT_FIELDS \ -- struct uv__queue watchers; \ -+ void* watchers[2]; \ - int wd; \ - - #endif /* UV_LINUX_H */ -diff --git a/deps/uv/include/uv/threadpool.h b/deps/uv/include/uv/threadpool.h -index 24ce916fda4..9708ebdd530 100644 ---- a/deps/uv/include/uv/threadpool.h -+++ b/deps/uv/include/uv/threadpool.h -@@ -31,7 +31,7 @@ struct uv__work { - void (*work)(struct uv__work *w); - void (*done)(struct uv__work *w, int status); - struct uv_loop_s* loop; -- struct uv__queue wq; -+ void* wq[2]; - }; - - #endif /* UV_THREADPOOL_H_ */ -diff --git a/deps/uv/include/uv/unix.h b/deps/uv/include/uv/unix.h -index 09f88a56742..95447b8dd67 100644 ---- a/deps/uv/include/uv/unix.h -+++ b/deps/uv/include/uv/unix.h -@@ -92,8 +92,8 @@ typedef struct uv__io_s uv__io_t; - - struct uv__io_s { - uv__io_cb cb; -- struct uv__queue pending_queue; -- struct uv__queue watcher_queue; -+ void* pending_queue[2]; -+ void* watcher_queue[2]; - unsigned int pevents; /* Pending event mask i.e. mask at next tick. */ - unsigned int events; /* Current event mask. */ - int fd; -@@ -220,21 +220,21 @@ typedef struct { - #define UV_LOOP_PRIVATE_FIELDS \ - unsigned long flags; \ - int backend_fd; \ -- struct uv__queue pending_queue; \ -- struct uv__queue watcher_queue; \ -+ void* pending_queue[2]; \ -+ void* watcher_queue[2]; \ - uv__io_t** watchers; \ - unsigned int nwatchers; \ - unsigned int nfds; \ -- struct uv__queue wq; \ -+ void* wq[2]; \ - uv_mutex_t wq_mutex; \ - uv_async_t wq_async; \ - uv_rwlock_t cloexec_lock; \ - uv_handle_t* closing_handles; \ -- struct uv__queue process_handles; \ -- struct uv__queue prepare_handles; \ -- struct uv__queue check_handles; \ -- struct uv__queue idle_handles; \ -- struct uv__queue async_handles; \ -+ void* process_handles[2]; \ -+ void* prepare_handles[2]; \ -+ void* check_handles[2]; \ -+ void* idle_handles[2]; \ -+ void* async_handles[2]; \ - void (*async_unused)(void); /* TODO(bnoordhuis) Remove in libuv v2. */ \ - uv__io_t async_io_watcher; \ - int async_wfd; \ -@@ -257,7 +257,7 @@ typedef struct { - #define UV_PRIVATE_REQ_TYPES /* empty */ - - #define UV_WRITE_PRIVATE_FIELDS \ -- struct uv__queue queue; \ -+ void* queue[2]; \ - unsigned int write_index; \ - uv_buf_t* bufs; \ - unsigned int nbufs; \ -@@ -265,12 +265,12 @@ typedef struct { - uv_buf_t bufsml[4]; \ - - #define UV_CONNECT_PRIVATE_FIELDS \ -- struct uv__queue queue; \ -+ void* queue[2]; \ - - #define UV_SHUTDOWN_PRIVATE_FIELDS /* empty */ - - #define UV_UDP_SEND_PRIVATE_FIELDS \ -- struct uv__queue queue; \ -+ void* queue[2]; \ - struct sockaddr_storage addr; \ - unsigned int nbufs; \ - uv_buf_t* bufs; \ -@@ -286,8 +286,8 @@ typedef struct { - uv_connect_t *connect_req; \ - uv_shutdown_t *shutdown_req; \ - uv__io_t io_watcher; \ -- struct uv__queue write_queue; \ -- struct uv__queue write_completed_queue; \ -+ void* write_queue[2]; \ -+ void* write_completed_queue[2]; \ - uv_connection_cb connection_cb; \ - int delayed_error; \ - int accepted_fd; \ -@@ -300,8 +300,8 @@ typedef struct { - uv_alloc_cb alloc_cb; \ - uv_udp_recv_cb recv_cb; \ - uv__io_t io_watcher; \ -- struct uv__queue write_queue; \ -- struct uv__queue write_completed_queue; \ -+ void* write_queue[2]; \ -+ void* write_completed_queue[2]; \ - - #define UV_PIPE_PRIVATE_FIELDS \ - const char* pipe_fname; /* NULL or strdup'ed */ -@@ -311,19 +311,19 @@ typedef struct { - - #define UV_PREPARE_PRIVATE_FIELDS \ - uv_prepare_cb prepare_cb; \ -- struct uv__queue queue; \ -+ void* queue[2]; \ - - #define UV_CHECK_PRIVATE_FIELDS \ - uv_check_cb check_cb; \ -- struct uv__queue queue; \ -+ void* queue[2]; \ - - #define UV_IDLE_PRIVATE_FIELDS \ - uv_idle_cb idle_cb; \ -- struct uv__queue queue; \ -+ void* queue[2]; \ - - #define UV_ASYNC_PRIVATE_FIELDS \ - uv_async_cb async_cb; \ -- struct uv__queue queue; \ -+ void* queue[2]; \ - int pending; \ - - #define UV_TIMER_PRIVATE_FIELDS \ -@@ -352,7 +352,7 @@ typedef struct { - int retcode; - - #define UV_PROCESS_PRIVATE_FIELDS \ -- struct uv__queue queue; \ -+ void* queue[2]; \ - int status; \ - - #define UV_FS_PRIVATE_FIELDS \ -diff --git a/deps/uv/include/uv/win.h b/deps/uv/include/uv/win.h -index 6f8c47298e4..92a95fa15f1 100644 ---- a/deps/uv/include/uv/win.h -+++ b/deps/uv/include/uv/win.h -@@ -357,7 +357,7 @@ typedef struct { - /* Counter to started timer */ \ - uint64_t timer_counter; \ - /* Threadpool */ \ -- struct uv__queue wq; \ -+ void* wq[2]; \ - uv_mutex_t wq_mutex; \ - uv_async_t wq_async; - -@@ -486,7 +486,7 @@ typedef struct { - uint32_t payload_remaining; \ - uint64_t dummy; /* TODO: retained for ABI compat; remove this in v2.x. */ \ - } ipc_data_frame; \ -- struct uv__queue ipc_xfer_queue; \ -+ void* ipc_xfer_queue[2]; \ - int ipc_xfer_queue_length; \ - uv_write_t* non_overlapped_writes_tail; \ - CRITICAL_SECTION readfile_thread_lock; \ -diff --git a/deps/uv/src/queue.h b/deps/uv/src/queue.h -index 5f8489e9bc5..ff3540a0a51 100644 ---- a/deps/uv/src/queue.h -+++ b/deps/uv/src/queue.h -@@ -18,73 +18,91 @@ - - #include - --#define uv__queue_data(pointer, type, field) \ -- ((type*) ((char*) (pointer) - offsetof(type, field))) -- --#define uv__queue_foreach(q, h) \ -- for ((q) = (h)->next; (q) != (h); (q) = (q)->next) -- --static inline void uv__queue_init(struct uv__queue* q) { -- q->next = q; -- q->prev = q; --} -- --static inline int uv__queue_empty(const struct uv__queue* q) { -- return q == q->next; --} -- --static inline struct uv__queue* uv__queue_head(const struct uv__queue* q) { -- return q->next; --} -- --static inline struct uv__queue* uv__queue_next(const struct uv__queue* q) { -- return q->next; --} -- --static inline void uv__queue_add(struct uv__queue* h, struct uv__queue* n) { -- h->prev->next = n->next; -- n->next->prev = h->prev; -- h->prev = n->prev; -- h->prev->next = h; --} -- --static inline void uv__queue_split(struct uv__queue* h, -- struct uv__queue* q, -- struct uv__queue* n) { -- n->prev = h->prev; -- n->prev->next = n; -- n->next = q; -- h->prev = q->prev; -- h->prev->next = h; -- q->prev = n; --} -- --static inline void uv__queue_move(struct uv__queue* h, struct uv__queue* n) { -- if (uv__queue_empty(h)) -- uv__queue_init(n); -- else -- uv__queue_split(h, h->next, n); --} -- --static inline void uv__queue_insert_head(struct uv__queue* h, -- struct uv__queue* q) { -- q->next = h->next; -- q->prev = h; -- q->next->prev = q; -- h->next = q; --} -- --static inline void uv__queue_insert_tail(struct uv__queue* h, -- struct uv__queue* q) { -- q->next = h; -- q->prev = h->prev; -- q->prev->next = q; -- h->prev = q; --} -- --static inline void uv__queue_remove(struct uv__queue* q) { -- q->prev->next = q->next; -- q->next->prev = q->prev; --} -+typedef void *QUEUE[2]; -+ -+/* Private macros. */ -+#define QUEUE_NEXT(q) (*(QUEUE **) &((*(q))[0])) -+#define QUEUE_PREV(q) (*(QUEUE **) &((*(q))[1])) -+#define QUEUE_PREV_NEXT(q) (QUEUE_NEXT(QUEUE_PREV(q))) -+#define QUEUE_NEXT_PREV(q) (QUEUE_PREV(QUEUE_NEXT(q))) -+ -+/* Public macros. */ -+#define QUEUE_DATA(ptr, type, field) \ -+ ((type *) ((char *) (ptr) - offsetof(type, field))) -+ -+/* Important note: mutating the list while QUEUE_FOREACH is -+ * iterating over its elements results in undefined behavior. -+ */ -+#define QUEUE_FOREACH(q, h) \ -+ for ((q) = QUEUE_NEXT(h); (q) != (h); (q) = QUEUE_NEXT(q)) -+ -+#define QUEUE_EMPTY(q) \ -+ ((const QUEUE *) (q) == (const QUEUE *) QUEUE_NEXT(q)) -+ -+#define QUEUE_HEAD(q) \ -+ (QUEUE_NEXT(q)) -+ -+#define QUEUE_INIT(q) \ -+ do { \ -+ QUEUE_NEXT(q) = (q); \ -+ QUEUE_PREV(q) = (q); \ -+ } \ -+ while (0) -+ -+#define QUEUE_ADD(h, n) \ -+ do { \ -+ QUEUE_PREV_NEXT(h) = QUEUE_NEXT(n); \ -+ QUEUE_NEXT_PREV(n) = QUEUE_PREV(h); \ -+ QUEUE_PREV(h) = QUEUE_PREV(n); \ -+ QUEUE_PREV_NEXT(h) = (h); \ -+ } \ -+ while (0) -+ -+#define QUEUE_SPLIT(h, q, n) \ -+ do { \ -+ QUEUE_PREV(n) = QUEUE_PREV(h); \ -+ QUEUE_PREV_NEXT(n) = (n); \ -+ QUEUE_NEXT(n) = (q); \ -+ QUEUE_PREV(h) = QUEUE_PREV(q); \ -+ QUEUE_PREV_NEXT(h) = (h); \ -+ QUEUE_PREV(q) = (n); \ -+ } \ -+ while (0) -+ -+#define QUEUE_MOVE(h, n) \ -+ do { \ -+ if (QUEUE_EMPTY(h)) \ -+ QUEUE_INIT(n); \ -+ else { \ -+ QUEUE* q = QUEUE_HEAD(h); \ -+ QUEUE_SPLIT(h, q, n); \ -+ } \ -+ } \ -+ while (0) -+ -+#define QUEUE_INSERT_HEAD(h, q) \ -+ do { \ -+ QUEUE_NEXT(q) = QUEUE_NEXT(h); \ -+ QUEUE_PREV(q) = (h); \ -+ QUEUE_NEXT_PREV(q) = (q); \ -+ QUEUE_NEXT(h) = (q); \ -+ } \ -+ while (0) -+ -+#define QUEUE_INSERT_TAIL(h, q) \ -+ do { \ -+ QUEUE_NEXT(q) = (h); \ -+ QUEUE_PREV(q) = QUEUE_PREV(h); \ -+ QUEUE_PREV_NEXT(q) = (q); \ -+ QUEUE_PREV(h) = (q); \ -+ } \ -+ while (0) -+ -+#define QUEUE_REMOVE(q) \ -+ do { \ -+ QUEUE_PREV_NEXT(q) = QUEUE_NEXT(q); \ -+ QUEUE_NEXT_PREV(q) = QUEUE_PREV(q); \ -+ } \ -+ while (0) - - #endif /* QUEUE_H_ */ -diff --git a/deps/uv/src/threadpool.c b/deps/uv/src/threadpool.c -index dbef67f2f10..51962bf0021 100644 ---- a/deps/uv/src/threadpool.c -+++ b/deps/uv/src/threadpool.c -@@ -37,10 +37,10 @@ static unsigned int slow_io_work_running; - static unsigned int nthreads; - static uv_thread_t* threads; - static uv_thread_t default_threads[4]; --static struct uv__queue exit_message; --static struct uv__queue wq; --static struct uv__queue run_slow_work_message; --static struct uv__queue slow_io_pending_wq; -+static QUEUE exit_message; -+static QUEUE wq; -+static QUEUE run_slow_work_message; -+static QUEUE slow_io_pending_wq; - - static unsigned int slow_work_thread_threshold(void) { - return (nthreads + 1) / 2; -@@ -56,7 +56,7 @@ static void uv__cancelled(struct uv__work* w) { - */ - static void worker(void* arg) { - struct uv__work* w; -- struct uv__queue* q; -+ QUEUE* q; - int is_slow_work; - - uv_sem_post((uv_sem_t*) arg); -@@ -68,49 +68,49 @@ static void worker(void* arg) { - - /* Keep waiting while either no work is present or only slow I/O - and we're at the threshold for that. */ -- while (uv__queue_empty(&wq) || -- (uv__queue_head(&wq) == &run_slow_work_message && -- uv__queue_next(&run_slow_work_message) == &wq && -+ while (QUEUE_EMPTY(&wq) || -+ (QUEUE_HEAD(&wq) == &run_slow_work_message && -+ QUEUE_NEXT(&run_slow_work_message) == &wq && - slow_io_work_running >= slow_work_thread_threshold())) { - idle_threads += 1; - uv_cond_wait(&cond, &mutex); - idle_threads -= 1; - } - -- q = uv__queue_head(&wq); -+ q = QUEUE_HEAD(&wq); - if (q == &exit_message) { - uv_cond_signal(&cond); - uv_mutex_unlock(&mutex); - break; - } - -- uv__queue_remove(q); -- uv__queue_init(q); /* Signal uv_cancel() that the work req is executing. */ -+ QUEUE_REMOVE(q); -+ QUEUE_INIT(q); /* Signal uv_cancel() that the work req is executing. */ - - is_slow_work = 0; - if (q == &run_slow_work_message) { - /* If we're at the slow I/O threshold, re-schedule until after all - other work in the queue is done. */ - if (slow_io_work_running >= slow_work_thread_threshold()) { -- uv__queue_insert_tail(&wq, q); -+ QUEUE_INSERT_TAIL(&wq, q); - continue; - } - - /* If we encountered a request to run slow I/O work but there is none - to run, that means it's cancelled => Start over. */ -- if (uv__queue_empty(&slow_io_pending_wq)) -+ if (QUEUE_EMPTY(&slow_io_pending_wq)) - continue; - - is_slow_work = 1; - slow_io_work_running++; - -- q = uv__queue_head(&slow_io_pending_wq); -- uv__queue_remove(q); -- uv__queue_init(q); -+ q = QUEUE_HEAD(&slow_io_pending_wq); -+ QUEUE_REMOVE(q); -+ QUEUE_INIT(q); - - /* If there is more slow I/O work, schedule it to be run as well. */ -- if (!uv__queue_empty(&slow_io_pending_wq)) { -- uv__queue_insert_tail(&wq, &run_slow_work_message); -+ if (!QUEUE_EMPTY(&slow_io_pending_wq)) { -+ QUEUE_INSERT_TAIL(&wq, &run_slow_work_message); - if (idle_threads > 0) - uv_cond_signal(&cond); - } -@@ -118,13 +118,13 @@ static void worker(void* arg) { - - uv_mutex_unlock(&mutex); - -- w = uv__queue_data(q, struct uv__work, wq); -+ w = QUEUE_DATA(q, struct uv__work, wq); - w->work(w); - - uv_mutex_lock(&w->loop->wq_mutex); - w->work = NULL; /* Signal uv_cancel() that the work req is done - executing. */ -- uv__queue_insert_tail(&w->loop->wq, &w->wq); -+ QUEUE_INSERT_TAIL(&w->loop->wq, &w->wq); - uv_async_send(&w->loop->wq_async); - uv_mutex_unlock(&w->loop->wq_mutex); - -@@ -139,12 +139,12 @@ static void worker(void* arg) { - } - - --static void post(struct uv__queue* q, enum uv__work_kind kind) { -+static void post(QUEUE* q, enum uv__work_kind kind) { - uv_mutex_lock(&mutex); - if (kind == UV__WORK_SLOW_IO) { - /* Insert into a separate queue. */ -- uv__queue_insert_tail(&slow_io_pending_wq, q); -- if (!uv__queue_empty(&run_slow_work_message)) { -+ QUEUE_INSERT_TAIL(&slow_io_pending_wq, q); -+ if (!QUEUE_EMPTY(&run_slow_work_message)) { - /* Running slow I/O tasks is already scheduled => Nothing to do here. - The worker that runs said other task will schedule this one as well. */ - uv_mutex_unlock(&mutex); -@@ -153,7 +153,7 @@ static void post(struct uv__queue* q, enum uv__work_kind kind) { - q = &run_slow_work_message; - } - -- uv__queue_insert_tail(&wq, q); -+ QUEUE_INSERT_TAIL(&wq, q); - if (idle_threads > 0) - uv_cond_signal(&cond); - uv_mutex_unlock(&mutex); -@@ -220,9 +220,9 @@ static void init_threads(void) { - if (uv_mutex_init(&mutex)) - abort(); - -- uv__queue_init(&wq); -- uv__queue_init(&slow_io_pending_wq); -- uv__queue_init(&run_slow_work_message); -+ QUEUE_INIT(&wq); -+ QUEUE_INIT(&slow_io_pending_wq); -+ QUEUE_INIT(&run_slow_work_message); - - if (uv_sem_init(&sem, 0)) - abort(); -@@ -285,9 +285,9 @@ static int uv__work_cancel(uv_loop_t* loop, uv_req_t* req, struct uv__work* w) { - uv_mutex_lock(&mutex); - uv_mutex_lock(&w->loop->wq_mutex); - -- cancelled = !uv__queue_empty(&w->wq) && w->work != NULL; -+ cancelled = !QUEUE_EMPTY(&w->wq) && w->work != NULL; - if (cancelled) -- uv__queue_remove(&w->wq); -+ QUEUE_REMOVE(&w->wq); - - uv_mutex_unlock(&w->loop->wq_mutex); - uv_mutex_unlock(&mutex); -@@ -297,7 +297,7 @@ static int uv__work_cancel(uv_loop_t* loop, uv_req_t* req, struct uv__work* w) { - - w->work = uv__cancelled; - uv_mutex_lock(&loop->wq_mutex); -- uv__queue_insert_tail(&loop->wq, &w->wq); -+ QUEUE_INSERT_TAIL(&loop->wq, &w->wq); - uv_async_send(&loop->wq_async); - uv_mutex_unlock(&loop->wq_mutex); - -@@ -308,21 +308,21 @@ static int uv__work_cancel(uv_loop_t* loop, uv_req_t* req, struct uv__work* w) { - void uv__work_done(uv_async_t* handle) { - struct uv__work* w; - uv_loop_t* loop; -- struct uv__queue* q; -- struct uv__queue wq; -+ QUEUE* q; -+ QUEUE wq; - int err; - int nevents; - - loop = container_of(handle, uv_loop_t, wq_async); - uv_mutex_lock(&loop->wq_mutex); -- uv__queue_move(&loop->wq, &wq); -+ QUEUE_MOVE(&loop->wq, &wq); - uv_mutex_unlock(&loop->wq_mutex); - - nevents = 0; - -- while (!uv__queue_empty(&wq)) { -- q = uv__queue_head(&wq); -- uv__queue_remove(q); -+ while (!QUEUE_EMPTY(&wq)) { -+ q = QUEUE_HEAD(&wq); -+ QUEUE_REMOVE(q); - - w = container_of(q, struct uv__work, wq); - err = (w->work == uv__cancelled) ? UV_ECANCELED : 0; -diff --git a/deps/uv/src/unix/aix.c b/deps/uv/src/unix/aix.c -index 3af3009a216..f1afbed49ec 100644 ---- a/deps/uv/src/unix/aix.c -+++ b/deps/uv/src/unix/aix.c -@@ -136,7 +136,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - struct pollfd pqry; - struct pollfd* pe; - struct poll_ctl pc; -- struct uv__queue* q; -+ QUEUE* q; - uv__io_t* w; - uint64_t base; - uint64_t diff; -@@ -151,18 +151,18 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - int reset_timeout; - - if (loop->nfds == 0) { -- assert(uv__queue_empty(&loop->watcher_queue)); -+ assert(QUEUE_EMPTY(&loop->watcher_queue)); - return; - } - - lfields = uv__get_internal_fields(loop); - -- while (!uv__queue_empty(&loop->watcher_queue)) { -- q = uv__queue_head(&loop->watcher_queue); -- uv__queue_remove(q); -- uv__queue_init(q); -+ while (!QUEUE_EMPTY(&loop->watcher_queue)) { -+ q = QUEUE_HEAD(&loop->watcher_queue); -+ QUEUE_REMOVE(q); -+ QUEUE_INIT(q); - -- w = uv__queue_data(q, uv__io_t, watcher_queue); -+ w = QUEUE_DATA(q, uv__io_t, watcher_queue); - assert(w->pevents != 0); - assert(w->fd >= 0); - assert(w->fd < (int) loop->nwatchers); -diff --git a/deps/uv/src/unix/async.c b/deps/uv/src/unix/async.c -index 0ff2669e30a..5751b6d02be 100644 ---- a/deps/uv/src/unix/async.c -+++ b/deps/uv/src/unix/async.c -@@ -55,7 +55,7 @@ int uv_async_init(uv_loop_t* loop, uv_async_t* handle, uv_async_cb async_cb) { - handle->pending = 0; - handle->u.fd = 0; /* This will be used as a busy flag. */ - -- uv__queue_insert_tail(&loop->async_handles, &handle->queue); -+ QUEUE_INSERT_TAIL(&loop->async_handles, &handle->queue); - uv__handle_start(handle); - - return 0; -@@ -124,7 +124,7 @@ static void uv__async_spin(uv_async_t* handle) { - - void uv__async_close(uv_async_t* handle) { - uv__async_spin(handle); -- uv__queue_remove(&handle->queue); -+ QUEUE_REMOVE(&handle->queue); - uv__handle_stop(handle); - } - -@@ -132,8 +132,8 @@ void uv__async_close(uv_async_t* handle) { - static void uv__async_io(uv_loop_t* loop, uv__io_t* w, unsigned int events) { - char buf[1024]; - ssize_t r; -- struct uv__queue queue; -- struct uv__queue* q; -+ QUEUE queue; -+ QUEUE* q; - uv_async_t* h; - _Atomic int *pending; - -@@ -157,13 +157,13 @@ static void uv__async_io(uv_loop_t* loop, uv__io_t* w, unsigned int events) { - abort(); - } - -- uv__queue_move(&loop->async_handles, &queue); -- while (!uv__queue_empty(&queue)) { -- q = uv__queue_head(&queue); -- h = uv__queue_data(q, uv_async_t, queue); -+ QUEUE_MOVE(&loop->async_handles, &queue); -+ while (!QUEUE_EMPTY(&queue)) { -+ q = QUEUE_HEAD(&queue); -+ h = QUEUE_DATA(q, uv_async_t, queue); - -- uv__queue_remove(q); -- uv__queue_insert_tail(&loop->async_handles, q); -+ QUEUE_REMOVE(q); -+ QUEUE_INSERT_TAIL(&loop->async_handles, q); - - /* Atomically fetch and clear pending flag */ - pending = (_Atomic int*) &h->pending; -@@ -241,8 +241,8 @@ static int uv__async_start(uv_loop_t* loop) { - - - void uv__async_stop(uv_loop_t* loop) { -- struct uv__queue queue; -- struct uv__queue* q; -+ QUEUE queue; -+ QUEUE* q; - uv_async_t* h; - - if (loop->async_io_watcher.fd == -1) -@@ -251,13 +251,13 @@ void uv__async_stop(uv_loop_t* loop) { - /* Make sure no other thread is accessing the async handle fd after the loop - * cleanup. - */ -- uv__queue_move(&loop->async_handles, &queue); -- while (!uv__queue_empty(&queue)) { -- q = uv__queue_head(&queue); -- h = uv__queue_data(q, uv_async_t, queue); -+ QUEUE_MOVE(&loop->async_handles, &queue); -+ while (!QUEUE_EMPTY(&queue)) { -+ q = QUEUE_HEAD(&queue); -+ h = QUEUE_DATA(q, uv_async_t, queue); - -- uv__queue_remove(q); -- uv__queue_insert_tail(&loop->async_handles, q); -+ QUEUE_REMOVE(q); -+ QUEUE_INSERT_TAIL(&loop->async_handles, q); - - uv__async_spin(h); - } -@@ -275,20 +275,20 @@ void uv__async_stop(uv_loop_t* loop) { - - - int uv__async_fork(uv_loop_t* loop) { -- struct uv__queue queue; -- struct uv__queue* q; -+ QUEUE queue; -+ QUEUE* q; - uv_async_t* h; - - if (loop->async_io_watcher.fd == -1) /* never started */ - return 0; - -- uv__queue_move(&loop->async_handles, &queue); -- while (!uv__queue_empty(&queue)) { -- q = uv__queue_head(&queue); -- h = uv__queue_data(q, uv_async_t, queue); -+ QUEUE_MOVE(&loop->async_handles, &queue); -+ while (!QUEUE_EMPTY(&queue)) { -+ q = QUEUE_HEAD(&queue); -+ h = QUEUE_DATA(q, uv_async_t, queue); - -- uv__queue_remove(q); -- uv__queue_insert_tail(&loop->async_handles, q); -+ QUEUE_REMOVE(q); -+ QUEUE_INSERT_TAIL(&loop->async_handles, q); - - /* The state of any thread that set pending is now likely corrupt in this - * child because the user called fork, so just clear these flags and move -diff --git a/deps/uv/src/unix/core.c b/deps/uv/src/unix/core.c -index 25c5181f370..55aa962787e 100644 ---- a/deps/uv/src/unix/core.c -+++ b/deps/uv/src/unix/core.c -@@ -344,7 +344,7 @@ static void uv__finish_close(uv_handle_t* handle) { - } - - uv__handle_unref(handle); -- uv__queue_remove(&handle->handle_queue); -+ QUEUE_REMOVE(&handle->handle_queue); - - if (handle->close_cb) { - handle->close_cb(handle); -@@ -380,7 +380,7 @@ int uv_backend_fd(const uv_loop_t* loop) { - static int uv__loop_alive(const uv_loop_t* loop) { - return uv__has_active_handles(loop) || - uv__has_active_reqs(loop) || -- !uv__queue_empty(&loop->pending_queue) || -+ !QUEUE_EMPTY(&loop->pending_queue) || - loop->closing_handles != NULL; - } - -@@ -389,8 +389,8 @@ static int uv__backend_timeout(const uv_loop_t* loop) { - if (loop->stop_flag == 0 && - /* uv__loop_alive(loop) && */ - (uv__has_active_handles(loop) || uv__has_active_reqs(loop)) && -- uv__queue_empty(&loop->pending_queue) && -- uv__queue_empty(&loop->idle_handles) && -+ QUEUE_EMPTY(&loop->pending_queue) && -+ QUEUE_EMPTY(&loop->idle_handles) && - (loop->flags & UV_LOOP_REAP_CHILDREN) == 0 && - loop->closing_handles == NULL) - return uv__next_timeout(loop); -@@ -399,7 +399,7 @@ static int uv__backend_timeout(const uv_loop_t* loop) { - - - int uv_backend_timeout(const uv_loop_t* loop) { -- if (uv__queue_empty(&loop->watcher_queue)) -+ if (QUEUE_EMPTY(&loop->watcher_queue)) - return uv__backend_timeout(loop); - /* Need to call uv_run to update the backend fd state. */ - return 0; -@@ -431,8 +431,7 @@ int uv_run(uv_loop_t* loop, uv_run_mode mode) { - - while (r != 0 && loop->stop_flag == 0) { - can_sleep = -- uv__queue_empty(&loop->pending_queue) && -- uv__queue_empty(&loop->idle_handles); -+ QUEUE_EMPTY(&loop->pending_queue) && QUEUE_EMPTY(&loop->idle_handles); - - uv__run_pending(loop); - uv__run_idle(loop); -@@ -448,7 +447,7 @@ int uv_run(uv_loop_t* loop, uv_run_mode mode) { - - /* Process immediate callbacks (e.g. write_cb) a small fixed number of - * times to avoid loop starvation.*/ -- for (r = 0; r < 8 && !uv__queue_empty(&loop->pending_queue); r++) -+ for (r = 0; r < 8 && !QUEUE_EMPTY(&loop->pending_queue); r++) - uv__run_pending(loop); - - /* Run one final update on the provider_idle_time in case uv__io_poll -@@ -827,17 +826,17 @@ int uv_fileno(const uv_handle_t* handle, uv_os_fd_t* fd) { - - - static void uv__run_pending(uv_loop_t* loop) { -- struct uv__queue* q; -- struct uv__queue pq; -+ QUEUE* q; -+ QUEUE pq; - uv__io_t* w; - -- uv__queue_move(&loop->pending_queue, &pq); -+ QUEUE_MOVE(&loop->pending_queue, &pq); - -- while (!uv__queue_empty(&pq)) { -- q = uv__queue_head(&pq); -- uv__queue_remove(q); -- uv__queue_init(q); -- w = uv__queue_data(q, uv__io_t, pending_queue); -+ while (!QUEUE_EMPTY(&pq)) { -+ q = QUEUE_HEAD(&pq); -+ QUEUE_REMOVE(q); -+ QUEUE_INIT(q); -+ w = QUEUE_DATA(q, uv__io_t, pending_queue); - w->cb(loop, w, POLLOUT); - } - } -@@ -892,8 +891,8 @@ static void maybe_resize(uv_loop_t* loop, unsigned int len) { - void uv__io_init(uv__io_t* w, uv__io_cb cb, int fd) { - assert(cb != NULL); - assert(fd >= -1); -- uv__queue_init(&w->pending_queue); -- uv__queue_init(&w->watcher_queue); -+ QUEUE_INIT(&w->pending_queue); -+ QUEUE_INIT(&w->watcher_queue); - w->cb = cb; - w->fd = fd; - w->events = 0; -@@ -919,8 +918,8 @@ void uv__io_start(uv_loop_t* loop, uv__io_t* w, unsigned int events) { - return; - #endif - -- if (uv__queue_empty(&w->watcher_queue)) -- uv__queue_insert_tail(&loop->watcher_queue, &w->watcher_queue); -+ if (QUEUE_EMPTY(&w->watcher_queue)) -+ QUEUE_INSERT_TAIL(&loop->watcher_queue, &w->watcher_queue); - - if (loop->watchers[w->fd] == NULL) { - loop->watchers[w->fd] = w; -@@ -945,8 +944,8 @@ void uv__io_stop(uv_loop_t* loop, uv__io_t* w, unsigned int events) { - w->pevents &= ~events; - - if (w->pevents == 0) { -- uv__queue_remove(&w->watcher_queue); -- uv__queue_init(&w->watcher_queue); -+ QUEUE_REMOVE(&w->watcher_queue); -+ QUEUE_INIT(&w->watcher_queue); - w->events = 0; - - if (w == loop->watchers[w->fd]) { -@@ -955,14 +954,14 @@ void uv__io_stop(uv_loop_t* loop, uv__io_t* w, unsigned int events) { - loop->nfds--; - } - } -- else if (uv__queue_empty(&w->watcher_queue)) -- uv__queue_insert_tail(&loop->watcher_queue, &w->watcher_queue); -+ else if (QUEUE_EMPTY(&w->watcher_queue)) -+ QUEUE_INSERT_TAIL(&loop->watcher_queue, &w->watcher_queue); - } - - - void uv__io_close(uv_loop_t* loop, uv__io_t* w) { - uv__io_stop(loop, w, POLLIN | POLLOUT | UV__POLLRDHUP | UV__POLLPRI); -- uv__queue_remove(&w->pending_queue); -+ QUEUE_REMOVE(&w->pending_queue); - - /* Remove stale events for this file descriptor */ - if (w->fd != -1) -@@ -971,8 +970,8 @@ void uv__io_close(uv_loop_t* loop, uv__io_t* w) { - - - void uv__io_feed(uv_loop_t* loop, uv__io_t* w) { -- if (uv__queue_empty(&w->pending_queue)) -- uv__queue_insert_tail(&loop->pending_queue, &w->pending_queue); -+ if (QUEUE_EMPTY(&w->pending_queue)) -+ QUEUE_INSERT_TAIL(&loop->pending_queue, &w->pending_queue); - } - - -diff --git a/deps/uv/src/unix/fs.c b/deps/uv/src/unix/fs.c -index 6b051c124f2..00d385c24b7 100644 ---- a/deps/uv/src/unix/fs.c -+++ b/deps/uv/src/unix/fs.c -@@ -62,6 +62,7 @@ - - #if defined(__linux__) - # include -+# include - #endif - - #if defined(__sun) -@@ -903,6 +904,31 @@ out: - - - #ifdef __linux__ -+static unsigned uv__kernel_version(void) { -+ static _Atomic unsigned cached_version; -+ struct utsname u; -+ unsigned version; -+ unsigned major; -+ unsigned minor; -+ unsigned patch; -+ -+ version = atomic_load_explicit(&cached_version, memory_order_relaxed); -+ if (version != 0) -+ return version; -+ -+ if (-1 == uname(&u)) -+ return 0; -+ -+ if (3 != sscanf(u.release, "%u.%u.%u", &major, &minor, &patch)) -+ return 0; -+ -+ version = major * 65536 + minor * 256 + patch; -+ atomic_store_explicit(&cached_version, version, memory_order_relaxed); -+ -+ return version; -+} -+ -+ - /* Pre-4.20 kernels have a bug where CephFS uses the RADOS copy-from command - * in copy_file_range() when it shouldn't. There is no workaround except to - * fall back to a regular copy. -@@ -1905,9 +1931,6 @@ int uv_fs_link(uv_loop_t* loop, - uv_fs_cb cb) { - INIT(LINK); - PATH2; -- if (cb != NULL) -- if (uv__iou_fs_link(loop, req)) -- return 0; - POST; - } - -@@ -1920,9 +1943,6 @@ int uv_fs_mkdir(uv_loop_t* loop, - INIT(MKDIR); - PATH; - req->mode = mode; -- if (cb != NULL) -- if (uv__iou_fs_mkdir(loop, req)) -- return 0; - POST; - } - -@@ -2074,9 +2094,6 @@ int uv_fs_rename(uv_loop_t* loop, - uv_fs_cb cb) { - INIT(RENAME); - PATH2; -- if (cb != NULL) -- if (uv__iou_fs_rename(loop, req)) -- return 0; - POST; - } - -@@ -2123,9 +2140,6 @@ int uv_fs_symlink(uv_loop_t* loop, - INIT(SYMLINK); - PATH2; - req->flags = flags; -- if (cb != NULL) -- if (uv__iou_fs_symlink(loop, req)) -- return 0; - POST; - } - -@@ -2133,9 +2147,6 @@ int uv_fs_symlink(uv_loop_t* loop, - int uv_fs_unlink(uv_loop_t* loop, uv_fs_t* req, const char* path, uv_fs_cb cb) { - INIT(UNLINK); - PATH; -- if (cb != NULL) -- if (uv__iou_fs_unlink(loop, req)) -- return 0; - POST; - } - -diff --git a/deps/uv/src/unix/fsevents.c b/deps/uv/src/unix/fsevents.c -index df703f3635f..0535b4547aa 100644 ---- a/deps/uv/src/unix/fsevents.c -+++ b/deps/uv/src/unix/fsevents.c -@@ -80,13 +80,13 @@ enum uv__cf_loop_signal_type_e { - typedef enum uv__cf_loop_signal_type_e uv__cf_loop_signal_type_t; - - struct uv__cf_loop_signal_s { -- struct uv__queue member; -+ QUEUE member; - uv_fs_event_t* handle; - uv__cf_loop_signal_type_t type; - }; - - struct uv__fsevents_event_s { -- struct uv__queue member; -+ QUEUE member; - int events; - char path[1]; - }; -@@ -98,7 +98,7 @@ struct uv__cf_loop_state_s { - FSEventStreamRef fsevent_stream; - uv_sem_t fsevent_sem; - uv_mutex_t fsevent_mutex; -- struct uv__queue fsevent_handles; -+ void* fsevent_handles[2]; - unsigned int fsevent_handle_count; - }; - -@@ -150,22 +150,22 @@ static void (*pFSEventStreamStop)(FSEventStreamRef); - - #define UV__FSEVENTS_PROCESS(handle, block) \ - do { \ -- struct uv__queue events; \ -- struct uv__queue* q; \ -+ QUEUE events; \ -+ QUEUE* q; \ - uv__fsevents_event_t* event; \ - int err; \ - uv_mutex_lock(&(handle)->cf_mutex); \ - /* Split-off all events and empty original queue */ \ -- uv__queue_move(&(handle)->cf_events, &events); \ -+ QUEUE_MOVE(&(handle)->cf_events, &events); \ - /* Get error (if any) and zero original one */ \ - err = (handle)->cf_error; \ - (handle)->cf_error = 0; \ - uv_mutex_unlock(&(handle)->cf_mutex); \ - /* Loop through events, deallocating each after processing */ \ -- while (!uv__queue_empty(&events)) { \ -- q = uv__queue_head(&events); \ -- event = uv__queue_data(q, uv__fsevents_event_t, member); \ -- uv__queue_remove(q); \ -+ while (!QUEUE_EMPTY(&events)) { \ -+ q = QUEUE_HEAD(&events); \ -+ event = QUEUE_DATA(q, uv__fsevents_event_t, member); \ -+ QUEUE_REMOVE(q); \ - /* NOTE: Checking uv__is_active() is required here, because handle \ - * callback may close handle and invoking it after it will lead to \ - * incorrect behaviour */ \ -@@ -193,14 +193,14 @@ static void uv__fsevents_cb(uv_async_t* cb) { - - /* Runs in CF thread, pushed event into handle's event list */ - static void uv__fsevents_push_event(uv_fs_event_t* handle, -- struct uv__queue* events, -+ QUEUE* events, - int err) { - assert(events != NULL || err != 0); - uv_mutex_lock(&handle->cf_mutex); - - /* Concatenate two queues */ - if (events != NULL) -- uv__queue_add(&handle->cf_events, events); -+ QUEUE_ADD(&handle->cf_events, events); - - /* Propagate error */ - if (err != 0) -@@ -224,12 +224,12 @@ static void uv__fsevents_event_cb(const FSEventStreamRef streamRef, - char* path; - char* pos; - uv_fs_event_t* handle; -- struct uv__queue* q; -+ QUEUE* q; - uv_loop_t* loop; - uv__cf_loop_state_t* state; - uv__fsevents_event_t* event; - FSEventStreamEventFlags flags; -- struct uv__queue head; -+ QUEUE head; - - loop = info; - state = loop->cf_state; -@@ -238,9 +238,9 @@ static void uv__fsevents_event_cb(const FSEventStreamRef streamRef, - - /* For each handle */ - uv_mutex_lock(&state->fsevent_mutex); -- uv__queue_foreach(q, &state->fsevent_handles) { -- handle = uv__queue_data(q, uv_fs_event_t, cf_member); -- uv__queue_init(&head); -+ QUEUE_FOREACH(q, &state->fsevent_handles) { -+ handle = QUEUE_DATA(q, uv_fs_event_t, cf_member); -+ QUEUE_INIT(&head); - - /* Process and filter out events */ - for (i = 0; i < numEvents; i++) { -@@ -318,10 +318,10 @@ static void uv__fsevents_event_cb(const FSEventStreamRef streamRef, - event->events = UV_CHANGE; - } - -- uv__queue_insert_tail(&head, &event->member); -+ QUEUE_INSERT_TAIL(&head, &event->member); - } - -- if (!uv__queue_empty(&head)) -+ if (!QUEUE_EMPTY(&head)) - uv__fsevents_push_event(handle, &head, 0); - } - uv_mutex_unlock(&state->fsevent_mutex); -@@ -403,7 +403,7 @@ static void uv__fsevents_destroy_stream(uv__cf_loop_state_t* state) { - static void uv__fsevents_reschedule(uv__cf_loop_state_t* state, - uv_loop_t* loop, - uv__cf_loop_signal_type_t type) { -- struct uv__queue* q; -+ QUEUE* q; - uv_fs_event_t* curr; - CFArrayRef cf_paths; - CFStringRef* paths; -@@ -446,9 +446,9 @@ static void uv__fsevents_reschedule(uv__cf_loop_state_t* state, - - q = &state->fsevent_handles; - for (; i < path_count; i++) { -- q = uv__queue_next(q); -+ q = QUEUE_NEXT(q); - assert(q != &state->fsevent_handles); -- curr = uv__queue_data(q, uv_fs_event_t, cf_member); -+ curr = QUEUE_DATA(q, uv_fs_event_t, cf_member); - - assert(curr->realpath != NULL); - paths[i] = -@@ -486,8 +486,8 @@ final: - - /* Broadcast error to all handles */ - uv_mutex_lock(&state->fsevent_mutex); -- uv__queue_foreach(q, &state->fsevent_handles) { -- curr = uv__queue_data(q, uv_fs_event_t, cf_member); -+ QUEUE_FOREACH(q, &state->fsevent_handles) { -+ curr = QUEUE_DATA(q, uv_fs_event_t, cf_member); - uv__fsevents_push_event(curr, NULL, err); - } - uv_mutex_unlock(&state->fsevent_mutex); -@@ -606,7 +606,7 @@ static int uv__fsevents_loop_init(uv_loop_t* loop) { - if (err) - goto fail_sem_init; - -- uv__queue_init(&loop->cf_signals); -+ QUEUE_INIT(&loop->cf_signals); - - err = uv_sem_init(&state->fsevent_sem, 0); - if (err) -@@ -616,7 +616,7 @@ static int uv__fsevents_loop_init(uv_loop_t* loop) { - if (err) - goto fail_fsevent_mutex_init; - -- uv__queue_init(&state->fsevent_handles); -+ QUEUE_INIT(&state->fsevent_handles); - state->fsevent_need_reschedule = 0; - state->fsevent_handle_count = 0; - -@@ -675,7 +675,7 @@ fail_mutex_init: - void uv__fsevents_loop_delete(uv_loop_t* loop) { - uv__cf_loop_signal_t* s; - uv__cf_loop_state_t* state; -- struct uv__queue* q; -+ QUEUE* q; - - if (loop->cf_state == NULL) - return; -@@ -688,10 +688,10 @@ void uv__fsevents_loop_delete(uv_loop_t* loop) { - uv_mutex_destroy(&loop->cf_mutex); - - /* Free any remaining data */ -- while (!uv__queue_empty(&loop->cf_signals)) { -- q = uv__queue_head(&loop->cf_signals); -- s = uv__queue_data(q, uv__cf_loop_signal_t, member); -- uv__queue_remove(q); -+ while (!QUEUE_EMPTY(&loop->cf_signals)) { -+ q = QUEUE_HEAD(&loop->cf_signals); -+ s = QUEUE_DATA(q, uv__cf_loop_signal_t, member); -+ QUEUE_REMOVE(q); - uv__free(s); - } - -@@ -735,22 +735,22 @@ static void* uv__cf_loop_runner(void* arg) { - static void uv__cf_loop_cb(void* arg) { - uv_loop_t* loop; - uv__cf_loop_state_t* state; -- struct uv__queue* item; -- struct uv__queue split_head; -+ QUEUE* item; -+ QUEUE split_head; - uv__cf_loop_signal_t* s; - - loop = arg; - state = loop->cf_state; - - uv_mutex_lock(&loop->cf_mutex); -- uv__queue_move(&loop->cf_signals, &split_head); -+ QUEUE_MOVE(&loop->cf_signals, &split_head); - uv_mutex_unlock(&loop->cf_mutex); - -- while (!uv__queue_empty(&split_head)) { -- item = uv__queue_head(&split_head); -- uv__queue_remove(item); -+ while (!QUEUE_EMPTY(&split_head)) { -+ item = QUEUE_HEAD(&split_head); -+ QUEUE_REMOVE(item); - -- s = uv__queue_data(item, uv__cf_loop_signal_t, member); -+ s = QUEUE_DATA(item, uv__cf_loop_signal_t, member); - - /* This was a termination signal */ - if (s->handle == NULL) -@@ -778,7 +778,7 @@ int uv__cf_loop_signal(uv_loop_t* loop, - item->type = type; - - uv_mutex_lock(&loop->cf_mutex); -- uv__queue_insert_tail(&loop->cf_signals, &item->member); -+ QUEUE_INSERT_TAIL(&loop->cf_signals, &item->member); - - state = loop->cf_state; - assert(state != NULL); -@@ -807,7 +807,7 @@ int uv__fsevents_init(uv_fs_event_t* handle) { - handle->realpath_len = strlen(handle->realpath); - - /* Initialize event queue */ -- uv__queue_init(&handle->cf_events); -+ QUEUE_INIT(&handle->cf_events); - handle->cf_error = 0; - - /* -@@ -832,7 +832,7 @@ int uv__fsevents_init(uv_fs_event_t* handle) { - /* Insert handle into the list */ - state = handle->loop->cf_state; - uv_mutex_lock(&state->fsevent_mutex); -- uv__queue_insert_tail(&state->fsevent_handles, &handle->cf_member); -+ QUEUE_INSERT_TAIL(&state->fsevent_handles, &handle->cf_member); - state->fsevent_handle_count++; - state->fsevent_need_reschedule = 1; - uv_mutex_unlock(&state->fsevent_mutex); -@@ -872,7 +872,7 @@ int uv__fsevents_close(uv_fs_event_t* handle) { - /* Remove handle from the list */ - state = handle->loop->cf_state; - uv_mutex_lock(&state->fsevent_mutex); -- uv__queue_remove(&handle->cf_member); -+ QUEUE_REMOVE(&handle->cf_member); - state->fsevent_handle_count--; - state->fsevent_need_reschedule = 1; - uv_mutex_unlock(&state->fsevent_mutex); -diff --git a/deps/uv/src/unix/internal.h b/deps/uv/src/unix/internal.h -index fe588513603..6c5822e6a0d 100644 ---- a/deps/uv/src/unix/internal.h -+++ b/deps/uv/src/unix/internal.h -@@ -335,30 +335,20 @@ int uv__iou_fs_close(uv_loop_t* loop, uv_fs_t* req); - int uv__iou_fs_fsync_or_fdatasync(uv_loop_t* loop, - uv_fs_t* req, - uint32_t fsync_flags); --int uv__iou_fs_link(uv_loop_t* loop, uv_fs_t* req); --int uv__iou_fs_mkdir(uv_loop_t* loop, uv_fs_t* req); - int uv__iou_fs_open(uv_loop_t* loop, uv_fs_t* req); - int uv__iou_fs_read_or_write(uv_loop_t* loop, - uv_fs_t* req, - int is_read); --int uv__iou_fs_rename(uv_loop_t* loop, uv_fs_t* req); - int uv__iou_fs_statx(uv_loop_t* loop, - uv_fs_t* req, - int is_fstat, - int is_lstat); --int uv__iou_fs_symlink(uv_loop_t* loop, uv_fs_t* req); --int uv__iou_fs_unlink(uv_loop_t* loop, uv_fs_t* req); - #else - #define uv__iou_fs_close(loop, req) 0 - #define uv__iou_fs_fsync_or_fdatasync(loop, req, fsync_flags) 0 --#define uv__iou_fs_link(loop, req) 0 --#define uv__iou_fs_mkdir(loop, req) 0 - #define uv__iou_fs_open(loop, req) 0 - #define uv__iou_fs_read_or_write(loop, req, is_read) 0 --#define uv__iou_fs_rename(loop, req) 0 - #define uv__iou_fs_statx(loop, req, is_fstat, is_lstat) 0 --#define uv__iou_fs_symlink(loop, req) 0 --#define uv__iou_fs_unlink(loop, req) 0 - #endif - - #if defined(__APPLE__) -@@ -439,7 +429,6 @@ int uv__statx(int dirfd, - struct uv__statx* statxbuf); - void uv__statx_to_stat(const struct uv__statx* statxbuf, uv_stat_t* buf); - ssize_t uv__getrandom(void* buf, size_t buflen, unsigned flags); --unsigned uv__kernel_version(void); - #endif - - typedef int (*uv__peersockfunc)(int, struct sockaddr*, socklen_t*); -diff --git a/deps/uv/src/unix/kqueue.c b/deps/uv/src/unix/kqueue.c -index b78242d3be4..82916d65933 100644 ---- a/deps/uv/src/unix/kqueue.c -+++ b/deps/uv/src/unix/kqueue.c -@@ -133,7 +133,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - struct timespec spec; - unsigned int nevents; - unsigned int revents; -- struct uv__queue* q; -+ QUEUE* q; - uv__io_t* w; - uv_process_t* process; - sigset_t* pset; -@@ -152,19 +152,19 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - int reset_timeout; - - if (loop->nfds == 0) { -- assert(uv__queue_empty(&loop->watcher_queue)); -+ assert(QUEUE_EMPTY(&loop->watcher_queue)); - return; - } - - lfields = uv__get_internal_fields(loop); - nevents = 0; - -- while (!uv__queue_empty(&loop->watcher_queue)) { -- q = uv__queue_head(&loop->watcher_queue); -- uv__queue_remove(q); -- uv__queue_init(q); -+ while (!QUEUE_EMPTY(&loop->watcher_queue)) { -+ q = QUEUE_HEAD(&loop->watcher_queue); -+ QUEUE_REMOVE(q); -+ QUEUE_INIT(q); - -- w = uv__queue_data(q, uv__io_t, watcher_queue); -+ w = QUEUE_DATA(q, uv__io_t, watcher_queue); - assert(w->pevents != 0); - assert(w->fd >= 0); - assert(w->fd < (int) loop->nwatchers); -@@ -307,8 +307,8 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - - /* Handle kevent NOTE_EXIT results */ - if (ev->filter == EVFILT_PROC) { -- uv__queue_foreach(q, &loop->process_handles) { -- process = uv__queue_data(q, uv_process_t, queue); -+ QUEUE_FOREACH(q, &loop->process_handles) { -+ process = QUEUE_DATA(q, uv_process_t, queue); - if (process->pid == fd) { - process->flags |= UV_HANDLE_REAP; - loop->flags |= UV_LOOP_REAP_CHILDREN; -diff --git a/deps/uv/src/unix/linux.c b/deps/uv/src/unix/linux.c -index 48b9c2c43e1..5f84ad0eea3 100644 ---- a/deps/uv/src/unix/linux.c -+++ b/deps/uv/src/unix/linux.c -@@ -48,7 +48,6 @@ - #include - #include - #include --#include - #include - #include - -@@ -151,11 +150,6 @@ enum { - UV__IORING_OP_CLOSE = 19, - UV__IORING_OP_STATX = 21, - UV__IORING_OP_EPOLL_CTL = 29, -- UV__IORING_OP_RENAMEAT = 35, -- UV__IORING_OP_UNLINKAT = 36, -- UV__IORING_OP_MKDIRAT = 37, -- UV__IORING_OP_SYMLINKAT = 38, -- UV__IORING_OP_LINKAT = 39, - }; - - enum { -@@ -168,10 +162,6 @@ enum { - UV__IORING_SQ_CQ_OVERFLOW = 2u, - }; - --enum { -- UV__MKDIRAT_SYMLINKAT_LINKAT = 1u, --}; -- - struct uv__io_cqring_offsets { - uint32_t head; - uint32_t tail; -@@ -267,7 +257,7 @@ STATIC_ASSERT(EPOLL_CTL_MOD < 4); - - struct watcher_list { - RB_ENTRY(watcher_list) entry; -- struct uv__queue watchers; -+ QUEUE watchers; - int iterating; - char* path; - int wd; -@@ -310,31 +300,6 @@ static struct watcher_root* uv__inotify_watchers(uv_loop_t* loop) { - } - - --unsigned uv__kernel_version(void) { -- static _Atomic unsigned cached_version; -- struct utsname u; -- unsigned version; -- unsigned major; -- unsigned minor; -- unsigned patch; -- -- version = atomic_load_explicit(&cached_version, memory_order_relaxed); -- if (version != 0) -- return version; -- -- if (-1 == uname(&u)) -- return 0; -- -- if (3 != sscanf(u.release, "%u.%u.%u", &major, &minor, &patch)) -- return 0; -- -- version = major * 65536 + minor * 256 + patch; -- atomic_store_explicit(&cached_version, version, memory_order_relaxed); -- -- return version; --} -- -- - ssize_t - uv__fs_copy_file_range(int fd_in, - off_t* off_in, -@@ -420,9 +385,6 @@ int uv__io_uring_register(int fd, unsigned opcode, void* arg, unsigned nargs) { - - - static int uv__use_io_uring(void) { --#if defined(__ANDROID_API__) -- return 0; /* Possibly available but blocked by seccomp. */ --#else - /* Ternary: unknown=0, yes=1, no=-1 */ - static _Atomic int use_io_uring; - char* val; -@@ -437,7 +399,6 @@ static int uv__use_io_uring(void) { - } - - return use > 0; --#endif - } - - -@@ -542,10 +503,6 @@ static void uv__iou_init(int epollfd, - iou->sqelen = sqelen; - iou->ringfd = ringfd; - iou->in_flight = 0; -- iou->flags = 0; -- -- if (uv__kernel_version() >= /* 5.15.0 */ 0x050F00) -- iou->flags |= UV__MKDIRAT_SYMLINKAT_LINKAT; - - for (i = 0; i <= iou->sqmask; i++) - iou->sqarray[i] = i; /* Slot -> sqe identity mapping. */ -@@ -727,7 +684,7 @@ static struct uv__io_uring_sqe* uv__iou_get_sqe(struct uv__iou* iou, - req->work_req.loop = loop; - req->work_req.work = NULL; - req->work_req.done = NULL; -- uv__queue_init(&req->work_req.wq); -+ QUEUE_INIT(&req->work_req.wq); - - uv__req_register(loop, req); - iou->in_flight++; -@@ -757,17 +714,6 @@ int uv__iou_fs_close(uv_loop_t* loop, uv_fs_t* req) { - struct uv__io_uring_sqe* sqe; - struct uv__iou* iou; - -- /* Work around a poorly understood bug in older kernels where closing a file -- * descriptor pointing to /foo/bar results in ETXTBSY errors when trying to -- * execve("/foo/bar") later on. The bug seems to have been fixed somewhere -- * between 5.15.85 and 5.15.90. I couldn't pinpoint the responsible commit -- * but good candidates are the several data race fixes. Interestingly, it -- * seems to manifest only when running under Docker so the possibility of -- * a Docker bug can't be completely ruled out either. Yay, computers. -- */ -- if (uv__kernel_version() < /* 5.15.90 */ 0x050F5A) -- return 0; -- - iou = &uv__get_internal_fields(loop)->iou; - - sqe = uv__iou_get_sqe(iou, loop, req); -@@ -808,55 +754,6 @@ int uv__iou_fs_fsync_or_fdatasync(uv_loop_t* loop, - } - - --int uv__iou_fs_link(uv_loop_t* loop, uv_fs_t* req) { -- struct uv__io_uring_sqe* sqe; -- struct uv__iou* iou; -- -- iou = &uv__get_internal_fields(loop)->iou; -- -- if (!(iou->flags & UV__MKDIRAT_SYMLINKAT_LINKAT)) -- return 0; -- -- sqe = uv__iou_get_sqe(iou, loop, req); -- if (sqe == NULL) -- return 0; -- -- sqe->addr = (uintptr_t) req->path; -- sqe->fd = AT_FDCWD; -- sqe->addr2 = (uintptr_t) req->new_path; -- sqe->len = AT_FDCWD; -- sqe->opcode = UV__IORING_OP_LINKAT; -- -- uv__iou_submit(iou); -- -- return 1; --} -- -- --int uv__iou_fs_mkdir(uv_loop_t* loop, uv_fs_t* req) { -- struct uv__io_uring_sqe* sqe; -- struct uv__iou* iou; -- -- iou = &uv__get_internal_fields(loop)->iou; -- -- if (!(iou->flags & UV__MKDIRAT_SYMLINKAT_LINKAT)) -- return 0; -- -- sqe = uv__iou_get_sqe(iou, loop, req); -- if (sqe == NULL) -- return 0; -- -- sqe->addr = (uintptr_t) req->path; -- sqe->fd = AT_FDCWD; -- sqe->len = req->mode; -- sqe->opcode = UV__IORING_OP_MKDIRAT; -- -- uv__iou_submit(iou); -- -- return 1; --} -- -- - int uv__iou_fs_open(uv_loop_t* loop, uv_fs_t* req) { - struct uv__io_uring_sqe* sqe; - struct uv__iou* iou; -@@ -879,86 +776,16 @@ int uv__iou_fs_open(uv_loop_t* loop, uv_fs_t* req) { - } - - --int uv__iou_fs_rename(uv_loop_t* loop, uv_fs_t* req) { -- struct uv__io_uring_sqe* sqe; -- struct uv__iou* iou; -- -- iou = &uv__get_internal_fields(loop)->iou; -- -- sqe = uv__iou_get_sqe(iou, loop, req); -- if (sqe == NULL) -- return 0; -- -- sqe->addr = (uintptr_t) req->path; -- sqe->fd = AT_FDCWD; -- sqe->addr2 = (uintptr_t) req->new_path; -- sqe->len = AT_FDCWD; -- sqe->opcode = UV__IORING_OP_RENAMEAT; -- -- uv__iou_submit(iou); -- -- return 1; --} -- -- --int uv__iou_fs_symlink(uv_loop_t* loop, uv_fs_t* req) { -- struct uv__io_uring_sqe* sqe; -- struct uv__iou* iou; -- -- iou = &uv__get_internal_fields(loop)->iou; -- -- if (!(iou->flags & UV__MKDIRAT_SYMLINKAT_LINKAT)) -- return 0; -- -- sqe = uv__iou_get_sqe(iou, loop, req); -- if (sqe == NULL) -- return 0; -- -- sqe->addr = (uintptr_t) req->path; -- sqe->fd = AT_FDCWD; -- sqe->addr2 = (uintptr_t) req->new_path; -- sqe->opcode = UV__IORING_OP_SYMLINKAT; -- -- uv__iou_submit(iou); -- -- return 1; --} -- -- --int uv__iou_fs_unlink(uv_loop_t* loop, uv_fs_t* req) { -- struct uv__io_uring_sqe* sqe; -- struct uv__iou* iou; -- -- iou = &uv__get_internal_fields(loop)->iou; -- -- sqe = uv__iou_get_sqe(iou, loop, req); -- if (sqe == NULL) -- return 0; -- -- sqe->addr = (uintptr_t) req->path; -- sqe->fd = AT_FDCWD; -- sqe->opcode = UV__IORING_OP_UNLINKAT; -- -- uv__iou_submit(iou); -- -- return 1; --} -- -- - int uv__iou_fs_read_or_write(uv_loop_t* loop, - uv_fs_t* req, - int is_read) { - struct uv__io_uring_sqe* sqe; - struct uv__iou* iou; - -- /* If iovcnt is greater than IOV_MAX, cap it to IOV_MAX on reads and fallback -- * to the threadpool on writes */ -- if (req->nbufs > IOV_MAX) { -- if (is_read) -- req->nbufs = IOV_MAX; -- else -- return 0; -- } -+ /* For the moment, if iovcnt is greater than IOV_MAX, fallback to the -+ * threadpool. In the future we might take advantage of IOSQE_IO_LINK. */ -+ if (req->nbufs > IOV_MAX) -+ return 0; - - iou = &uv__get_internal_fields(loop)->iou; - -@@ -1265,7 +1092,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - struct uv__iou* ctl; - struct uv__iou* iou; - int real_timeout; -- struct uv__queue* q; -+ QUEUE* q; - uv__io_t* w; - sigset_t* sigmask; - sigset_t sigset; -@@ -1311,11 +1138,11 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - - memset(&e, 0, sizeof(e)); - -- while (!uv__queue_empty(&loop->watcher_queue)) { -- q = uv__queue_head(&loop->watcher_queue); -- w = uv__queue_data(q, uv__io_t, watcher_queue); -- uv__queue_remove(q); -- uv__queue_init(q); -+ while (!QUEUE_EMPTY(&loop->watcher_queue)) { -+ q = QUEUE_HEAD(&loop->watcher_queue); -+ w = QUEUE_DATA(q, uv__io_t, watcher_queue); -+ QUEUE_REMOVE(q); -+ QUEUE_INIT(q); - - op = EPOLL_CTL_MOD; - if (w->events == 0) -@@ -2272,8 +2099,8 @@ static int uv__inotify_fork(uv_loop_t* loop, struct watcher_list* root) { - struct watcher_list* tmp_watcher_list_iter; - struct watcher_list* watcher_list; - struct watcher_list tmp_watcher_list; -- struct uv__queue queue; -- struct uv__queue* q; -+ QUEUE queue; -+ QUEUE* q; - uv_fs_event_t* handle; - char* tmp_path; - -@@ -2285,41 +2112,41 @@ static int uv__inotify_fork(uv_loop_t* loop, struct watcher_list* root) { - */ - loop->inotify_watchers = root; - -- uv__queue_init(&tmp_watcher_list.watchers); -+ QUEUE_INIT(&tmp_watcher_list.watchers); - /* Note that the queue we use is shared with the start and stop() -- * functions, making uv__queue_foreach unsafe to use. So we use the -- * uv__queue_move trick to safely iterate. Also don't free the watcher -+ * functions, making QUEUE_FOREACH unsafe to use. So we use the -+ * QUEUE_MOVE trick to safely iterate. Also don't free the watcher - * list until we're done iterating. c.f. uv__inotify_read. - */ - RB_FOREACH_SAFE(watcher_list, watcher_root, - uv__inotify_watchers(loop), tmp_watcher_list_iter) { - watcher_list->iterating = 1; -- uv__queue_move(&watcher_list->watchers, &queue); -- while (!uv__queue_empty(&queue)) { -- q = uv__queue_head(&queue); -- handle = uv__queue_data(q, uv_fs_event_t, watchers); -+ QUEUE_MOVE(&watcher_list->watchers, &queue); -+ while (!QUEUE_EMPTY(&queue)) { -+ q = QUEUE_HEAD(&queue); -+ handle = QUEUE_DATA(q, uv_fs_event_t, watchers); - /* It's critical to keep a copy of path here, because it - * will be set to NULL by stop() and then deallocated by - * maybe_free_watcher_list - */ - tmp_path = uv__strdup(handle->path); - assert(tmp_path != NULL); -- uv__queue_remove(q); -- uv__queue_insert_tail(&watcher_list->watchers, q); -+ QUEUE_REMOVE(q); -+ QUEUE_INSERT_TAIL(&watcher_list->watchers, q); - uv_fs_event_stop(handle); - -- uv__queue_insert_tail(&tmp_watcher_list.watchers, &handle->watchers); -+ QUEUE_INSERT_TAIL(&tmp_watcher_list.watchers, &handle->watchers); - handle->path = tmp_path; - } - watcher_list->iterating = 0; - maybe_free_watcher_list(watcher_list, loop); - } - -- uv__queue_move(&tmp_watcher_list.watchers, &queue); -- while (!uv__queue_empty(&queue)) { -- q = uv__queue_head(&queue); -- uv__queue_remove(q); -- handle = uv__queue_data(q, uv_fs_event_t, watchers); -+ QUEUE_MOVE(&tmp_watcher_list.watchers, &queue); -+ while (!QUEUE_EMPTY(&queue)) { -+ q = QUEUE_HEAD(&queue); -+ QUEUE_REMOVE(q); -+ handle = QUEUE_DATA(q, uv_fs_event_t, watchers); - tmp_path = handle->path; - handle->path = NULL; - err = uv_fs_event_start(handle, handle->cb, tmp_path, 0); -@@ -2341,7 +2168,7 @@ static struct watcher_list* find_watcher(uv_loop_t* loop, int wd) { - - static void maybe_free_watcher_list(struct watcher_list* w, uv_loop_t* loop) { - /* if the watcher_list->watchers is being iterated over, we can't free it. */ -- if ((!w->iterating) && uv__queue_empty(&w->watchers)) { -+ if ((!w->iterating) && QUEUE_EMPTY(&w->watchers)) { - /* No watchers left for this path. Clean up. */ - RB_REMOVE(watcher_root, uv__inotify_watchers(loop), w); - inotify_rm_watch(loop->inotify_fd, w->wd); -@@ -2356,8 +2183,8 @@ static void uv__inotify_read(uv_loop_t* loop, - const struct inotify_event* e; - struct watcher_list* w; - uv_fs_event_t* h; -- struct uv__queue queue; -- struct uv__queue* q; -+ QUEUE queue; -+ QUEUE* q; - const char* path; - ssize_t size; - const char *p; -@@ -2400,7 +2227,7 @@ static void uv__inotify_read(uv_loop_t* loop, - * What can go wrong? - * A callback could call uv_fs_event_stop() - * and the queue can change under our feet. -- * So, we use uv__queue_move() trick to safely iterate over the queue. -+ * So, we use QUEUE_MOVE() trick to safely iterate over the queue. - * And we don't free the watcher_list until we're done iterating. - * - * First, -@@ -2408,13 +2235,13 @@ static void uv__inotify_read(uv_loop_t* loop, - * not to free watcher_list. - */ - w->iterating = 1; -- uv__queue_move(&w->watchers, &queue); -- while (!uv__queue_empty(&queue)) { -- q = uv__queue_head(&queue); -- h = uv__queue_data(q, uv_fs_event_t, watchers); -+ QUEUE_MOVE(&w->watchers, &queue); -+ while (!QUEUE_EMPTY(&queue)) { -+ q = QUEUE_HEAD(&queue); -+ h = QUEUE_DATA(q, uv_fs_event_t, watchers); - -- uv__queue_remove(q); -- uv__queue_insert_tail(&w->watchers, q); -+ QUEUE_REMOVE(q); -+ QUEUE_INSERT_TAIL(&w->watchers, q); - - h->cb(h, path, events, 0); - } -@@ -2476,13 +2303,13 @@ int uv_fs_event_start(uv_fs_event_t* handle, - - w->wd = wd; - w->path = memcpy(w + 1, path, len); -- uv__queue_init(&w->watchers); -+ QUEUE_INIT(&w->watchers); - w->iterating = 0; - RB_INSERT(watcher_root, uv__inotify_watchers(loop), w); - - no_insert: - uv__handle_start(handle); -- uv__queue_insert_tail(&w->watchers, &handle->watchers); -+ QUEUE_INSERT_TAIL(&w->watchers, &handle->watchers); - handle->path = w->path; - handle->cb = cb; - handle->wd = wd; -@@ -2503,7 +2330,7 @@ int uv_fs_event_stop(uv_fs_event_t* handle) { - handle->wd = -1; - handle->path = NULL; - uv__handle_stop(handle); -- uv__queue_remove(&handle->watchers); -+ QUEUE_REMOVE(&handle->watchers); - - maybe_free_watcher_list(w, handle->loop); - -diff --git a/deps/uv/src/unix/loop-watcher.c b/deps/uv/src/unix/loop-watcher.c -index 2db8b515df7..b8c1c2a7102 100644 ---- a/deps/uv/src/unix/loop-watcher.c -+++ b/deps/uv/src/unix/loop-watcher.c -@@ -32,7 +32,7 @@ - int uv_##name##_start(uv_##name##_t* handle, uv_##name##_cb cb) { \ - if (uv__is_active(handle)) return 0; \ - if (cb == NULL) return UV_EINVAL; \ -- uv__queue_insert_head(&handle->loop->name##_handles, &handle->queue); \ -+ QUEUE_INSERT_HEAD(&handle->loop->name##_handles, &handle->queue); \ - handle->name##_cb = cb; \ - uv__handle_start(handle); \ - return 0; \ -@@ -40,21 +40,21 @@ - \ - int uv_##name##_stop(uv_##name##_t* handle) { \ - if (!uv__is_active(handle)) return 0; \ -- uv__queue_remove(&handle->queue); \ -+ QUEUE_REMOVE(&handle->queue); \ - uv__handle_stop(handle); \ - return 0; \ - } \ - \ - void uv__run_##name(uv_loop_t* loop) { \ - uv_##name##_t* h; \ -- struct uv__queue queue; \ -- struct uv__queue* q; \ -- uv__queue_move(&loop->name##_handles, &queue); \ -- while (!uv__queue_empty(&queue)) { \ -- q = uv__queue_head(&queue); \ -- h = uv__queue_data(q, uv_##name##_t, queue); \ -- uv__queue_remove(q); \ -- uv__queue_insert_tail(&loop->name##_handles, q); \ -+ QUEUE queue; \ -+ QUEUE* q; \ -+ QUEUE_MOVE(&loop->name##_handles, &queue); \ -+ while (!QUEUE_EMPTY(&queue)) { \ -+ q = QUEUE_HEAD(&queue); \ -+ h = QUEUE_DATA(q, uv_##name##_t, queue); \ -+ QUEUE_REMOVE(q); \ -+ QUEUE_INSERT_TAIL(&loop->name##_handles, q); \ - h->name##_cb(h); \ - } \ - } \ -diff --git a/deps/uv/src/unix/loop.c b/deps/uv/src/unix/loop.c -index a9468e8e19c..90a51b339de 100644 ---- a/deps/uv/src/unix/loop.c -+++ b/deps/uv/src/unix/loop.c -@@ -50,20 +50,20 @@ int uv_loop_init(uv_loop_t* loop) { - sizeof(lfields->loop_metrics.metrics)); - - heap_init((struct heap*) &loop->timer_heap); -- uv__queue_init(&loop->wq); -- uv__queue_init(&loop->idle_handles); -- uv__queue_init(&loop->async_handles); -- uv__queue_init(&loop->check_handles); -- uv__queue_init(&loop->prepare_handles); -- uv__queue_init(&loop->handle_queue); -+ QUEUE_INIT(&loop->wq); -+ QUEUE_INIT(&loop->idle_handles); -+ QUEUE_INIT(&loop->async_handles); -+ QUEUE_INIT(&loop->check_handles); -+ QUEUE_INIT(&loop->prepare_handles); -+ QUEUE_INIT(&loop->handle_queue); - - loop->active_handles = 0; - loop->active_reqs.count = 0; - loop->nfds = 0; - loop->watchers = NULL; - loop->nwatchers = 0; -- uv__queue_init(&loop->pending_queue); -- uv__queue_init(&loop->watcher_queue); -+ QUEUE_INIT(&loop->pending_queue); -+ QUEUE_INIT(&loop->watcher_queue); - - loop->closing_handles = NULL; - uv__update_time(loop); -@@ -85,7 +85,7 @@ int uv_loop_init(uv_loop_t* loop) { - err = uv__process_init(loop); - if (err) - goto fail_signal_init; -- uv__queue_init(&loop->process_handles); -+ QUEUE_INIT(&loop->process_handles); - - err = uv_rwlock_init(&loop->cloexec_lock); - if (err) -@@ -152,9 +152,9 @@ int uv_loop_fork(uv_loop_t* loop) { - if (w == NULL) - continue; - -- if (w->pevents != 0 && uv__queue_empty(&w->watcher_queue)) { -+ if (w->pevents != 0 && QUEUE_EMPTY(&w->watcher_queue)) { - w->events = 0; /* Force re-registration in uv__io_poll. */ -- uv__queue_insert_tail(&loop->watcher_queue, &w->watcher_queue); -+ QUEUE_INSERT_TAIL(&loop->watcher_queue, &w->watcher_queue); - } - } - -@@ -180,7 +180,7 @@ void uv__loop_close(uv_loop_t* loop) { - } - - uv_mutex_lock(&loop->wq_mutex); -- assert(uv__queue_empty(&loop->wq) && "thread pool work queue not empty!"); -+ assert(QUEUE_EMPTY(&loop->wq) && "thread pool work queue not empty!"); - assert(!uv__has_active_reqs(loop)); - uv_mutex_unlock(&loop->wq_mutex); - uv_mutex_destroy(&loop->wq_mutex); -@@ -192,8 +192,8 @@ void uv__loop_close(uv_loop_t* loop) { - uv_rwlock_destroy(&loop->cloexec_lock); - - #if 0 -- assert(uv__queue_empty(&loop->pending_queue)); -- assert(uv__queue_empty(&loop->watcher_queue)); -+ assert(QUEUE_EMPTY(&loop->pending_queue)); -+ assert(QUEUE_EMPTY(&loop->watcher_queue)); - assert(loop->nfds == 0); - #endif - -diff --git a/deps/uv/src/unix/os390-syscalls.c b/deps/uv/src/unix/os390-syscalls.c -index 7f90c270906..5861aaaa202 100644 ---- a/deps/uv/src/unix/os390-syscalls.c -+++ b/deps/uv/src/unix/os390-syscalls.c -@@ -27,7 +27,7 @@ - #include - #include - --static struct uv__queue global_epoll_queue; -+static QUEUE global_epoll_queue; - static uv_mutex_t global_epoll_lock; - static uv_once_t once = UV_ONCE_INIT; - -@@ -178,18 +178,18 @@ static void after_fork(void) { - - - static void child_fork(void) { -- struct uv__queue* q; -+ QUEUE* q; - uv_once_t child_once = UV_ONCE_INIT; - - /* reset once */ - memcpy(&once, &child_once, sizeof(child_once)); - - /* reset epoll list */ -- while (!uv__queue_empty(&global_epoll_queue)) { -+ while (!QUEUE_EMPTY(&global_epoll_queue)) { - uv__os390_epoll* lst; -- q = uv__queue_head(&global_epoll_queue); -- uv__queue_remove(q); -- lst = uv__queue_data(q, uv__os390_epoll, member); -+ q = QUEUE_HEAD(&global_epoll_queue); -+ QUEUE_REMOVE(q); -+ lst = QUEUE_DATA(q, uv__os390_epoll, member); - uv__free(lst->items); - lst->items = NULL; - lst->size = 0; -@@ -201,7 +201,7 @@ static void child_fork(void) { - - - static void epoll_init(void) { -- uv__queue_init(&global_epoll_queue); -+ QUEUE_INIT(&global_epoll_queue); - if (uv_mutex_init(&global_epoll_lock)) - abort(); - -@@ -225,7 +225,7 @@ uv__os390_epoll* epoll_create1(int flags) { - lst->items[lst->size - 1].revents = 0; - uv_once(&once, epoll_init); - uv_mutex_lock(&global_epoll_lock); -- uv__queue_insert_tail(&global_epoll_queue, &lst->member); -+ QUEUE_INSERT_TAIL(&global_epoll_queue, &lst->member); - uv_mutex_unlock(&global_epoll_lock); - } - -@@ -352,14 +352,14 @@ int epoll_wait(uv__os390_epoll* lst, struct epoll_event* events, - - - int epoll_file_close(int fd) { -- struct uv__queue* q; -+ QUEUE* q; - - uv_once(&once, epoll_init); - uv_mutex_lock(&global_epoll_lock); -- uv__queue_foreach(q, &global_epoll_queue) { -+ QUEUE_FOREACH(q, &global_epoll_queue) { - uv__os390_epoll* lst; - -- lst = uv__queue_data(q, uv__os390_epoll, member); -+ lst = QUEUE_DATA(q, uv__os390_epoll, member); - if (fd < lst->size && lst->items != NULL && lst->items[fd].fd != -1) - lst->items[fd].fd = -1; - } -@@ -371,7 +371,7 @@ int epoll_file_close(int fd) { - void epoll_queue_close(uv__os390_epoll* lst) { - /* Remove epoll instance from global queue */ - uv_mutex_lock(&global_epoll_lock); -- uv__queue_remove(&lst->member); -+ QUEUE_REMOVE(&lst->member); - uv_mutex_unlock(&global_epoll_lock); - - /* Free resources */ -diff --git a/deps/uv/src/unix/os390-syscalls.h b/deps/uv/src/unix/os390-syscalls.h -index d5f3bcf8b1c..9f504171d85 100644 ---- a/deps/uv/src/unix/os390-syscalls.h -+++ b/deps/uv/src/unix/os390-syscalls.h -@@ -45,7 +45,7 @@ struct epoll_event { - }; - - typedef struct { -- struct uv__queue member; -+ QUEUE member; - struct pollfd* items; - unsigned long size; - int msg_queue; -diff --git a/deps/uv/src/unix/os390.c b/deps/uv/src/unix/os390.c -index bbd37692d1d..a87c2d77faf 100644 ---- a/deps/uv/src/unix/os390.c -+++ b/deps/uv/src/unix/os390.c -@@ -815,7 +815,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - uv__os390_epoll* ep; - int have_signals; - int real_timeout; -- struct uv__queue* q; -+ QUEUE* q; - uv__io_t* w; - uint64_t base; - int count; -@@ -827,19 +827,19 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - int reset_timeout; - - if (loop->nfds == 0) { -- assert(uv__queue_empty(&loop->watcher_queue)); -+ assert(QUEUE_EMPTY(&loop->watcher_queue)); - return; - } - - lfields = uv__get_internal_fields(loop); - -- while (!uv__queue_empty(&loop->watcher_queue)) { -+ while (!QUEUE_EMPTY(&loop->watcher_queue)) { - uv_stream_t* stream; - -- q = uv__queue_head(&loop->watcher_queue); -- uv__queue_remove(q); -- uv__queue_init(q); -- w = uv__queue_data(q, uv__io_t, watcher_queue); -+ q = QUEUE_HEAD(&loop->watcher_queue); -+ QUEUE_REMOVE(q); -+ QUEUE_INIT(q); -+ w = QUEUE_DATA(q, uv__io_t, watcher_queue); - - assert(w->pevents != 0); - assert(w->fd >= 0); -diff --git a/deps/uv/src/unix/pipe.c b/deps/uv/src/unix/pipe.c -index d332f351830..ce91ac49b30 100644 ---- a/deps/uv/src/unix/pipe.c -+++ b/deps/uv/src/unix/pipe.c -@@ -297,7 +297,7 @@ out: - uv__req_init(handle->loop, req, UV_CONNECT); - req->handle = (uv_stream_t*)handle; - req->cb = cb; -- uv__queue_init(&req->queue); -+ QUEUE_INIT(&req->queue); - - /* Force callback to run on next tick in case of error. */ - if (err) -diff --git a/deps/uv/src/unix/posix-poll.c b/deps/uv/src/unix/posix-poll.c -index 2e016c2fbae..7e7de86845d 100644 ---- a/deps/uv/src/unix/posix-poll.c -+++ b/deps/uv/src/unix/posix-poll.c -@@ -137,7 +137,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - sigset_t set; - uint64_t time_base; - uint64_t time_diff; -- struct uv__queue* q; -+ QUEUE* q; - uv__io_t* w; - size_t i; - unsigned int nevents; -@@ -149,19 +149,19 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - int reset_timeout; - - if (loop->nfds == 0) { -- assert(uv__queue_empty(&loop->watcher_queue)); -+ assert(QUEUE_EMPTY(&loop->watcher_queue)); - return; - } - - lfields = uv__get_internal_fields(loop); - - /* Take queued watchers and add their fds to our poll fds array. */ -- while (!uv__queue_empty(&loop->watcher_queue)) { -- q = uv__queue_head(&loop->watcher_queue); -- uv__queue_remove(q); -- uv__queue_init(q); -+ while (!QUEUE_EMPTY(&loop->watcher_queue)) { -+ q = QUEUE_HEAD(&loop->watcher_queue); -+ QUEUE_REMOVE(q); -+ QUEUE_INIT(q); - -- w = uv__queue_data(q, uv__io_t, watcher_queue); -+ w = QUEUE_DATA(q, uv__io_t, watcher_queue); - assert(w->pevents != 0); - assert(w->fd >= 0); - assert(w->fd < (int) loop->nwatchers); -diff --git a/deps/uv/src/unix/process.c b/deps/uv/src/unix/process.c -index dd58c18d9b9..bbf367b57d3 100644 ---- a/deps/uv/src/unix/process.c -+++ b/deps/uv/src/unix/process.c -@@ -108,17 +108,17 @@ void uv__wait_children(uv_loop_t* loop) { - int status; - int options; - pid_t pid; -- struct uv__queue pending; -- struct uv__queue* q; -- struct uv__queue* h; -+ QUEUE pending; -+ QUEUE* q; -+ QUEUE* h; - -- uv__queue_init(&pending); -+ QUEUE_INIT(&pending); - - h = &loop->process_handles; -- q = uv__queue_head(h); -+ q = QUEUE_HEAD(h); - while (q != h) { -- process = uv__queue_data(q, uv_process_t, queue); -- q = uv__queue_next(q); -+ process = QUEUE_DATA(q, uv_process_t, queue); -+ q = QUEUE_NEXT(q); - - #ifndef UV_USE_SIGCHLD - if ((process->flags & UV_HANDLE_REAP) == 0) -@@ -149,18 +149,18 @@ void uv__wait_children(uv_loop_t* loop) { - - assert(pid == process->pid); - process->status = status; -- uv__queue_remove(&process->queue); -- uv__queue_insert_tail(&pending, &process->queue); -+ QUEUE_REMOVE(&process->queue); -+ QUEUE_INSERT_TAIL(&pending, &process->queue); - } - - h = &pending; -- q = uv__queue_head(h); -+ q = QUEUE_HEAD(h); - while (q != h) { -- process = uv__queue_data(q, uv_process_t, queue); -- q = uv__queue_next(q); -+ process = QUEUE_DATA(q, uv_process_t, queue); -+ q = QUEUE_NEXT(q); - -- uv__queue_remove(&process->queue); -- uv__queue_init(&process->queue); -+ QUEUE_REMOVE(&process->queue); -+ QUEUE_INIT(&process->queue); - uv__handle_stop(process); - - if (process->exit_cb == NULL) -@@ -176,7 +176,7 @@ void uv__wait_children(uv_loop_t* loop) { - - process->exit_cb(process, exit_status, term_signal); - } -- assert(uv__queue_empty(&pending)); -+ assert(QUEUE_EMPTY(&pending)); - } - - /* -@@ -978,7 +978,7 @@ int uv_spawn(uv_loop_t* loop, - UV_PROCESS_WINDOWS_VERBATIM_ARGUMENTS))); - - uv__handle_init(loop, (uv_handle_t*)process, UV_PROCESS); -- uv__queue_init(&process->queue); -+ QUEUE_INIT(&process->queue); - process->status = 0; - - stdio_count = options->stdio_count; -@@ -1041,7 +1041,7 @@ int uv_spawn(uv_loop_t* loop, - - process->pid = pid; - process->exit_cb = options->exit_cb; -- uv__queue_insert_tail(&loop->process_handles, &process->queue); -+ QUEUE_INSERT_TAIL(&loop->process_handles, &process->queue); - uv__handle_start(process); - } - -@@ -1103,10 +1103,10 @@ int uv_kill(int pid, int signum) { - - - void uv__process_close(uv_process_t* handle) { -- uv__queue_remove(&handle->queue); -+ QUEUE_REMOVE(&handle->queue); - uv__handle_stop(handle); - #ifdef UV_USE_SIGCHLD -- if (uv__queue_empty(&handle->loop->process_handles)) -+ if (QUEUE_EMPTY(&handle->loop->process_handles)) - uv_signal_stop(&handle->loop->child_watcher); - #endif - } -diff --git a/deps/uv/src/unix/signal.c b/deps/uv/src/unix/signal.c -index 63aba5a60e0..bb70523f561 100644 ---- a/deps/uv/src/unix/signal.c -+++ b/deps/uv/src/unix/signal.c -@@ -291,16 +291,16 @@ int uv__signal_loop_fork(uv_loop_t* loop) { - - - void uv__signal_loop_cleanup(uv_loop_t* loop) { -- struct uv__queue* q; -+ QUEUE* q; - - /* Stop all the signal watchers that are still attached to this loop. This - * ensures that the (shared) signal tree doesn't contain any invalid entries - * entries, and that signal handlers are removed when appropriate. -- * It's safe to use uv__queue_foreach here because the handles and the handle -+ * It's safe to use QUEUE_FOREACH here because the handles and the handle - * queue are not modified by uv__signal_stop(). - */ -- uv__queue_foreach(q, &loop->handle_queue) { -- uv_handle_t* handle = uv__queue_data(q, uv_handle_t, handle_queue); -+ QUEUE_FOREACH(q, &loop->handle_queue) { -+ uv_handle_t* handle = QUEUE_DATA(q, uv_handle_t, handle_queue); - - if (handle->type == UV_SIGNAL) - uv__signal_stop((uv_signal_t*) handle); -diff --git a/deps/uv/src/unix/stream.c b/deps/uv/src/unix/stream.c -index 28c4d5463c4..03f92b5045a 100644 ---- a/deps/uv/src/unix/stream.c -+++ b/deps/uv/src/unix/stream.c -@@ -94,8 +94,8 @@ void uv__stream_init(uv_loop_t* loop, - stream->accepted_fd = -1; - stream->queued_fds = NULL; - stream->delayed_error = 0; -- uv__queue_init(&stream->write_queue); -- uv__queue_init(&stream->write_completed_queue); -+ QUEUE_INIT(&stream->write_queue); -+ QUEUE_INIT(&stream->write_completed_queue); - stream->write_queue_size = 0; - - if (loop->emfile_fd == -1) { -@@ -439,15 +439,15 @@ int uv__stream_open(uv_stream_t* stream, int fd, int flags) { - - void uv__stream_flush_write_queue(uv_stream_t* stream, int error) { - uv_write_t* req; -- struct uv__queue* q; -- while (!uv__queue_empty(&stream->write_queue)) { -- q = uv__queue_head(&stream->write_queue); -- uv__queue_remove(q); -+ QUEUE* q; -+ while (!QUEUE_EMPTY(&stream->write_queue)) { -+ q = QUEUE_HEAD(&stream->write_queue); -+ QUEUE_REMOVE(q); - -- req = uv__queue_data(q, uv_write_t, queue); -+ req = QUEUE_DATA(q, uv_write_t, queue); - req->error = error; - -- uv__queue_insert_tail(&stream->write_completed_queue, &req->queue); -+ QUEUE_INSERT_TAIL(&stream->write_completed_queue, &req->queue); - } - } - -@@ -627,7 +627,7 @@ static void uv__drain(uv_stream_t* stream) { - uv_shutdown_t* req; - int err; - -- assert(uv__queue_empty(&stream->write_queue)); -+ assert(QUEUE_EMPTY(&stream->write_queue)); - if (!(stream->flags & UV_HANDLE_CLOSING)) { - uv__io_stop(stream->loop, &stream->io_watcher, POLLOUT); - uv__stream_osx_interrupt_select(stream); -@@ -714,7 +714,7 @@ static void uv__write_req_finish(uv_write_t* req) { - uv_stream_t* stream = req->handle; - - /* Pop the req off tcp->write_queue. */ -- uv__queue_remove(&req->queue); -+ QUEUE_REMOVE(&req->queue); - - /* Only free when there was no error. On error, we touch up write_queue_size - * right before making the callback. The reason we don't do that right away -@@ -731,7 +731,7 @@ static void uv__write_req_finish(uv_write_t* req) { - /* Add it to the write_completed_queue where it will have its - * callback called in the near future. - */ -- uv__queue_insert_tail(&stream->write_completed_queue, &req->queue); -+ QUEUE_INSERT_TAIL(&stream->write_completed_queue, &req->queue); - uv__io_feed(stream->loop, &stream->io_watcher); - } - -@@ -837,7 +837,7 @@ static int uv__try_write(uv_stream_t* stream, - } - - static void uv__write(uv_stream_t* stream) { -- struct uv__queue* q; -+ QUEUE* q; - uv_write_t* req; - ssize_t n; - int count; -@@ -851,11 +851,11 @@ static void uv__write(uv_stream_t* stream) { - count = 32; - - for (;;) { -- if (uv__queue_empty(&stream->write_queue)) -+ if (QUEUE_EMPTY(&stream->write_queue)) - return; - -- q = uv__queue_head(&stream->write_queue); -- req = uv__queue_data(q, uv_write_t, queue); -+ q = QUEUE_HEAD(&stream->write_queue); -+ req = QUEUE_DATA(q, uv_write_t, queue); - assert(req->handle == stream); - - n = uv__try_write(stream, -@@ -899,19 +899,19 @@ error: - - static void uv__write_callbacks(uv_stream_t* stream) { - uv_write_t* req; -- struct uv__queue* q; -- struct uv__queue pq; -+ QUEUE* q; -+ QUEUE pq; - -- if (uv__queue_empty(&stream->write_completed_queue)) -+ if (QUEUE_EMPTY(&stream->write_completed_queue)) - return; - -- uv__queue_move(&stream->write_completed_queue, &pq); -+ QUEUE_MOVE(&stream->write_completed_queue, &pq); - -- while (!uv__queue_empty(&pq)) { -+ while (!QUEUE_EMPTY(&pq)) { - /* Pop a req off write_completed_queue. */ -- q = uv__queue_head(&pq); -- req = uv__queue_data(q, uv_write_t, queue); -- uv__queue_remove(q); -+ q = QUEUE_HEAD(&pq); -+ req = QUEUE_DATA(q, uv_write_t, queue); -+ QUEUE_REMOVE(q); - uv__req_unregister(stream->loop, req); - - if (req->bufs != NULL) { -@@ -1174,7 +1174,7 @@ int uv_shutdown(uv_shutdown_t* req, uv_stream_t* stream, uv_shutdown_cb cb) { - stream->shutdown_req = req; - stream->flags &= ~UV_HANDLE_WRITABLE; - -- if (uv__queue_empty(&stream->write_queue)) -+ if (QUEUE_EMPTY(&stream->write_queue)) - uv__io_feed(stream->loop, &stream->io_watcher); - - return 0; -@@ -1227,7 +1227,7 @@ static void uv__stream_io(uv_loop_t* loop, uv__io_t* w, unsigned int events) { - uv__write_callbacks(stream); - - /* Write queue drained. */ -- if (uv__queue_empty(&stream->write_queue)) -+ if (QUEUE_EMPTY(&stream->write_queue)) - uv__drain(stream); - } - } -@@ -1270,7 +1270,7 @@ static void uv__stream_connect(uv_stream_t* stream) { - stream->connect_req = NULL; - uv__req_unregister(stream->loop, req); - -- if (error < 0 || uv__queue_empty(&stream->write_queue)) { -+ if (error < 0 || QUEUE_EMPTY(&stream->write_queue)) { - uv__io_stop(stream->loop, &stream->io_watcher, POLLOUT); - } - -@@ -1352,7 +1352,7 @@ int uv_write2(uv_write_t* req, - req->handle = stream; - req->error = 0; - req->send_handle = send_handle; -- uv__queue_init(&req->queue); -+ QUEUE_INIT(&req->queue); - - req->bufs = req->bufsml; - if (nbufs > ARRAY_SIZE(req->bufsml)) -@@ -1367,7 +1367,7 @@ int uv_write2(uv_write_t* req, - stream->write_queue_size += uv__count_bufs(bufs, nbufs); - - /* Append the request to write_queue. */ -- uv__queue_insert_tail(&stream->write_queue, &req->queue); -+ QUEUE_INSERT_TAIL(&stream->write_queue, &req->queue); - - /* If the queue was empty when this function began, we should attempt to - * do the write immediately. Otherwise start the write_watcher and wait -diff --git a/deps/uv/src/unix/sunos.c b/deps/uv/src/unix/sunos.c -index 2d6bae79604..75b6fbad493 100644 ---- a/deps/uv/src/unix/sunos.c -+++ b/deps/uv/src/unix/sunos.c -@@ -148,7 +148,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - struct port_event events[1024]; - struct port_event* pe; - struct timespec spec; -- struct uv__queue* q; -+ QUEUE* q; - uv__io_t* w; - sigset_t* pset; - sigset_t set; -@@ -166,16 +166,16 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - int reset_timeout; - - if (loop->nfds == 0) { -- assert(uv__queue_empty(&loop->watcher_queue)); -+ assert(QUEUE_EMPTY(&loop->watcher_queue)); - return; - } - -- while (!uv__queue_empty(&loop->watcher_queue)) { -- q = uv__queue_head(&loop->watcher_queue); -- uv__queue_remove(q); -- uv__queue_init(q); -+ while (!QUEUE_EMPTY(&loop->watcher_queue)) { -+ q = QUEUE_HEAD(&loop->watcher_queue); -+ QUEUE_REMOVE(q); -+ QUEUE_INIT(q); - -- w = uv__queue_data(q, uv__io_t, watcher_queue); -+ w = QUEUE_DATA(q, uv__io_t, watcher_queue); - assert(w->pevents != 0); - - if (port_associate(loop->backend_fd, -@@ -316,8 +316,8 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - continue; /* Disabled by callback. */ - - /* Events Ports operates in oneshot mode, rearm timer on next run. */ -- if (w->pevents != 0 && uv__queue_empty(&w->watcher_queue)) -- uv__queue_insert_tail(&loop->watcher_queue, &w->watcher_queue); -+ if (w->pevents != 0 && QUEUE_EMPTY(&w->watcher_queue)) -+ QUEUE_INSERT_TAIL(&loop->watcher_queue, &w->watcher_queue); - } - - uv__metrics_inc_events(loop, nevents); -diff --git a/deps/uv/src/unix/tcp.c b/deps/uv/src/unix/tcp.c -index d6c848f4610..ab4e06c2f67 100644 ---- a/deps/uv/src/unix/tcp.c -+++ b/deps/uv/src/unix/tcp.c -@@ -124,7 +124,7 @@ int uv_tcp_init_ex(uv_loop_t* loop, uv_tcp_t* tcp, unsigned int flags) { - if (domain != AF_UNSPEC) { - err = new_socket(tcp, domain, 0); - if (err) { -- uv__queue_remove(&tcp->handle_queue); -+ QUEUE_REMOVE(&tcp->handle_queue); - if (tcp->io_watcher.fd != -1) - uv__close(tcp->io_watcher.fd); - tcp->io_watcher.fd = -1; -@@ -252,7 +252,7 @@ out: - uv__req_init(handle->loop, req, UV_CONNECT); - req->cb = cb; - req->handle = (uv_stream_t*) handle; -- uv__queue_init(&req->queue); -+ QUEUE_INIT(&req->queue); - handle->connect_req = req; - - uv__io_start(handle->loop, &handle->io_watcher, POLLOUT); -diff --git a/deps/uv/src/unix/tty.c b/deps/uv/src/unix/tty.c -index d099bdb3b67..7a5390c1a8b 100644 ---- a/deps/uv/src/unix/tty.c -+++ b/deps/uv/src/unix/tty.c -@@ -222,7 +222,7 @@ skip: - int rc = r; - if (newfd != -1) - uv__close(newfd); -- uv__queue_remove(&tty->handle_queue); -+ QUEUE_REMOVE(&tty->handle_queue); - do - r = fcntl(fd, F_SETFL, saved_flags); - while (r == -1 && errno == EINTR); -diff --git a/deps/uv/src/unix/udp.c b/deps/uv/src/unix/udp.c -index c2814512a5f..f556808fbae 100644 ---- a/deps/uv/src/unix/udp.c -+++ b/deps/uv/src/unix/udp.c -@@ -62,18 +62,18 @@ void uv__udp_close(uv_udp_t* handle) { - - void uv__udp_finish_close(uv_udp_t* handle) { - uv_udp_send_t* req; -- struct uv__queue* q; -+ QUEUE* q; - - assert(!uv__io_active(&handle->io_watcher, POLLIN | POLLOUT)); - assert(handle->io_watcher.fd == -1); - -- while (!uv__queue_empty(&handle->write_queue)) { -- q = uv__queue_head(&handle->write_queue); -- uv__queue_remove(q); -+ while (!QUEUE_EMPTY(&handle->write_queue)) { -+ q = QUEUE_HEAD(&handle->write_queue); -+ QUEUE_REMOVE(q); - -- req = uv__queue_data(q, uv_udp_send_t, queue); -+ req = QUEUE_DATA(q, uv_udp_send_t, queue); - req->status = UV_ECANCELED; -- uv__queue_insert_tail(&handle->write_completed_queue, &req->queue); -+ QUEUE_INSERT_TAIL(&handle->write_completed_queue, &req->queue); - } - - uv__udp_run_completed(handle); -@@ -90,16 +90,16 @@ void uv__udp_finish_close(uv_udp_t* handle) { - - static void uv__udp_run_completed(uv_udp_t* handle) { - uv_udp_send_t* req; -- struct uv__queue* q; -+ QUEUE* q; - - assert(!(handle->flags & UV_HANDLE_UDP_PROCESSING)); - handle->flags |= UV_HANDLE_UDP_PROCESSING; - -- while (!uv__queue_empty(&handle->write_completed_queue)) { -- q = uv__queue_head(&handle->write_completed_queue); -- uv__queue_remove(q); -+ while (!QUEUE_EMPTY(&handle->write_completed_queue)) { -+ q = QUEUE_HEAD(&handle->write_completed_queue); -+ QUEUE_REMOVE(q); - -- req = uv__queue_data(q, uv_udp_send_t, queue); -+ req = QUEUE_DATA(q, uv_udp_send_t, queue); - uv__req_unregister(handle->loop, req); - - handle->send_queue_size -= uv__count_bufs(req->bufs, req->nbufs); -@@ -121,7 +121,7 @@ static void uv__udp_run_completed(uv_udp_t* handle) { - req->send_cb(req, req->status); - } - -- if (uv__queue_empty(&handle->write_queue)) { -+ if (QUEUE_EMPTY(&handle->write_queue)) { - /* Pending queue and completion queue empty, stop watcher. */ - uv__io_stop(handle->loop, &handle->io_watcher, POLLOUT); - if (!uv__io_active(&handle->io_watcher, POLLIN)) -@@ -280,20 +280,20 @@ static void uv__udp_sendmsg(uv_udp_t* handle) { - uv_udp_send_t* req; - struct mmsghdr h[20]; - struct mmsghdr* p; -- struct uv__queue* q; -+ QUEUE* q; - ssize_t npkts; - size_t pkts; - size_t i; - -- if (uv__queue_empty(&handle->write_queue)) -+ if (QUEUE_EMPTY(&handle->write_queue)) - return; - - write_queue_drain: -- for (pkts = 0, q = uv__queue_head(&handle->write_queue); -+ for (pkts = 0, q = QUEUE_HEAD(&handle->write_queue); - pkts < ARRAY_SIZE(h) && q != &handle->write_queue; -- ++pkts, q = uv__queue_head(q)) { -+ ++pkts, q = QUEUE_HEAD(q)) { - assert(q != NULL); -- req = uv__queue_data(q, uv_udp_send_t, queue); -+ req = QUEUE_DATA(q, uv_udp_send_t, queue); - assert(req != NULL); - - p = &h[pkts]; -@@ -325,16 +325,16 @@ write_queue_drain: - if (npkts < 1) { - if (errno == EAGAIN || errno == EWOULDBLOCK || errno == ENOBUFS) - return; -- for (i = 0, q = uv__queue_head(&handle->write_queue); -+ for (i = 0, q = QUEUE_HEAD(&handle->write_queue); - i < pkts && q != &handle->write_queue; -- ++i, q = uv__queue_head(&handle->write_queue)) { -+ ++i, q = QUEUE_HEAD(&handle->write_queue)) { - assert(q != NULL); -- req = uv__queue_data(q, uv_udp_send_t, queue); -+ req = QUEUE_DATA(q, uv_udp_send_t, queue); - assert(req != NULL); - - req->status = UV__ERR(errno); -- uv__queue_remove(&req->queue); -- uv__queue_insert_tail(&handle->write_completed_queue, &req->queue); -+ QUEUE_REMOVE(&req->queue); -+ QUEUE_INSERT_TAIL(&handle->write_completed_queue, &req->queue); - } - uv__io_feed(handle->loop, &handle->io_watcher); - return; -@@ -343,11 +343,11 @@ write_queue_drain: - /* Safety: npkts known to be >0 below. Hence cast from ssize_t - * to size_t safe. - */ -- for (i = 0, q = uv__queue_head(&handle->write_queue); -+ for (i = 0, q = QUEUE_HEAD(&handle->write_queue); - i < (size_t)npkts && q != &handle->write_queue; -- ++i, q = uv__queue_head(&handle->write_queue)) { -+ ++i, q = QUEUE_HEAD(&handle->write_queue)) { - assert(q != NULL); -- req = uv__queue_data(q, uv_udp_send_t, queue); -+ req = QUEUE_DATA(q, uv_udp_send_t, queue); - assert(req != NULL); - - req->status = req->bufs[0].len; -@@ -357,25 +357,25 @@ write_queue_drain: - * why we don't handle partial writes. Just pop the request - * off the write queue and onto the completed queue, done. - */ -- uv__queue_remove(&req->queue); -- uv__queue_insert_tail(&handle->write_completed_queue, &req->queue); -+ QUEUE_REMOVE(&req->queue); -+ QUEUE_INSERT_TAIL(&handle->write_completed_queue, &req->queue); - } - - /* couldn't batch everything, continue sending (jump to avoid stack growth) */ -- if (!uv__queue_empty(&handle->write_queue)) -+ if (!QUEUE_EMPTY(&handle->write_queue)) - goto write_queue_drain; - uv__io_feed(handle->loop, &handle->io_watcher); - #else /* __linux__ || ____FreeBSD__ */ - uv_udp_send_t* req; - struct msghdr h; -- struct uv__queue* q; -+ QUEUE* q; - ssize_t size; - -- while (!uv__queue_empty(&handle->write_queue)) { -- q = uv__queue_head(&handle->write_queue); -+ while (!QUEUE_EMPTY(&handle->write_queue)) { -+ q = QUEUE_HEAD(&handle->write_queue); - assert(q != NULL); - -- req = uv__queue_data(q, uv_udp_send_t, queue); -+ req = QUEUE_DATA(q, uv_udp_send_t, queue); - assert(req != NULL); - - memset(&h, 0, sizeof h); -@@ -414,8 +414,8 @@ write_queue_drain: - * why we don't handle partial writes. Just pop the request - * off the write queue and onto the completed queue, done. - */ -- uv__queue_remove(&req->queue); -- uv__queue_insert_tail(&handle->write_completed_queue, &req->queue); -+ QUEUE_REMOVE(&req->queue); -+ QUEUE_INSERT_TAIL(&handle->write_completed_queue, &req->queue); - uv__io_feed(handle->loop, &handle->io_watcher); - } - #endif /* __linux__ || ____FreeBSD__ */ -@@ -729,7 +729,7 @@ int uv__udp_send(uv_udp_send_t* req, - memcpy(req->bufs, bufs, nbufs * sizeof(bufs[0])); - handle->send_queue_size += uv__count_bufs(req->bufs, req->nbufs); - handle->send_queue_count++; -- uv__queue_insert_tail(&handle->write_queue, &req->queue); -+ QUEUE_INSERT_TAIL(&handle->write_queue, &req->queue); - uv__handle_start(handle); - - if (empty_queue && !(handle->flags & UV_HANDLE_UDP_PROCESSING)) { -@@ -739,7 +739,7 @@ int uv__udp_send(uv_udp_send_t* req, - * away. In such cases the `io_watcher` has to be queued for asynchronous - * write. - */ -- if (!uv__queue_empty(&handle->write_queue)) -+ if (!QUEUE_EMPTY(&handle->write_queue)) - uv__io_start(handle->loop, &handle->io_watcher, POLLOUT); - } else { - uv__io_start(handle->loop, &handle->io_watcher, POLLOUT); -@@ -1007,8 +1007,8 @@ int uv__udp_init_ex(uv_loop_t* loop, - handle->send_queue_size = 0; - handle->send_queue_count = 0; - uv__io_init(&handle->io_watcher, uv__udp_io, fd); -- uv__queue_init(&handle->write_queue); -- uv__queue_init(&handle->write_completed_queue); -+ QUEUE_INIT(&handle->write_queue); -+ QUEUE_INIT(&handle->write_completed_queue); - - return 0; - } -diff --git a/deps/uv/src/uv-common.c b/deps/uv/src/uv-common.c -index 916f3f4e006..cec771fab21 100644 ---- a/deps/uv/src/uv-common.c -+++ b/deps/uv/src/uv-common.c -@@ -533,17 +533,17 @@ int uv_udp_recv_stop(uv_udp_t* handle) { - - - void uv_walk(uv_loop_t* loop, uv_walk_cb walk_cb, void* arg) { -- struct uv__queue queue; -- struct uv__queue* q; -+ QUEUE queue; -+ QUEUE* q; - uv_handle_t* h; - -- uv__queue_move(&loop->handle_queue, &queue); -- while (!uv__queue_empty(&queue)) { -- q = uv__queue_head(&queue); -- h = uv__queue_data(q, uv_handle_t, handle_queue); -+ QUEUE_MOVE(&loop->handle_queue, &queue); -+ while (!QUEUE_EMPTY(&queue)) { -+ q = QUEUE_HEAD(&queue); -+ h = QUEUE_DATA(q, uv_handle_t, handle_queue); - -- uv__queue_remove(q); -- uv__queue_insert_tail(&loop->handle_queue, q); -+ QUEUE_REMOVE(q); -+ QUEUE_INSERT_TAIL(&loop->handle_queue, q); - - if (h->flags & UV_HANDLE_INTERNAL) continue; - walk_cb(h, arg); -@@ -553,14 +553,14 @@ void uv_walk(uv_loop_t* loop, uv_walk_cb walk_cb, void* arg) { - - static void uv__print_handles(uv_loop_t* loop, int only_active, FILE* stream) { - const char* type; -- struct uv__queue* q; -+ QUEUE* q; - uv_handle_t* h; - - if (loop == NULL) - loop = uv_default_loop(); - -- uv__queue_foreach(q, &loop->handle_queue) { -- h = uv__queue_data(q, uv_handle_t, handle_queue); -+ QUEUE_FOREACH(q, &loop->handle_queue) { -+ h = QUEUE_DATA(q, uv_handle_t, handle_queue); - - if (only_active && !uv__is_active(h)) - continue; -@@ -846,7 +846,7 @@ uv_loop_t* uv_loop_new(void) { - - - int uv_loop_close(uv_loop_t* loop) { -- struct uv__queue* q; -+ QUEUE* q; - uv_handle_t* h; - #ifndef NDEBUG - void* saved_data; -@@ -855,8 +855,8 @@ int uv_loop_close(uv_loop_t* loop) { - if (uv__has_active_reqs(loop)) - return UV_EBUSY; - -- uv__queue_foreach(q, &loop->handle_queue) { -- h = uv__queue_data(q, uv_handle_t, handle_queue); -+ QUEUE_FOREACH(q, &loop->handle_queue) { -+ h = QUEUE_DATA(q, uv_handle_t, handle_queue); - if (!(h->flags & UV_HANDLE_INTERNAL)) - return UV_EBUSY; - } -diff --git a/deps/uv/src/uv-common.h b/deps/uv/src/uv-common.h -index cd57e5a3515..decde5362c8 100644 ---- a/deps/uv/src/uv-common.h -+++ b/deps/uv/src/uv-common.h -@@ -323,7 +323,7 @@ void uv__threadpool_cleanup(void); - (h)->loop = (loop_); \ - (h)->type = (type_); \ - (h)->flags = UV_HANDLE_REF; /* Ref the loop when active. */ \ -- uv__queue_insert_tail(&(loop_)->handle_queue, &(h)->handle_queue); \ -+ QUEUE_INSERT_TAIL(&(loop_)->handle_queue, &(h)->handle_queue); \ - uv__handle_platform_init(h); \ - } \ - while (0) -@@ -415,7 +415,6 @@ struct uv__iou { - size_t sqelen; - int ringfd; - uint32_t in_flight; -- uint32_t flags; - }; - #endif /* __linux__ */ - -diff --git a/deps/uv/src/win/core.c b/deps/uv/src/win/core.c -index e9885a0f1ff..9a3be58849a 100644 ---- a/deps/uv/src/win/core.c -+++ b/deps/uv/src/win/core.c -@@ -255,8 +255,8 @@ int uv_loop_init(uv_loop_t* loop) { - loop->time = 0; - uv_update_time(loop); - -- uv__queue_init(&loop->wq); -- uv__queue_init(&loop->handle_queue); -+ QUEUE_INIT(&loop->wq); -+ QUEUE_INIT(&loop->handle_queue); - loop->active_reqs.count = 0; - loop->active_handles = 0; - -@@ -358,7 +358,7 @@ void uv__loop_close(uv_loop_t* loop) { - } - - uv_mutex_lock(&loop->wq_mutex); -- assert(uv__queue_empty(&loop->wq) && "thread pool work queue not empty!"); -+ assert(QUEUE_EMPTY(&loop->wq) && "thread pool work queue not empty!"); - assert(!uv__has_active_reqs(loop)); - uv_mutex_unlock(&loop->wq_mutex); - uv_mutex_destroy(&loop->wq_mutex); -diff --git a/deps/uv/src/win/fs.c b/deps/uv/src/win/fs.c -index fc209c54f47..deb9438d689 100644 ---- a/deps/uv/src/win/fs.c -+++ b/deps/uv/src/win/fs.c -@@ -144,97 +144,26 @@ void uv__fs_init(void) { - } - - --static int32_t fs__decode_wtf8_char(const char** input) { -- uint32_t code_point; -- uint8_t b1; -- uint8_t b2; -- uint8_t b3; -- uint8_t b4; -- -- b1 = **input; -- if (b1 <= 0x7F) -- return b1; /* ASCII code point */ -- if (b1 < 0xC2) -- return -1; /* invalid: continuation byte */ -- code_point = b1; -- -- b2 = *++*input; -- if ((b2 & 0xC0) != 0x80) -- return -1; /* invalid: not a continuation byte */ -- code_point = (code_point << 6) | (b2 & 0x3F); -- if (b1 <= 0xDF) -- return 0x7FF & code_point; /* two-byte character */ -- -- b3 = *++*input; -- if ((b3 & 0xC0) != 0x80) -- return -1; /* invalid: not a continuation byte */ -- code_point = (code_point << 6) | (b3 & 0x3F); -- if (b1 <= 0xEF) -- return 0xFFFF & code_point; /* three-byte character */ -- -- b4 = *++*input; -- if ((b4 & 0xC0) != 0x80) -- return -1; /* invalid: not a continuation byte */ -- code_point = (code_point << 6) | (b4 & 0x3F); -- if (b1 <= 0xF4) -- if (code_point <= 0x10FFFF) -- return code_point; /* four-byte character */ -- -- /* code point too large */ -- return -1; --} -- -- --static ssize_t fs__get_length_wtf8(const char* source_ptr) { -- size_t w_target_len = 0; -- int32_t code_point; -- -- do { -- code_point = fs__decode_wtf8_char(&source_ptr); -- if (code_point < 0) -- return -1; -- if (code_point > 0xFFFF) -- w_target_len++; -- w_target_len++; -- } while (*source_ptr++); -- return w_target_len; --} -- -- --static void fs__wtf8_to_wide(const char* source_ptr, WCHAR* w_target) { -- int32_t code_point; -- -- do { -- code_point = fs__decode_wtf8_char(&source_ptr); -- /* fs__get_length_wtf8 should have been called and checked first. */ -- assert(code_point >= 0); -- if (code_point > 0x10000) { -- assert(code_point < 0x10FFFF); -- *w_target++ = (((code_point - 0x10000) >> 10) + 0xD800); -- *w_target++ = ((code_point - 0x10000) & 0x3FF) + 0xDC00; -- } else { -- *w_target++ = code_point; -- } -- } while (*source_ptr++); --} -- -- - INLINE static int fs__capture_path(uv_fs_t* req, const char* path, - const char* new_path, const int copy_path) { -- WCHAR* buf; -- WCHAR* pos; -- size_t buf_sz = 0; -- size_t path_len = 0; -- ssize_t pathw_len = 0; -- ssize_t new_pathw_len = 0; -+ char* buf; -+ char* pos; -+ ssize_t buf_sz = 0, path_len = 0, pathw_len = 0, new_pathw_len = 0; - - /* new_path can only be set if path is also set. */ - assert(new_path == NULL || path != NULL); - - if (path != NULL) { -- pathw_len = fs__get_length_wtf8(path); -- if (pathw_len < 0) -- return ERROR_INVALID_NAME; -+ pathw_len = MultiByteToWideChar(CP_UTF8, -+ 0, -+ path, -+ -1, -+ NULL, -+ 0); -+ if (pathw_len == 0) { -+ return GetLastError(); -+ } -+ - buf_sz += pathw_len * sizeof(WCHAR); - } - -@@ -244,9 +173,16 @@ INLINE static int fs__capture_path(uv_fs_t* req, const char* path, - } - - if (new_path != NULL) { -- new_pathw_len = fs__get_length_wtf8(new_path); -- if (new_pathw_len < 0) -- return ERROR_INVALID_NAME; -+ new_pathw_len = MultiByteToWideChar(CP_UTF8, -+ 0, -+ new_path, -+ -1, -+ NULL, -+ 0); -+ if (new_pathw_len == 0) { -+ return GetLastError(); -+ } -+ - buf_sz += new_pathw_len * sizeof(WCHAR); - } - -@@ -258,7 +194,7 @@ INLINE static int fs__capture_path(uv_fs_t* req, const char* path, - return 0; - } - -- buf = uv__malloc(buf_sz); -+ buf = (char*) uv__malloc(buf_sz); - if (buf == NULL) { - return ERROR_OUTOFMEMORY; - } -@@ -266,17 +202,29 @@ INLINE static int fs__capture_path(uv_fs_t* req, const char* path, - pos = buf; - - if (path != NULL) { -- fs__wtf8_to_wide(path, pos); -- req->file.pathw = pos; -- pos += pathw_len; -+ DWORD r = MultiByteToWideChar(CP_UTF8, -+ 0, -+ path, -+ -1, -+ (WCHAR*) pos, -+ pathw_len); -+ assert(r == (DWORD) pathw_len); -+ req->file.pathw = (WCHAR*) pos; -+ pos += r * sizeof(WCHAR); - } else { - req->file.pathw = NULL; - } - - if (new_path != NULL) { -- fs__wtf8_to_wide(new_path, pos); -- req->fs.info.new_pathw = pos; -- pos += new_pathw_len; -+ DWORD r = MultiByteToWideChar(CP_UTF8, -+ 0, -+ new_path, -+ -1, -+ (WCHAR*) pos, -+ new_pathw_len); -+ assert(r == (DWORD) new_pathw_len); -+ req->fs.info.new_pathw = (WCHAR*) pos; -+ pos += r * sizeof(WCHAR); - } else { - req->fs.info.new_pathw = NULL; - } -@@ -284,8 +232,8 @@ INLINE static int fs__capture_path(uv_fs_t* req, const char* path, - req->path = path; - if (path != NULL && copy_path) { - memcpy(pos, path, path_len); -- assert(path_len == buf_sz - (pos - buf) * sizeof(WCHAR)); -- req->path = (char*) pos; -+ assert(path_len == buf_sz - (pos - buf)); -+ req->path = pos; - } - - req->flags |= UV_FS_FREE_PATHS; -@@ -311,115 +259,57 @@ INLINE static void uv__fs_req_init(uv_loop_t* loop, uv_fs_t* req, - } - - --static int32_t fs__get_surrogate_value(const WCHAR* w_source_ptr, -- size_t w_source_len) { -- WCHAR u; -- WCHAR next; -- -- u = w_source_ptr[0]; -- if (u >= 0xD800 && u <= 0xDBFF && w_source_len > 1) { -- next = w_source_ptr[1]; -- if (next >= 0xDC00 && next <= 0xDFFF) -- return 0x10000 + ((u - 0xD800) << 10) + (next - 0xDC00); -- } -- return u; --} -- -- --static size_t fs__get_length_wide(const WCHAR* w_source_ptr, -- size_t w_source_len) { -- size_t target_len; -- int32_t code_point; -+static int fs__wide_to_utf8(WCHAR* w_source_ptr, -+ DWORD w_source_len, -+ char** target_ptr, -+ uint64_t* target_len_ptr) { -+ int r; -+ int target_len; -+ char* target; -+ target_len = WideCharToMultiByte(CP_UTF8, -+ 0, -+ w_source_ptr, -+ w_source_len, -+ NULL, -+ 0, -+ NULL, -+ NULL); - -- target_len = 0; -- for (; w_source_len; w_source_len--, w_source_ptr++) { -- code_point = fs__get_surrogate_value(w_source_ptr, w_source_len); -- /* Can be invalid UTF-8 but must be valid WTF-8. */ -- assert(code_point >= 0); -- if (code_point < 0x80) -- target_len += 1; -- else if (code_point < 0x800) -- target_len += 2; -- else if (code_point < 0x10000) -- target_len += 3; -- else { -- target_len += 4; -- w_source_ptr++; -- w_source_len--; -- } -+ if (target_len == 0) { -+ return -1; - } -- return target_len; --} - -- --static int fs__wide_to_wtf8(WCHAR* w_source_ptr, -- size_t w_source_len, -- char** target_ptr, -- size_t* target_len_ptr) { -- size_t target_len; -- char* target; -- int32_t code_point; -- -- /* If *target_ptr is provided, then *target_len_ptr must be its length -- * (excluding space for null), otherwise we will compute the target_len_ptr -- * length and may return a new allocation in *target_ptr if target_ptr is -- * provided. */ -- if (target_ptr == NULL || *target_ptr == NULL) { -- target_len = fs__get_length_wide(w_source_ptr, w_source_len); -- if (target_len_ptr != NULL) -- *target_len_ptr = target_len; -- } else { -- target_len = *target_len_ptr; -+ if (target_len_ptr != NULL) { -+ *target_len_ptr = target_len; - } - -- if (target_ptr == NULL) -+ if (target_ptr == NULL) { - return 0; -- -- if (*target_ptr == NULL) { -- target = uv__malloc(target_len + 1); -- if (target == NULL) { -- SetLastError(ERROR_OUTOFMEMORY); -- return -1; -- } -- *target_ptr = target; -- } else { -- target = *target_ptr; -- } -- -- for (; w_source_len; w_source_len--, w_source_ptr++) { -- code_point = fs__get_surrogate_value(w_source_ptr, w_source_len); -- /* Can be invalid UTF-8 but must be valid WTF-8. */ -- assert(code_point >= 0); -- -- if (code_point < 0x80) { -- *target++ = code_point; -- } else if (code_point < 0x800) { -- *target++ = 0xC0 | (code_point >> 6); -- *target++ = 0x80 | (code_point & 0x3F); -- } else if (code_point < 0x10000) { -- *target++ = 0xE0 | (code_point >> 12); -- *target++ = 0x80 | ((code_point >> 6) & 0x3F); -- *target++ = 0x80 | (code_point & 0x3F); -- } else { -- *target++ = 0xF0 | (code_point >> 18); -- *target++ = 0x80 | ((code_point >> 12) & 0x3F); -- *target++ = 0x80 | ((code_point >> 6) & 0x3F); -- *target++ = 0x80 | (code_point & 0x3F); -- w_source_ptr++; -- w_source_len--; -- } - } -- assert((size_t) (target - *target_ptr) == target_len); - -- *target++ = '\0'; -+ target = uv__malloc(target_len + 1); -+ if (target == NULL) { -+ SetLastError(ERROR_OUTOFMEMORY); -+ return -1; -+ } - -+ r = WideCharToMultiByte(CP_UTF8, -+ 0, -+ w_source_ptr, -+ w_source_len, -+ target, -+ target_len, -+ NULL, -+ NULL); -+ assert(r == target_len); -+ target[target_len] = '\0'; -+ *target_ptr = target; - return 0; - } - - --INLINE static int fs__readlink_handle(HANDLE handle, -- char** target_ptr, -- size_t* target_len_ptr) { -+INLINE static int fs__readlink_handle(HANDLE handle, char** target_ptr, -+ uint64_t* target_len_ptr) { - char buffer[MAXIMUM_REPARSE_DATA_BUFFER_SIZE]; - REPARSE_DATA_BUFFER* reparse_data = (REPARSE_DATA_BUFFER*) buffer; - WCHAR* w_target; -@@ -549,8 +439,7 @@ INLINE static int fs__readlink_handle(HANDLE handle, - return -1; - } - -- assert(target_ptr == NULL || *target_ptr == NULL); -- return fs__wide_to_wtf8(w_target, w_target_len, target_ptr, target_len_ptr); -+ return fs__wide_to_utf8(w_target, w_target_len, target_ptr, target_len_ptr); - } - - -@@ -1540,8 +1429,7 @@ void fs__scandir(uv_fs_t* req) { - uv__dirent_t* dirent; - - size_t wchar_len; -- size_t wtf8_len; -- char* wtf8; -+ size_t utf8_len; - - /* Obtain a pointer to the current directory entry. */ - position += next_entry_offset; -@@ -1568,8 +1456,11 @@ void fs__scandir(uv_fs_t* req) { - info->FileName[1] == L'.') - continue; - -- /* Compute the space required to store the filename as WTF-8. */ -- wtf8_len = fs__get_length_wide(&info->FileName[0], wchar_len); -+ /* Compute the space required to store the filename as UTF-8. */ -+ utf8_len = WideCharToMultiByte( -+ CP_UTF8, 0, &info->FileName[0], wchar_len, NULL, 0, NULL, NULL); -+ if (utf8_len == 0) -+ goto win32_error; - - /* Resize the dirent array if needed. */ - if (dirents_used >= dirents_size) { -@@ -1589,17 +1480,26 @@ void fs__scandir(uv_fs_t* req) { - * includes room for the first character of the filename, but `utf8_len` - * doesn't count the NULL terminator at this point. - */ -- dirent = uv__malloc(sizeof *dirent + wtf8_len); -+ dirent = uv__malloc(sizeof *dirent + utf8_len); - if (dirent == NULL) - goto out_of_memory_error; - - dirents[dirents_used++] = dirent; - - /* Convert file name to UTF-8. */ -- wtf8 = &dirent->d_name[0]; -- if (fs__wide_to_wtf8(&info->FileName[0], wchar_len, &wtf8, &wtf8_len) == -1) -+ if (WideCharToMultiByte(CP_UTF8, -+ 0, -+ &info->FileName[0], -+ wchar_len, -+ &dirent->d_name[0], -+ utf8_len, -+ NULL, -+ NULL) == 0) - goto win32_error; - -+ /* Add a null terminator to the filename. */ -+ dirent->d_name[utf8_len] = '\0'; -+ - /* Fill out the type field. */ - if (info->FileAttributes & FILE_ATTRIBUTE_DEVICE) - dirent->d_type = UV__DT_CHAR; -@@ -1808,7 +1708,6 @@ void fs__closedir(uv_fs_t* req) { - - INLINE static int fs__stat_handle(HANDLE handle, uv_stat_t* statbuf, - int do_lstat) { -- size_t target_length = 0; - FILE_FS_DEVICE_INFORMATION device_info; - FILE_ALL_INFORMATION file_info; - FILE_FS_VOLUME_INFORMATION volume_info; -@@ -1904,10 +1803,9 @@ INLINE static int fs__stat_handle(HANDLE handle, uv_stat_t* statbuf, - * to be treated as a regular file. The higher level lstat function will - * detect this failure and retry without do_lstat if appropriate. - */ -- if (fs__readlink_handle(handle, NULL, &target_length) != 0) -+ if (fs__readlink_handle(handle, NULL, &statbuf->st_size) != 0) - return -1; - statbuf->st_mode |= S_IFLNK; -- statbuf->st_size = target_length; - } - - if (statbuf->st_mode == 0) { -@@ -2763,7 +2661,6 @@ static void fs__readlink(uv_fs_t* req) { - return; - } - -- assert(req->ptr == NULL); - if (fs__readlink_handle(handle, (char**) &req->ptr, NULL) != 0) { - DWORD error = GetLastError(); - SET_REQ_WIN32_ERROR(req, error); -@@ -2823,8 +2720,7 @@ static ssize_t fs__realpath_handle(HANDLE handle, char** realpath_ptr) { - return -1; - } - -- assert(*realpath_ptr == NULL); -- r = fs__wide_to_wtf8(w_realpath_ptr, w_realpath_len, realpath_ptr, NULL); -+ r = fs__wide_to_utf8(w_realpath_ptr, w_realpath_len, realpath_ptr, NULL); - uv__free(w_realpath_buf); - return r; - } -@@ -2844,7 +2740,6 @@ static void fs__realpath(uv_fs_t* req) { - return; - } - -- assert(req->ptr == NULL); - if (fs__realpath_handle(handle, (char**) &req->ptr) == -1) { - CloseHandle(handle); - SET_REQ_WIN32_ERROR(req, GetLastError()); -diff --git a/deps/uv/src/win/handle-inl.h b/deps/uv/src/win/handle-inl.h -index 4722e85790a..5c843c241ef 100644 ---- a/deps/uv/src/win/handle-inl.h -+++ b/deps/uv/src/win/handle-inl.h -@@ -75,7 +75,7 @@ - - #define uv__handle_close(handle) \ - do { \ -- uv__queue_remove(&(handle)->handle_queue); \ -+ QUEUE_REMOVE(&(handle)->handle_queue); \ - uv__active_handle_rm((uv_handle_t*) (handle)); \ - \ - (handle)->flags |= UV_HANDLE_CLOSED; \ -diff --git a/deps/uv/src/win/pipe.c b/deps/uv/src/win/pipe.c -index f0cac382256..5e4276387ac 100644 ---- a/deps/uv/src/win/pipe.c -+++ b/deps/uv/src/win/pipe.c -@@ -55,7 +55,7 @@ static const int pipe_prefix_len = sizeof(pipe_prefix) - 1; - typedef struct { - uv__ipc_socket_xfer_type_t xfer_type; - uv__ipc_socket_xfer_info_t xfer_info; -- struct uv__queue member; -+ QUEUE member; - } uv__ipc_xfer_queue_item_t; - - /* IPC frame header flags. */ -@@ -111,7 +111,7 @@ int uv_pipe_init(uv_loop_t* loop, uv_pipe_t* handle, int ipc) { - handle->name = NULL; - handle->pipe.conn.ipc_remote_pid = 0; - handle->pipe.conn.ipc_data_frame.payload_remaining = 0; -- uv__queue_init(&handle->pipe.conn.ipc_xfer_queue); -+ QUEUE_INIT(&handle->pipe.conn.ipc_xfer_queue); - handle->pipe.conn.ipc_xfer_queue_length = 0; - handle->ipc = ipc; - handle->pipe.conn.non_overlapped_writes_tail = NULL; -@@ -637,13 +637,13 @@ void uv__pipe_endgame(uv_loop_t* loop, uv_pipe_t* handle) { - - if (handle->flags & UV_HANDLE_CONNECTION) { - /* Free pending sockets */ -- while (!uv__queue_empty(&handle->pipe.conn.ipc_xfer_queue)) { -- struct uv__queue* q; -+ while (!QUEUE_EMPTY(&handle->pipe.conn.ipc_xfer_queue)) { -+ QUEUE* q; - SOCKET socket; - -- q = uv__queue_head(&handle->pipe.conn.ipc_xfer_queue); -- uv__queue_remove(q); -- xfer_queue_item = uv__queue_data(q, uv__ipc_xfer_queue_item_t, member); -+ q = QUEUE_HEAD(&handle->pipe.conn.ipc_xfer_queue); -+ QUEUE_REMOVE(q); -+ xfer_queue_item = QUEUE_DATA(q, uv__ipc_xfer_queue_item_t, member); - - /* Materialize socket and close it */ - socket = WSASocketW(FROM_PROTOCOL_INFO, -@@ -1124,20 +1124,20 @@ int uv__pipe_accept(uv_pipe_t* server, uv_stream_t* client) { - uv_loop_t* loop = server->loop; - uv_pipe_t* pipe_client; - uv_pipe_accept_t* req; -- struct uv__queue* q; -+ QUEUE* q; - uv__ipc_xfer_queue_item_t* item; - int err; - - if (server->ipc) { -- if (uv__queue_empty(&server->pipe.conn.ipc_xfer_queue)) { -+ if (QUEUE_EMPTY(&server->pipe.conn.ipc_xfer_queue)) { - /* No valid pending sockets. */ - return WSAEWOULDBLOCK; - } - -- q = uv__queue_head(&server->pipe.conn.ipc_xfer_queue); -- uv__queue_remove(q); -+ q = QUEUE_HEAD(&server->pipe.conn.ipc_xfer_queue); -+ QUEUE_REMOVE(q); - server->pipe.conn.ipc_xfer_queue_length--; -- item = uv__queue_data(q, uv__ipc_xfer_queue_item_t, member); -+ item = QUEUE_DATA(q, uv__ipc_xfer_queue_item_t, member); - - err = uv__tcp_xfer_import( - (uv_tcp_t*) client, item->xfer_type, &item->xfer_info); -@@ -1891,7 +1891,7 @@ static void uv__pipe_queue_ipc_xfer_info( - item->xfer_type = xfer_type; - item->xfer_info = *xfer_info; - -- uv__queue_insert_tail(&handle->pipe.conn.ipc_xfer_queue, &item->member); -+ QUEUE_INSERT_TAIL(&handle->pipe.conn.ipc_xfer_queue, &item->member); - handle->pipe.conn.ipc_xfer_queue_length++; - } - -diff --git a/deps/uv/src/win/tcp.c b/deps/uv/src/win/tcp.c -index 187f36e2a61..6b282e0b501 100644 ---- a/deps/uv/src/win/tcp.c -+++ b/deps/uv/src/win/tcp.c -@@ -175,14 +175,14 @@ int uv_tcp_init_ex(uv_loop_t* loop, uv_tcp_t* handle, unsigned int flags) { - sock = socket(domain, SOCK_STREAM, 0); - if (sock == INVALID_SOCKET) { - err = WSAGetLastError(); -- uv__queue_remove(&handle->handle_queue); -+ QUEUE_REMOVE(&handle->handle_queue); - return uv_translate_sys_error(err); - } - - err = uv__tcp_set_socket(handle->loop, handle, sock, domain, 0); - if (err) { - closesocket(sock); -- uv__queue_remove(&handle->handle_queue); -+ QUEUE_REMOVE(&handle->handle_queue); - return uv_translate_sys_error(err); - } - -diff --git a/deps/uv/src/win/udp.c b/deps/uv/src/win/udp.c -index eab53842d4f..8a982d1907d 100644 ---- a/deps/uv/src/win/udp.c -+++ b/deps/uv/src/win/udp.c -@@ -146,14 +146,14 @@ int uv__udp_init_ex(uv_loop_t* loop, - sock = socket(domain, SOCK_DGRAM, 0); - if (sock == INVALID_SOCKET) { - err = WSAGetLastError(); -- uv__queue_remove(&handle->handle_queue); -+ QUEUE_REMOVE(&handle->handle_queue); - return uv_translate_sys_error(err); - } - - err = uv__udp_set_socket(handle->loop, handle, sock, domain); - if (err) { - closesocket(sock); -- uv__queue_remove(&handle->handle_queue); -+ QUEUE_REMOVE(&handle->handle_queue); - return uv_translate_sys_error(err); - } - } diff --git a/meta-oe/recipes-devtools/nodejs/nodejs/0001-deps-disable-io_uring-support-in-libuv.patch b/meta-oe/recipes-devtools/nodejs/nodejs/0001-deps-disable-io_uring-support-in-libuv.patch new file mode 100644 index 000000000..f11cd08cb --- /dev/null +++ b/meta-oe/recipes-devtools/nodejs/nodejs/0001-deps-disable-io_uring-support-in-libuv.patch @@ -0,0 +1,35 @@ +From 9838be9c710ab4249df86726fa390232a3b6a6e7 Mon Sep 17 00:00:00 2001 +From: Changqing Li +Date: Fri, 1 Mar 2024 15:46:11 +0800 +Subject: [PATCH] deps: disable io_uring support in libuv + +Refer [1], Pseudo fails to intercept some of the syscalls when io_uring +enabled. Refer [2], always disable io_uring support in libuv to fix +issue in [1]. + +[1] https://git.openembedded.org/meta-openembedded/commit/?id=d08453978c31ee41d28206c6ff198d7d9d701d88 +[2] https://github.com/nodejs/node/commit/686da19abb + +Upstream-Status: Inappropriate [oe specific] + +Signed-off-by: Changqing Li +--- + deps/uv/src/unix/linux.c | 2 +- + 1 file changed, 1 insertion(+), 1 deletion(-) + +diff --git a/deps/uv/src/unix/linux.c b/deps/uv/src/unix/linux.c +index 0c997185..7508409d 100644 +--- a/deps/uv/src/unix/linux.c ++++ b/deps/uv/src/unix/linux.c +@@ -433,7 +433,7 @@ static int uv__use_io_uring(void) { + if (use == 0) { + /* Disable io_uring by default due to CVE-2024-22017. */ + val = getenv("UV_USE_IO_URING"); +- use = val != NULL && atoi(val) ? 1 : -1; ++ use = 0; + atomic_store_explicit(&use_io_uring, use, memory_order_relaxed); + } + +-- +2.25.1 + diff --git a/meta-oe/recipes-devtools/nodejs/nodejs/0002-Revert-io_uring-changes-from-libuv-1.45.0.patch b/meta-oe/recipes-devtools/nodejs/nodejs/0002-Revert-io_uring-changes-from-libuv-1.45.0.patch deleted file mode 100644 index 77cd53b75..000000000 --- a/meta-oe/recipes-devtools/nodejs/nodejs/0002-Revert-io_uring-changes-from-libuv-1.45.0.patch +++ /dev/null @@ -1,1803 +0,0 @@ -From 6d2ef4c8ba2304ee4941a6719b3ad9bd63e415a9 Mon Sep 17 00:00:00 2001 -From: Martin Jansa -Date: Wed, 18 Oct 2023 21:09:44 +0200 -Subject: [PATCH] Revert io_uring changes from libuv-1.45.0 - -This reverts https://github.com/libuv/libuv/pull/3952/commits/26c79a942b92573a1388c0ee8a6ad4397f009318 - -Included in nodejs-20.3.0 with the libuv upgrade to 1.45.0 in: -https://github.com/nodejs/node/commit/bfcb3d1d9a876f399013d326bd65804f9eda77e4 - -Reverted libuv commits: -Revert "linux: fix WRITEV with lots of bufs using io_uring (#4004)" -This reverts commit ef6a9a624df0a00687037474025a3608472f722a. -Revert "linux: work around EOWNERDEAD io_uring kernel bug (#4002)" -This reverts commit d23a20f62cc50b9fd7694992263f1d296d8f5cb4. -Revert "unix: handle CQ overflow in iou ring (#3991)" -This reverts commit 30fc896cc1b5822e9f1eb462587fe4b368a6215c. -Revert "unix: constrained_memory should return UINT64_MAX (#3753)" -This reverts commit 6ad347fae4520f39520d34bd7c7f5ddafab13a69. -Revert "linux: use io_uring to batch epoll_ctl calls (#3979)" -This reverts commit 6e073ef5daf93b708a654008959b823b58029e88. -Revert "linux: fix logic bug in sqe ring space check (#3980)" -This reverts commit f27208224084fc972b9d2802486d97ef31b51a39. -Revert "src: fix events/events_waiting metrics counter (#3957)" -This reverts commit e02642cf3b768b2c58a41f97fa38507e032ae415. -Revert "linux: remove bug workaround for obsolete kernels (#3965)" -This reverts commit 1c935a34454167b23f8eef7f0f63d7119f0de747. -Revert "linux: add IORING_OP_CLOSE support (#3964)" -This reverts commit dfae365f844e127621128a76bce7165e3f99a8d9. -Revert "linux: add IORING_OP_OPENAT support (#3963)" -This reverts commit 5ca5e475bb1711e65323ef1594a31818e5a1a9eb. -Revert "linux: fix academic valgrind warning (#3960)" -This reverts commit a7ff759ca1deacb2e0e6ae3c2d3dce91cc637dfe. -Revert "linux: introduce io_uring support (#3952)" -This reverts commit d2c31f429b87b476a7f1344d145dad4752a406d4. - -Dropped deps/uv/docs deps/uv/test changes as these dirs aren't included -in nodejs tarballs. - -Signed-off-by: Martin Jansa ---- -Upstream-Status: Inappropriate [OE specific] - - deps/uv/docs/src/fs.rst | 6 - - deps/uv/docs/src/misc.rst | 5 +- - deps/uv/src/threadpool.c | 21 - - deps/uv/src/unix/aix.c | 11 +- - deps/uv/src/unix/fs.c | 57 +- - deps/uv/src/unix/internal.h | 23 - - deps/uv/src/unix/kqueue.c | 10 +- - deps/uv/src/unix/linux.c | 998 +++----------------------- - deps/uv/src/unix/os390.c | 11 +- - deps/uv/src/unix/posix-poll.c | 11 +- - deps/uv/src/uv-common.h | 28 - - deps/uv/src/win/core.c | 20 +- - deps/uv/test/test-fs.c | 33 - - deps/uv/test/test-list.h | 4 - - deps/uv/test/test-metrics.c | 151 ---- - deps/uv/test/test-threadpool-cancel.c | 34 +- - 16 files changed, 138 insertions(+), 1285 deletions(-) - -diff --git a/deps/uv/src/threadpool.c b/deps/uv/src/threadpool.c -index 51962bf0021..a3da53026f9 100644 ---- a/deps/uv/src/threadpool.c -+++ b/deps/uv/src/threadpool.c -@@ -275,13 +275,9 @@ void uv__work_submit(uv_loop_t* loop, - } - - --/* TODO(bnoordhuis) teach libuv how to cancel file operations -- * that go through io_uring instead of the thread pool. -- */ - static int uv__work_cancel(uv_loop_t* loop, uv_req_t* req, struct uv__work* w) { - int cancelled; - -- uv_once(&once, init_once); /* Ensure |mutex| is initialized. */ - uv_mutex_lock(&mutex); - uv_mutex_lock(&w->loop->wq_mutex); - -@@ -311,15 +307,12 @@ void uv__work_done(uv_async_t* handle) { - QUEUE* q; - QUEUE wq; - int err; -- int nevents; - - loop = container_of(handle, uv_loop_t, wq_async); - uv_mutex_lock(&loop->wq_mutex); - QUEUE_MOVE(&loop->wq, &wq); - uv_mutex_unlock(&loop->wq_mutex); - -- nevents = 0; -- - while (!QUEUE_EMPTY(&wq)) { - q = QUEUE_HEAD(&wq); - QUEUE_REMOVE(q); -@@ -327,20 +320,6 @@ void uv__work_done(uv_async_t* handle) { - w = container_of(q, struct uv__work, wq); - err = (w->work == uv__cancelled) ? UV_ECANCELED : 0; - w->done(w, err); -- nevents++; -- } -- -- /* This check accomplishes 2 things: -- * 1. Even if the queue was empty, the call to uv__work_done() should count -- * as an event. Which will have been added by the event loop when -- * calling this callback. -- * 2. Prevents accidental wrap around in case nevents == 0 events == 0. -- */ -- if (nevents > 1) { -- /* Subtract 1 to counter the call to uv__work_done(). */ -- uv__metrics_inc_events(loop, nevents - 1); -- if (uv__get_internal_fields(loop)->current_timeout == 0) -- uv__metrics_inc_events_waiting(loop, nevents - 1); - } - } - -diff --git a/deps/uv/src/unix/aix.c b/deps/uv/src/unix/aix.c -index f1afbed49ec..b855282ebc8 100644 ---- a/deps/uv/src/unix/aix.c -+++ b/deps/uv/src/unix/aix.c -@@ -131,7 +131,6 @@ int uv__io_check_fd(uv_loop_t* loop, int fd) { - - - void uv__io_poll(uv_loop_t* loop, int timeout) { -- uv__loop_internal_fields_t* lfields; - struct pollfd events[1024]; - struct pollfd pqry; - struct pollfd* pe; -@@ -155,8 +154,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - return; - } - -- lfields = uv__get_internal_fields(loop); -- - while (!QUEUE_EMPTY(&loop->watcher_queue)) { - q = QUEUE_HEAD(&loop->watcher_queue); - QUEUE_REMOVE(q); -@@ -220,7 +217,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - base = loop->time; - count = 48; /* Benchmarks suggest this gives the best throughput. */ - -- if (lfields->flags & UV_METRICS_IDLE_TIME) { -+ if (uv__get_internal_fields(loop)->flags & UV_METRICS_IDLE_TIME) { - reset_timeout = 1; - user_timeout = timeout; - timeout = 0; -@@ -235,12 +232,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - if (timeout != 0) - uv__metrics_set_provider_entry_time(loop); - -- /* Store the current timeout in a location that's globally accessible so -- * other locations like uv__work_done() can determine whether the queue -- * of events in the callback were waiting when poll was called. -- */ -- lfields->current_timeout = timeout; -- - nfds = pollset_poll(loop->backend_fd, - events, - ARRAY_SIZE(events), -diff --git a/deps/uv/src/unix/fs.c b/deps/uv/src/unix/fs.c -index 00d385c24b7..c696a8d5500 100644 ---- a/deps/uv/src/unix/fs.c -+++ b/deps/uv/src/unix/fs.c -@@ -61,11 +61,10 @@ - #endif - - #if defined(__linux__) --# include --# include -+# include "sys/utsname.h" - #endif - --#if defined(__sun) -+#if defined(__linux__) || defined(__sun) - # include - # include - #endif -@@ -1554,7 +1553,26 @@ static int uv__fs_statx(int fd, - return UV_ENOSYS; - } - -- uv__statx_to_stat(&statxbuf, buf); -+ buf->st_dev = makedev(statxbuf.stx_dev_major, statxbuf.stx_dev_minor); -+ buf->st_mode = statxbuf.stx_mode; -+ buf->st_nlink = statxbuf.stx_nlink; -+ buf->st_uid = statxbuf.stx_uid; -+ buf->st_gid = statxbuf.stx_gid; -+ buf->st_rdev = makedev(statxbuf.stx_rdev_major, statxbuf.stx_rdev_minor); -+ buf->st_ino = statxbuf.stx_ino; -+ buf->st_size = statxbuf.stx_size; -+ buf->st_blksize = statxbuf.stx_blksize; -+ buf->st_blocks = statxbuf.stx_blocks; -+ buf->st_atim.tv_sec = statxbuf.stx_atime.tv_sec; -+ buf->st_atim.tv_nsec = statxbuf.stx_atime.tv_nsec; -+ buf->st_mtim.tv_sec = statxbuf.stx_mtime.tv_sec; -+ buf->st_mtim.tv_nsec = statxbuf.stx_mtime.tv_nsec; -+ buf->st_ctim.tv_sec = statxbuf.stx_ctime.tv_sec; -+ buf->st_ctim.tv_nsec = statxbuf.stx_ctime.tv_nsec; -+ buf->st_birthtim.tv_sec = statxbuf.stx_btime.tv_sec; -+ buf->st_birthtim.tv_nsec = statxbuf.stx_btime.tv_nsec; -+ buf->st_flags = 0; -+ buf->st_gen = 0; - - return 0; - #else -@@ -1798,9 +1816,6 @@ int uv_fs_chown(uv_loop_t* loop, - int uv_fs_close(uv_loop_t* loop, uv_fs_t* req, uv_file file, uv_fs_cb cb) { - INIT(CLOSE); - req->file = file; -- if (cb != NULL) -- if (uv__iou_fs_close(loop, req)) -- return 0; - POST; - } - -@@ -1848,9 +1863,6 @@ int uv_fs_lchown(uv_loop_t* loop, - int uv_fs_fdatasync(uv_loop_t* loop, uv_fs_t* req, uv_file file, uv_fs_cb cb) { - INIT(FDATASYNC); - req->file = file; -- if (cb != NULL) -- if (uv__iou_fs_fsync_or_fdatasync(loop, req, /* IORING_FSYNC_DATASYNC */ 1)) -- return 0; - POST; - } - -@@ -1858,9 +1870,6 @@ int uv_fs_fdatasync(uv_loop_t* loop, uv_fs_t* req, uv_file file, uv_fs_cb cb) { - int uv_fs_fstat(uv_loop_t* loop, uv_fs_t* req, uv_file file, uv_fs_cb cb) { - INIT(FSTAT); - req->file = file; -- if (cb != NULL) -- if (uv__iou_fs_statx(loop, req, /* is_fstat */ 1, /* is_lstat */ 0)) -- return 0; - POST; - } - -@@ -1868,9 +1877,6 @@ int uv_fs_fstat(uv_loop_t* loop, uv_fs_t* req, uv_file file, uv_fs_cb cb) { - int uv_fs_fsync(uv_loop_t* loop, uv_fs_t* req, uv_file file, uv_fs_cb cb) { - INIT(FSYNC); - req->file = file; -- if (cb != NULL) -- if (uv__iou_fs_fsync_or_fdatasync(loop, req, /* no flags */ 0)) -- return 0; - POST; - } - -@@ -1917,9 +1923,6 @@ int uv_fs_lutime(uv_loop_t* loop, - int uv_fs_lstat(uv_loop_t* loop, uv_fs_t* req, const char* path, uv_fs_cb cb) { - INIT(LSTAT); - PATH; -- if (cb != NULL) -- if (uv__iou_fs_statx(loop, req, /* is_fstat */ 0, /* is_lstat */ 1)) -- return 0; - POST; - } - -@@ -1981,9 +1984,6 @@ int uv_fs_open(uv_loop_t* loop, - PATH; - req->flags = flags; - req->mode = mode; -- if (cb != NULL) -- if (uv__iou_fs_open(loop, req)) -- return 0; - POST; - } - -@@ -2012,11 +2012,6 @@ int uv_fs_read(uv_loop_t* loop, uv_fs_t* req, - memcpy(req->bufs, bufs, nbufs * sizeof(*bufs)); - - req->off = off; -- -- if (cb != NULL) -- if (uv__iou_fs_read_or_write(loop, req, /* is_read */ 1)) -- return 0; -- - POST; - } - -@@ -2124,9 +2119,6 @@ int uv_fs_sendfile(uv_loop_t* loop, - int uv_fs_stat(uv_loop_t* loop, uv_fs_t* req, const char* path, uv_fs_cb cb) { - INIT(STAT); - PATH; -- if (cb != NULL) -- if (uv__iou_fs_statx(loop, req, /* is_fstat */ 0, /* is_lstat */ 0)) -- return 0; - POST; - } - -@@ -2190,11 +2182,6 @@ int uv_fs_write(uv_loop_t* loop, - memcpy(req->bufs, bufs, nbufs * sizeof(*bufs)); - - req->off = off; -- -- if (cb != NULL) -- if (uv__iou_fs_read_or_write(loop, req, /* is_read */ 0)) -- return 0; -- - POST; - } - -diff --git a/deps/uv/src/unix/internal.h b/deps/uv/src/unix/internal.h -index 6c5822e6a0d..d439ae6dd3d 100644 ---- a/deps/uv/src/unix/internal.h -+++ b/deps/uv/src/unix/internal.h -@@ -329,28 +329,6 @@ int uv__random_getentropy(void* buf, size_t buflen); - int uv__random_readpath(const char* path, void* buf, size_t buflen); - int uv__random_sysctl(void* buf, size_t buflen); - --/* io_uring */ --#ifdef __linux__ --int uv__iou_fs_close(uv_loop_t* loop, uv_fs_t* req); --int uv__iou_fs_fsync_or_fdatasync(uv_loop_t* loop, -- uv_fs_t* req, -- uint32_t fsync_flags); --int uv__iou_fs_open(uv_loop_t* loop, uv_fs_t* req); --int uv__iou_fs_read_or_write(uv_loop_t* loop, -- uv_fs_t* req, -- int is_read); --int uv__iou_fs_statx(uv_loop_t* loop, -- uv_fs_t* req, -- int is_fstat, -- int is_lstat); --#else --#define uv__iou_fs_close(loop, req) 0 --#define uv__iou_fs_fsync_or_fdatasync(loop, req, fsync_flags) 0 --#define uv__iou_fs_open(loop, req) 0 --#define uv__iou_fs_read_or_write(loop, req, is_read) 0 --#define uv__iou_fs_statx(loop, req, is_fstat, is_lstat) 0 --#endif -- - #if defined(__APPLE__) - int uv___stream_fd(const uv_stream_t* handle); - #define uv__stream_fd(handle) (uv___stream_fd((const uv_stream_t*) (handle))) -@@ -427,7 +405,6 @@ int uv__statx(int dirfd, - int flags, - unsigned int mask, - struct uv__statx* statxbuf); --void uv__statx_to_stat(const struct uv__statx* statxbuf, uv_stat_t* buf); - ssize_t uv__getrandom(void* buf, size_t buflen, unsigned flags); - #endif - -diff --git a/deps/uv/src/unix/kqueue.c b/deps/uv/src/unix/kqueue.c -index 82916d65933..deb486bae7a 100644 ---- a/deps/uv/src/unix/kqueue.c -+++ b/deps/uv/src/unix/kqueue.c -@@ -127,7 +127,6 @@ static void uv__kqueue_delete(int kqfd, const struct kevent *ev) { - - - void uv__io_poll(uv_loop_t* loop, int timeout) { -- uv__loop_internal_fields_t* lfields; - struct kevent events[1024]; - struct kevent* ev; - struct timespec spec; -@@ -156,7 +155,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - return; - } - -- lfields = uv__get_internal_fields(loop); - nevents = 0; - - while (!QUEUE_EMPTY(&loop->watcher_queue)) { -@@ -224,7 +222,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - base = loop->time; - count = 48; /* Benchmarks suggest this gives the best throughput. */ - -- if (lfields->flags & UV_METRICS_IDLE_TIME) { -+ if (uv__get_internal_fields(loop)->flags & UV_METRICS_IDLE_TIME) { - reset_timeout = 1; - user_timeout = timeout; - timeout = 0; -@@ -247,12 +245,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - if (pset != NULL) - pthread_sigmask(SIG_BLOCK, pset, NULL); - -- /* Store the current timeout in a location that's globally accessible so -- * other locations like uv__work_done() can determine whether the queue -- * of events in the callback were waiting when poll was called. -- */ -- lfields->current_timeout = timeout; -- - nfds = kevent(loop->backend_fd, - events, - nevents, -diff --git a/deps/uv/src/unix/linux.c b/deps/uv/src/unix/linux.c -index 5f84ad0eea3..343e37f2527 100644 ---- a/deps/uv/src/unix/linux.c -+++ b/deps/uv/src/unix/linux.c -@@ -27,8 +27,6 @@ - #include "internal.h" - - #include --#include --#include /* offsetof */ - #include - #include - #include -@@ -40,29 +38,15 @@ - #include - #include - #include --#include - #include - #include - #include - #include - #include --#include - #include - #include - #include - --#ifndef __NR_io_uring_setup --# define __NR_io_uring_setup 425 --#endif -- --#ifndef __NR_io_uring_enter --# define __NR_io_uring_enter 426 --#endif -- --#ifndef __NR_io_uring_register --# define __NR_io_uring_register 427 --#endif -- - #ifndef __NR_copy_file_range - # if defined(__x86_64__) - # define __NR_copy_file_range 326 -@@ -132,129 +116,6 @@ - # include - #endif /* HAVE_IFADDRS_H */ - --enum { -- UV__IORING_SETUP_SQPOLL = 2u, --}; -- --enum { -- UV__IORING_FEAT_SINGLE_MMAP = 1u, -- UV__IORING_FEAT_NODROP = 2u, -- UV__IORING_FEAT_RSRC_TAGS = 1024u, /* linux v5.13 */ --}; -- --enum { -- UV__IORING_OP_READV = 1, -- UV__IORING_OP_WRITEV = 2, -- UV__IORING_OP_FSYNC = 3, -- UV__IORING_OP_OPENAT = 18, -- UV__IORING_OP_CLOSE = 19, -- UV__IORING_OP_STATX = 21, -- UV__IORING_OP_EPOLL_CTL = 29, --}; -- --enum { -- UV__IORING_ENTER_GETEVENTS = 1u, -- UV__IORING_ENTER_SQ_WAKEUP = 2u, --}; -- --enum { -- UV__IORING_SQ_NEED_WAKEUP = 1u, -- UV__IORING_SQ_CQ_OVERFLOW = 2u, --}; -- --struct uv__io_cqring_offsets { -- uint32_t head; -- uint32_t tail; -- uint32_t ring_mask; -- uint32_t ring_entries; -- uint32_t overflow; -- uint32_t cqes; -- uint64_t reserved0; -- uint64_t reserved1; --}; -- --STATIC_ASSERT(40 == sizeof(struct uv__io_cqring_offsets)); -- --struct uv__io_sqring_offsets { -- uint32_t head; -- uint32_t tail; -- uint32_t ring_mask; -- uint32_t ring_entries; -- uint32_t flags; -- uint32_t dropped; -- uint32_t array; -- uint32_t reserved0; -- uint64_t reserved1; --}; -- --STATIC_ASSERT(40 == sizeof(struct uv__io_sqring_offsets)); -- --struct uv__io_uring_cqe { -- uint64_t user_data; -- int32_t res; -- uint32_t flags; --}; -- --STATIC_ASSERT(16 == sizeof(struct uv__io_uring_cqe)); -- --struct uv__io_uring_sqe { -- uint8_t opcode; -- uint8_t flags; -- uint16_t ioprio; -- int32_t fd; -- union { -- uint64_t off; -- uint64_t addr2; -- }; -- union { -- uint64_t addr; -- }; -- uint32_t len; -- union { -- uint32_t rw_flags; -- uint32_t fsync_flags; -- uint32_t open_flags; -- uint32_t statx_flags; -- }; -- uint64_t user_data; -- union { -- uint16_t buf_index; -- uint64_t pad[3]; -- }; --}; -- --STATIC_ASSERT(64 == sizeof(struct uv__io_uring_sqe)); --STATIC_ASSERT(0 == offsetof(struct uv__io_uring_sqe, opcode)); --STATIC_ASSERT(1 == offsetof(struct uv__io_uring_sqe, flags)); --STATIC_ASSERT(2 == offsetof(struct uv__io_uring_sqe, ioprio)); --STATIC_ASSERT(4 == offsetof(struct uv__io_uring_sqe, fd)); --STATIC_ASSERT(8 == offsetof(struct uv__io_uring_sqe, off)); --STATIC_ASSERT(16 == offsetof(struct uv__io_uring_sqe, addr)); --STATIC_ASSERT(24 == offsetof(struct uv__io_uring_sqe, len)); --STATIC_ASSERT(28 == offsetof(struct uv__io_uring_sqe, rw_flags)); --STATIC_ASSERT(32 == offsetof(struct uv__io_uring_sqe, user_data)); --STATIC_ASSERT(40 == offsetof(struct uv__io_uring_sqe, buf_index)); -- --struct uv__io_uring_params { -- uint32_t sq_entries; -- uint32_t cq_entries; -- uint32_t flags; -- uint32_t sq_thread_cpu; -- uint32_t sq_thread_idle; -- uint32_t features; -- uint32_t reserved[4]; -- struct uv__io_sqring_offsets sq_off; /* 40 bytes */ -- struct uv__io_cqring_offsets cq_off; /* 40 bytes */ --}; -- --STATIC_ASSERT(40 + 40 + 40 == sizeof(struct uv__io_uring_params)); --STATIC_ASSERT(40 == offsetof(struct uv__io_uring_params, sq_off)); --STATIC_ASSERT(80 == offsetof(struct uv__io_uring_params, cq_off)); -- --STATIC_ASSERT(EPOLL_CTL_ADD < 4); --STATIC_ASSERT(EPOLL_CTL_DEL < 4); --STATIC_ASSERT(EPOLL_CTL_MOD < 4); -- - struct watcher_list { - RB_ENTRY(watcher_list) entry; - QUEUE watchers; -@@ -276,17 +137,6 @@ static int compare_watchers(const struct watcher_list* a, - static void maybe_free_watcher_list(struct watcher_list* w, - uv_loop_t* loop); - --static void uv__epoll_ctl_flush(int epollfd, -- struct uv__iou* ctl, -- struct epoll_event (*events)[256]); -- --static void uv__epoll_ctl_prep(int epollfd, -- struct uv__iou* ctl, -- struct epoll_event (*events)[256], -- int op, -- int fd, -- struct epoll_event* e); -- - RB_GENERATE_STATIC(watcher_root, watcher_list, entry, compare_watchers) - - -@@ -356,187 +206,7 @@ ssize_t uv__getrandom(void* buf, size_t buflen, unsigned flags) { - } - - --int uv__io_uring_setup(int entries, struct uv__io_uring_params* params) { -- return syscall(__NR_io_uring_setup, entries, params); --} -- -- --int uv__io_uring_enter(int fd, -- unsigned to_submit, -- unsigned min_complete, -- unsigned flags) { -- /* io_uring_enter used to take a sigset_t but it's unused -- * in newer kernels unless IORING_ENTER_EXT_ARG is set, -- * in which case it takes a struct io_uring_getevents_arg. -- */ -- return syscall(__NR_io_uring_enter, -- fd, -- to_submit, -- min_complete, -- flags, -- NULL, -- 0L); --} -- -- --int uv__io_uring_register(int fd, unsigned opcode, void* arg, unsigned nargs) { -- return syscall(__NR_io_uring_register, fd, opcode, arg, nargs); --} -- -- --static int uv__use_io_uring(void) { -- /* Ternary: unknown=0, yes=1, no=-1 */ -- static _Atomic int use_io_uring; -- char* val; -- int use; -- -- use = atomic_load_explicit(&use_io_uring, memory_order_relaxed); -- -- if (use == 0) { -- val = getenv("UV_USE_IO_URING"); -- use = val == NULL || atoi(val) ? 1 : -1; -- atomic_store_explicit(&use_io_uring, use, memory_order_relaxed); -- } -- -- return use > 0; --} -- -- --static void uv__iou_init(int epollfd, -- struct uv__iou* iou, -- uint32_t entries, -- uint32_t flags) { -- struct uv__io_uring_params params; -- struct epoll_event e; -- size_t cqlen; -- size_t sqlen; -- size_t maxlen; -- size_t sqelen; -- uint32_t i; -- char* sq; -- char* sqe; -- int ringfd; -- -- sq = MAP_FAILED; -- sqe = MAP_FAILED; -- -- if (!uv__use_io_uring()) -- return; -- -- /* SQPOLL required CAP_SYS_NICE until linux v5.12 relaxed that requirement. -- * Mostly academic because we check for a v5.13 kernel afterwards anyway. -- */ -- memset(¶ms, 0, sizeof(params)); -- params.flags = flags; -- -- if (flags & UV__IORING_SETUP_SQPOLL) -- params.sq_thread_idle = 10; /* milliseconds */ -- -- /* Kernel returns a file descriptor with O_CLOEXEC flag set. */ -- ringfd = uv__io_uring_setup(entries, ¶ms); -- if (ringfd == -1) -- return; -- -- /* IORING_FEAT_RSRC_TAGS is used to detect linux v5.13 but what we're -- * actually detecting is whether IORING_OP_STATX works with SQPOLL. -- */ -- if (!(params.features & UV__IORING_FEAT_RSRC_TAGS)) -- goto fail; -- -- /* Implied by IORING_FEAT_RSRC_TAGS but checked explicitly anyway. */ -- if (!(params.features & UV__IORING_FEAT_SINGLE_MMAP)) -- goto fail; -- -- /* Implied by IORING_FEAT_RSRC_TAGS but checked explicitly anyway. */ -- if (!(params.features & UV__IORING_FEAT_NODROP)) -- goto fail; -- -- sqlen = params.sq_off.array + params.sq_entries * sizeof(uint32_t); -- cqlen = -- params.cq_off.cqes + params.cq_entries * sizeof(struct uv__io_uring_cqe); -- maxlen = sqlen < cqlen ? cqlen : sqlen; -- sqelen = params.sq_entries * sizeof(struct uv__io_uring_sqe); -- -- sq = mmap(0, -- maxlen, -- PROT_READ | PROT_WRITE, -- MAP_SHARED | MAP_POPULATE, -- ringfd, -- 0); /* IORING_OFF_SQ_RING */ -- -- sqe = mmap(0, -- sqelen, -- PROT_READ | PROT_WRITE, -- MAP_SHARED | MAP_POPULATE, -- ringfd, -- 0x10000000ull); /* IORING_OFF_SQES */ -- -- if (sq == MAP_FAILED || sqe == MAP_FAILED) -- goto fail; -- -- if (flags & UV__IORING_SETUP_SQPOLL) { -- /* Only interested in completion events. To get notified when -- * the kernel pulls items from the submission ring, add POLLOUT. -- */ -- memset(&e, 0, sizeof(e)); -- e.events = POLLIN; -- e.data.fd = ringfd; -- -- if (epoll_ctl(epollfd, EPOLL_CTL_ADD, ringfd, &e)) -- goto fail; -- } -- -- iou->sqhead = (uint32_t*) (sq + params.sq_off.head); -- iou->sqtail = (uint32_t*) (sq + params.sq_off.tail); -- iou->sqmask = *(uint32_t*) (sq + params.sq_off.ring_mask); -- iou->sqarray = (uint32_t*) (sq + params.sq_off.array); -- iou->sqflags = (uint32_t*) (sq + params.sq_off.flags); -- iou->cqhead = (uint32_t*) (sq + params.cq_off.head); -- iou->cqtail = (uint32_t*) (sq + params.cq_off.tail); -- iou->cqmask = *(uint32_t*) (sq + params.cq_off.ring_mask); -- iou->sq = sq; -- iou->cqe = sq + params.cq_off.cqes; -- iou->sqe = sqe; -- iou->sqlen = sqlen; -- iou->cqlen = cqlen; -- iou->maxlen = maxlen; -- iou->sqelen = sqelen; -- iou->ringfd = ringfd; -- iou->in_flight = 0; -- -- for (i = 0; i <= iou->sqmask; i++) -- iou->sqarray[i] = i; /* Slot -> sqe identity mapping. */ -- -- return; -- --fail: -- if (sq != MAP_FAILED) -- munmap(sq, maxlen); -- -- if (sqe != MAP_FAILED) -- munmap(sqe, sqelen); -- -- uv__close(ringfd); --} -- -- --static void uv__iou_delete(struct uv__iou* iou) { -- if (iou->ringfd != -1) { -- munmap(iou->sq, iou->maxlen); -- munmap(iou->sqe, iou->sqelen); -- uv__close(iou->ringfd); -- iou->ringfd = -1; -- } --} -- -- - int uv__platform_loop_init(uv_loop_t* loop) { -- uv__loop_internal_fields_t* lfields; -- -- lfields = uv__get_internal_fields(loop); -- lfields->ctl.ringfd = -1; -- lfields->iou.ringfd = -1; -- - loop->inotify_watchers = NULL; - loop->inotify_fd = -1; - loop->backend_fd = epoll_create1(O_CLOEXEC); -@@ -544,9 +214,6 @@ int uv__platform_loop_init(uv_loop_t* loop) { - if (loop->backend_fd == -1) - return UV__ERR(errno); - -- uv__iou_init(loop->backend_fd, &lfields->iou, 64, UV__IORING_SETUP_SQPOLL); -- uv__iou_init(loop->backend_fd, &lfields->ctl, 256, 0); -- - return 0; - } - -@@ -559,8 +226,6 @@ int uv__io_fork(uv_loop_t* loop) { - - uv__close(loop->backend_fd); - loop->backend_fd = -1; -- -- /* TODO(bnoordhuis) Loses items from the submission and completion rings. */ - uv__platform_loop_delete(loop); - - err = uv__platform_loop_init(loop); -@@ -572,62 +237,42 @@ int uv__io_fork(uv_loop_t* loop) { - - - void uv__platform_loop_delete(uv_loop_t* loop) { -- uv__loop_internal_fields_t* lfields; -- -- lfields = uv__get_internal_fields(loop); -- uv__iou_delete(&lfields->ctl); -- uv__iou_delete(&lfields->iou); -- -- if (loop->inotify_fd != -1) { -- uv__io_stop(loop, &loop->inotify_read_watcher, POLLIN); -- uv__close(loop->inotify_fd); -- loop->inotify_fd = -1; -- } -+ if (loop->inotify_fd == -1) return; -+ uv__io_stop(loop, &loop->inotify_read_watcher, POLLIN); -+ uv__close(loop->inotify_fd); -+ loop->inotify_fd = -1; - } - - --struct uv__invalidate { -- struct epoll_event (*prep)[256]; -- struct epoll_event* events; -- int nfds; --}; -- -- - void uv__platform_invalidate_fd(uv_loop_t* loop, int fd) { -- uv__loop_internal_fields_t* lfields; -- struct uv__invalidate* inv; -+ struct epoll_event* events; - struct epoll_event dummy; -- int i; -+ uintptr_t i; -+ uintptr_t nfds; - -- lfields = uv__get_internal_fields(loop); -- inv = lfields->inv; -+ assert(loop->watchers != NULL); -+ assert(fd >= 0); - -- /* Invalidate events with same file descriptor */ -- if (inv != NULL) -- for (i = 0; i < inv->nfds; i++) -- if (inv->events[i].data.fd == fd) -- inv->events[i].data.fd = -1; -+ events = (struct epoll_event*) loop->watchers[loop->nwatchers]; -+ nfds = (uintptr_t) loop->watchers[loop->nwatchers + 1]; -+ if (events != NULL) -+ /* Invalidate events with same file descriptor */ -+ for (i = 0; i < nfds; i++) -+ if (events[i].data.fd == fd) -+ events[i].data.fd = -1; - - /* Remove the file descriptor from the epoll. - * This avoids a problem where the same file description remains open - * in another process, causing repeated junk epoll events. - * - * We pass in a dummy epoll_event, to work around a bug in old kernels. -- * -- * Work around a bug in kernels 3.10 to 3.19 where passing a struct that -- * has the EPOLLWAKEUP flag set generates spurious audit syslog warnings. - */ -- memset(&dummy, 0, sizeof(dummy)); -- -- if (inv == NULL) { -+ if (loop->backend_fd >= 0) { -+ /* Work around a bug in kernels 3.10 to 3.19 where passing a struct that -+ * has the EPOLLWAKEUP flag set generates spurious audit syslog warnings. -+ */ -+ memset(&dummy, 0, sizeof(dummy)); - epoll_ctl(loop->backend_fd, EPOLL_CTL_DEL, fd, &dummy); -- } else { -- uv__epoll_ctl_prep(loop->backend_fd, -- &lfields->ctl, -- inv->prep, -- EPOLL_CTL_DEL, -- fd, -- &dummy); - } - } - -@@ -653,454 +298,27 @@ int uv__io_check_fd(uv_loop_t* loop, int fd) { - } - - --/* Caller must initialize SQE and call uv__iou_submit(). */ --static struct uv__io_uring_sqe* uv__iou_get_sqe(struct uv__iou* iou, -- uv_loop_t* loop, -- uv_fs_t* req) { -- struct uv__io_uring_sqe* sqe; -- uint32_t head; -- uint32_t tail; -- uint32_t mask; -- uint32_t slot; -- -- if (iou->ringfd == -1) -- return NULL; -- -- head = atomic_load_explicit((_Atomic uint32_t*) iou->sqhead, -- memory_order_acquire); -- tail = *iou->sqtail; -- mask = iou->sqmask; -- -- if ((head & mask) == ((tail + 1) & mask)) -- return NULL; /* No room in ring buffer. TODO(bnoordhuis) maybe flush it? */ -- -- slot = tail & mask; -- sqe = iou->sqe; -- sqe = &sqe[slot]; -- memset(sqe, 0, sizeof(*sqe)); -- sqe->user_data = (uintptr_t) req; -- -- /* Pacify uv_cancel(). */ -- req->work_req.loop = loop; -- req->work_req.work = NULL; -- req->work_req.done = NULL; -- QUEUE_INIT(&req->work_req.wq); -- -- uv__req_register(loop, req); -- iou->in_flight++; -- -- return sqe; --} -- -- --static void uv__iou_submit(struct uv__iou* iou) { -- uint32_t flags; -- -- atomic_store_explicit((_Atomic uint32_t*) iou->sqtail, -- *iou->sqtail + 1, -- memory_order_release); -- -- flags = atomic_load_explicit((_Atomic uint32_t*) iou->sqflags, -- memory_order_acquire); -- -- if (flags & UV__IORING_SQ_NEED_WAKEUP) -- if (uv__io_uring_enter(iou->ringfd, 0, 0, UV__IORING_ENTER_SQ_WAKEUP)) -- if (errno != EOWNERDEAD) /* Kernel bug. Harmless, ignore. */ -- perror("libuv: io_uring_enter(wakeup)"); /* Can't happen. */ --} -- -- --int uv__iou_fs_close(uv_loop_t* loop, uv_fs_t* req) { -- struct uv__io_uring_sqe* sqe; -- struct uv__iou* iou; -- -- iou = &uv__get_internal_fields(loop)->iou; -- -- sqe = uv__iou_get_sqe(iou, loop, req); -- if (sqe == NULL) -- return 0; -- -- sqe->fd = req->file; -- sqe->opcode = UV__IORING_OP_CLOSE; -- -- uv__iou_submit(iou); -- -- return 1; --} -- -- --int uv__iou_fs_fsync_or_fdatasync(uv_loop_t* loop, -- uv_fs_t* req, -- uint32_t fsync_flags) { -- struct uv__io_uring_sqe* sqe; -- struct uv__iou* iou; -- -- iou = &uv__get_internal_fields(loop)->iou; -- -- sqe = uv__iou_get_sqe(iou, loop, req); -- if (sqe == NULL) -- return 0; -- -- /* Little known fact: setting seq->off and seq->len turns -- * it into an asynchronous sync_file_range() operation. -- */ -- sqe->fd = req->file; -- sqe->fsync_flags = fsync_flags; -- sqe->opcode = UV__IORING_OP_FSYNC; -- -- uv__iou_submit(iou); -- -- return 1; --} -- -- --int uv__iou_fs_open(uv_loop_t* loop, uv_fs_t* req) { -- struct uv__io_uring_sqe* sqe; -- struct uv__iou* iou; -- -- iou = &uv__get_internal_fields(loop)->iou; -- -- sqe = uv__iou_get_sqe(iou, loop, req); -- if (sqe == NULL) -- return 0; -- -- sqe->addr = (uintptr_t) req->path; -- sqe->fd = AT_FDCWD; -- sqe->len = req->mode; -- sqe->opcode = UV__IORING_OP_OPENAT; -- sqe->open_flags = req->flags | O_CLOEXEC; -- -- uv__iou_submit(iou); -- -- return 1; --} -- -- --int uv__iou_fs_read_or_write(uv_loop_t* loop, -- uv_fs_t* req, -- int is_read) { -- struct uv__io_uring_sqe* sqe; -- struct uv__iou* iou; -- -- /* For the moment, if iovcnt is greater than IOV_MAX, fallback to the -- * threadpool. In the future we might take advantage of IOSQE_IO_LINK. */ -- if (req->nbufs > IOV_MAX) -- return 0; -- -- iou = &uv__get_internal_fields(loop)->iou; -- -- sqe = uv__iou_get_sqe(iou, loop, req); -- if (sqe == NULL) -- return 0; -- -- sqe->addr = (uintptr_t) req->bufs; -- sqe->fd = req->file; -- sqe->len = req->nbufs; -- sqe->off = req->off < 0 ? -1 : req->off; -- sqe->opcode = is_read ? UV__IORING_OP_READV : UV__IORING_OP_WRITEV; -- -- uv__iou_submit(iou); -- -- return 1; --} -- -- --int uv__iou_fs_statx(uv_loop_t* loop, -- uv_fs_t* req, -- int is_fstat, -- int is_lstat) { -- struct uv__io_uring_sqe* sqe; -- struct uv__statx* statxbuf; -- struct uv__iou* iou; -- -- statxbuf = uv__malloc(sizeof(*statxbuf)); -- if (statxbuf == NULL) -- return 0; -- -- iou = &uv__get_internal_fields(loop)->iou; -- -- sqe = uv__iou_get_sqe(iou, loop, req); -- if (sqe == NULL) { -- uv__free(statxbuf); -- return 0; -- } -- -- req->ptr = statxbuf; -- -- sqe->addr = (uintptr_t) req->path; -- sqe->addr2 = (uintptr_t) statxbuf; -- sqe->fd = AT_FDCWD; -- sqe->len = 0xFFF; /* STATX_BASIC_STATS + STATX_BTIME */ -- sqe->opcode = UV__IORING_OP_STATX; -- -- if (is_fstat) { -- sqe->addr = (uintptr_t) ""; -- sqe->fd = req->file; -- sqe->statx_flags |= 0x1000; /* AT_EMPTY_PATH */ -- } -- -- if (is_lstat) -- sqe->statx_flags |= AT_SYMLINK_NOFOLLOW; -- -- uv__iou_submit(iou); -- -- return 1; --} -- -- --void uv__statx_to_stat(const struct uv__statx* statxbuf, uv_stat_t* buf) { -- buf->st_dev = makedev(statxbuf->stx_dev_major, statxbuf->stx_dev_minor); -- buf->st_mode = statxbuf->stx_mode; -- buf->st_nlink = statxbuf->stx_nlink; -- buf->st_uid = statxbuf->stx_uid; -- buf->st_gid = statxbuf->stx_gid; -- buf->st_rdev = makedev(statxbuf->stx_rdev_major, statxbuf->stx_rdev_minor); -- buf->st_ino = statxbuf->stx_ino; -- buf->st_size = statxbuf->stx_size; -- buf->st_blksize = statxbuf->stx_blksize; -- buf->st_blocks = statxbuf->stx_blocks; -- buf->st_atim.tv_sec = statxbuf->stx_atime.tv_sec; -- buf->st_atim.tv_nsec = statxbuf->stx_atime.tv_nsec; -- buf->st_mtim.tv_sec = statxbuf->stx_mtime.tv_sec; -- buf->st_mtim.tv_nsec = statxbuf->stx_mtime.tv_nsec; -- buf->st_ctim.tv_sec = statxbuf->stx_ctime.tv_sec; -- buf->st_ctim.tv_nsec = statxbuf->stx_ctime.tv_nsec; -- buf->st_birthtim.tv_sec = statxbuf->stx_btime.tv_sec; -- buf->st_birthtim.tv_nsec = statxbuf->stx_btime.tv_nsec; -- buf->st_flags = 0; -- buf->st_gen = 0; --} -- -- --static void uv__iou_fs_statx_post(uv_fs_t* req) { -- struct uv__statx* statxbuf; -- uv_stat_t* buf; -- -- buf = &req->statbuf; -- statxbuf = req->ptr; -- req->ptr = NULL; -- -- if (req->result == 0) { -- uv__msan_unpoison(statxbuf, sizeof(*statxbuf)); -- uv__statx_to_stat(statxbuf, buf); -- req->ptr = buf; -- } -- -- uv__free(statxbuf); --} -- -- --static void uv__poll_io_uring(uv_loop_t* loop, struct uv__iou* iou) { -- struct uv__io_uring_cqe* cqe; -- struct uv__io_uring_cqe* e; -- uv_fs_t* req; -- uint32_t head; -- uint32_t tail; -- uint32_t mask; -- uint32_t i; -- uint32_t flags; -- int nevents; -- int rc; -- -- head = *iou->cqhead; -- tail = atomic_load_explicit((_Atomic uint32_t*) iou->cqtail, -- memory_order_acquire); -- mask = iou->cqmask; -- cqe = iou->cqe; -- nevents = 0; -- -- for (i = head; i != tail; i++) { -- e = &cqe[i & mask]; -- -- req = (uv_fs_t*) (uintptr_t) e->user_data; -- assert(req->type == UV_FS); -- -- uv__req_unregister(loop, req); -- iou->in_flight--; -- -- /* io_uring stores error codes as negative numbers, same as libuv. */ -- req->result = e->res; -- -- switch (req->fs_type) { -- case UV_FS_FSTAT: -- case UV_FS_LSTAT: -- case UV_FS_STAT: -- uv__iou_fs_statx_post(req); -- break; -- default: /* Squelch -Wswitch warnings. */ -- break; -- } -- -- uv__metrics_update_idle_time(loop); -- req->cb(req); -- nevents++; -- } -- -- atomic_store_explicit((_Atomic uint32_t*) iou->cqhead, -- tail, -- memory_order_release); -- -- /* Check whether CQE's overflowed, if so enter the kernel to make them -- * available. Don't grab them immediately but in the next loop iteration to -- * avoid loop starvation. */ -- flags = atomic_load_explicit((_Atomic uint32_t*) iou->sqflags, -- memory_order_acquire); -- -- if (flags & UV__IORING_SQ_CQ_OVERFLOW) { -- do -- rc = uv__io_uring_enter(iou->ringfd, 0, 0, UV__IORING_ENTER_GETEVENTS); -- while (rc == -1 && errno == EINTR); -- -- if (rc < 0) -- perror("libuv: io_uring_enter(getevents)"); /* Can't happen. */ -- } -- -- uv__metrics_inc_events(loop, nevents); -- if (uv__get_internal_fields(loop)->current_timeout == 0) -- uv__metrics_inc_events_waiting(loop, nevents); --} -- -- --static void uv__epoll_ctl_prep(int epollfd, -- struct uv__iou* ctl, -- struct epoll_event (*events)[256], -- int op, -- int fd, -- struct epoll_event* e) { -- struct uv__io_uring_sqe* sqe; -- struct epoll_event* pe; -- uint32_t mask; -- uint32_t slot; -- -- if (ctl->ringfd == -1) { -- if (!epoll_ctl(epollfd, op, fd, e)) -- return; -- -- if (op == EPOLL_CTL_DEL) -- return; /* Ignore errors, may be racing with another thread. */ -- -- if (op != EPOLL_CTL_ADD) -- abort(); -- -- if (errno != EEXIST) -- abort(); -- -- /* File descriptor that's been watched before, update event mask. */ -- if (!epoll_ctl(epollfd, EPOLL_CTL_MOD, fd, e)) -- return; -- -- abort(); -- } else { -- mask = ctl->sqmask; -- slot = (*ctl->sqtail)++ & mask; -- -- pe = &(*events)[slot]; -- *pe = *e; -- -- sqe = ctl->sqe; -- sqe = &sqe[slot]; -- -- memset(sqe, 0, sizeof(*sqe)); -- sqe->addr = (uintptr_t) pe; -- sqe->fd = epollfd; -- sqe->len = op; -- sqe->off = fd; -- sqe->opcode = UV__IORING_OP_EPOLL_CTL; -- sqe->user_data = op | slot << 2 | (int64_t) fd << 32; -- -- if ((*ctl->sqhead & mask) == (*ctl->sqtail & mask)) -- uv__epoll_ctl_flush(epollfd, ctl, events); -- } --} -- -- --static void uv__epoll_ctl_flush(int epollfd, -- struct uv__iou* ctl, -- struct epoll_event (*events)[256]) { -- struct epoll_event oldevents[256]; -- struct uv__io_uring_cqe* cqe; -- uint32_t oldslot; -- uint32_t slot; -- uint32_t n; -- int fd; -- int op; -- int rc; -- -- STATIC_ASSERT(sizeof(oldevents) == sizeof(*events)); -- assert(ctl->ringfd != -1); -- assert(*ctl->sqhead != *ctl->sqtail); -- -- n = *ctl->sqtail - *ctl->sqhead; -- do -- rc = uv__io_uring_enter(ctl->ringfd, n, n, UV__IORING_ENTER_GETEVENTS); -- while (rc == -1 && errno == EINTR); -- -- if (rc < 0) -- perror("libuv: io_uring_enter(getevents)"); /* Can't happen. */ -- -- if (rc != (int) n) -- abort(); -- -- assert(*ctl->sqhead == *ctl->sqtail); -- -- memcpy(oldevents, *events, sizeof(*events)); -- -- /* Failed submissions are either EPOLL_CTL_DEL commands for file descriptors -- * that have been closed, or EPOLL_CTL_ADD commands for file descriptors -- * that we are already watching. Ignore the former and retry the latter -- * with EPOLL_CTL_MOD. -- */ -- while (*ctl->cqhead != *ctl->cqtail) { -- slot = (*ctl->cqhead)++ & ctl->cqmask; -- -- cqe = ctl->cqe; -- cqe = &cqe[slot]; -- -- if (cqe->res == 0) -- continue; -- -- fd = cqe->user_data >> 32; -- op = 3 & cqe->user_data; -- oldslot = 255 & (cqe->user_data >> 2); -- -- if (op == EPOLL_CTL_DEL) -- continue; -- -- if (op != EPOLL_CTL_ADD) -- abort(); -- -- if (cqe->res != -EEXIST) -- abort(); -- -- uv__epoll_ctl_prep(epollfd, -- ctl, -- events, -- EPOLL_CTL_MOD, -- fd, -- &oldevents[oldslot]); -- } --} -- -- - void uv__io_poll(uv_loop_t* loop, int timeout) { -- uv__loop_internal_fields_t* lfields; -+ /* A bug in kernels < 2.6.37 makes timeouts larger than ~30 minutes -+ * effectively infinite on 32 bits architectures. To avoid blocking -+ * indefinitely, we cap the timeout and poll again if necessary. -+ * -+ * Note that "30 minutes" is a simplification because it depends on -+ * the value of CONFIG_HZ. The magic constant assumes CONFIG_HZ=1200, -+ * that being the largest value I have seen in the wild (and only once.) -+ */ -+ static const int max_safe_timeout = 1789569; - struct epoll_event events[1024]; -- struct epoll_event prep[256]; -- struct uv__invalidate inv; - struct epoll_event* pe; - struct epoll_event e; -- struct uv__iou* ctl; -- struct uv__iou* iou; - int real_timeout; - QUEUE* q; - uv__io_t* w; - sigset_t* sigmask; - sigset_t sigset; - uint64_t base; -- int have_iou_events; - int have_signals; - int nevents; -- int epollfd; - int count; - int nfds; - int fd; -@@ -1109,9 +327,47 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - int user_timeout; - int reset_timeout; - -- lfields = uv__get_internal_fields(loop); -- ctl = &lfields->ctl; -- iou = &lfields->iou; -+ if (loop->nfds == 0) { -+ assert(QUEUE_EMPTY(&loop->watcher_queue)); -+ return; -+ } -+ -+ memset(&e, 0, sizeof(e)); -+ -+ while (!QUEUE_EMPTY(&loop->watcher_queue)) { -+ q = QUEUE_HEAD(&loop->watcher_queue); -+ QUEUE_REMOVE(q); -+ QUEUE_INIT(q); -+ -+ w = QUEUE_DATA(q, uv__io_t, watcher_queue); -+ assert(w->pevents != 0); -+ assert(w->fd >= 0); -+ assert(w->fd < (int) loop->nwatchers); -+ -+ e.events = w->pevents; -+ e.data.fd = w->fd; -+ -+ if (w->events == 0) -+ op = EPOLL_CTL_ADD; -+ else -+ op = EPOLL_CTL_MOD; -+ -+ /* XXX Future optimization: do EPOLL_CTL_MOD lazily if we stop watching -+ * events, skip the syscall and squelch the events after epoll_wait(). -+ */ -+ if (epoll_ctl(loop->backend_fd, op, w->fd, &e)) { -+ if (errno != EEXIST) -+ abort(); -+ -+ assert(op == EPOLL_CTL_ADD); -+ -+ /* We've reactivated a file descriptor that's been watched before. */ -+ if (epoll_ctl(loop->backend_fd, EPOLL_CTL_MOD, w->fd, &e)) -+ abort(); -+ } -+ -+ w->events = w->pevents; -+ } - - sigmask = NULL; - if (loop->flags & UV_LOOP_BLOCK_SIGPROF) { -@@ -1125,7 +381,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - count = 48; /* Benchmarks suggest this gives the best throughput. */ - real_timeout = timeout; - -- if (lfields->flags & UV_METRICS_IDLE_TIME) { -+ if (uv__get_internal_fields(loop)->flags & UV_METRICS_IDLE_TIME) { - reset_timeout = 1; - user_timeout = timeout; - timeout = 0; -@@ -1134,56 +390,24 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - user_timeout = 0; - } - -- epollfd = loop->backend_fd; -- -- memset(&e, 0, sizeof(e)); -- -- while (!QUEUE_EMPTY(&loop->watcher_queue)) { -- q = QUEUE_HEAD(&loop->watcher_queue); -- w = QUEUE_DATA(q, uv__io_t, watcher_queue); -- QUEUE_REMOVE(q); -- QUEUE_INIT(q); -- -- op = EPOLL_CTL_MOD; -- if (w->events == 0) -- op = EPOLL_CTL_ADD; -- -- w->events = w->pevents; -- e.events = w->pevents; -- e.data.fd = w->fd; -- -- uv__epoll_ctl_prep(epollfd, ctl, &prep, op, w->fd, &e); -- } -- -- inv.events = events; -- inv.prep = &prep; -- inv.nfds = -1; -- - for (;;) { -- if (loop->nfds == 0) -- if (iou->in_flight == 0) -- break; -- -- /* All event mask mutations should be visible to the kernel before -- * we enter epoll_pwait(). -- */ -- if (ctl->ringfd != -1) -- while (*ctl->sqhead != *ctl->sqtail) -- uv__epoll_ctl_flush(epollfd, ctl, &prep); -- - /* Only need to set the provider_entry_time if timeout != 0. The function - * will return early if the loop isn't configured with UV_METRICS_IDLE_TIME. - */ - if (timeout != 0) - uv__metrics_set_provider_entry_time(loop); - -- /* Store the current timeout in a location that's globally accessible so -- * other locations like uv__work_done() can determine whether the queue -- * of events in the callback were waiting when poll was called. -+ /* See the comment for max_safe_timeout for an explanation of why -+ * this is necessary. Executive summary: kernel bug workaround. - */ -- lfields->current_timeout = timeout; -+ if (sizeof(int32_t) == sizeof(long) && timeout >= max_safe_timeout) -+ timeout = max_safe_timeout; - -- nfds = epoll_pwait(epollfd, events, ARRAY_SIZE(events), timeout, sigmask); -+ nfds = epoll_pwait(loop->backend_fd, -+ events, -+ ARRAY_SIZE(events), -+ timeout, -+ sigmask); - - /* Update loop->time unconditionally. It's tempting to skip the update when - * timeout == 0 (i.e. non-blocking poll) but there is no guarantee that the -@@ -1203,7 +427,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - continue; - - if (timeout == 0) -- break; -+ return; - - /* We may have been inside the system call for longer than |timeout| - * milliseconds so we need to update the timestamp to avoid drift. -@@ -1224,18 +448,27 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - continue; - - if (timeout == 0) -- break; -+ return; - - /* Interrupted by a signal. Update timeout and poll again. */ - goto update_timeout; - } - -- have_iou_events = 0; - have_signals = 0; - nevents = 0; - -- inv.nfds = nfds; -- lfields->inv = &inv; -+ { -+ /* Squelch a -Waddress-of-packed-member warning with gcc >= 9. */ -+ union { -+ struct epoll_event* events; -+ uv__io_t* watchers; -+ } x; -+ -+ x.events = events; -+ assert(loop->watchers != NULL); -+ loop->watchers[loop->nwatchers] = x.watchers; -+ loop->watchers[loop->nwatchers + 1] = (void*) (uintptr_t) nfds; -+ } - - for (i = 0; i < nfds; i++) { - pe = events + i; -@@ -1245,12 +478,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - if (fd == -1) - continue; - -- if (fd == iou->ringfd) { -- uv__poll_io_uring(loop, iou); -- have_iou_events = 1; -- continue; -- } -- - assert(fd >= 0); - assert((unsigned) fd < loop->nwatchers); - -@@ -1262,7 +489,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - * Ignore all errors because we may be racing with another thread - * when the file descriptor is closed. - */ -- uv__epoll_ctl_prep(epollfd, ctl, &prep, EPOLL_CTL_DEL, fd, pe); -+ epoll_ctl(loop->backend_fd, EPOLL_CTL_DEL, fd, pe); - continue; - } - -@@ -1319,13 +546,11 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - loop->signal_io_watcher.cb(loop, &loop->signal_io_watcher, POLLIN); - } - -- lfields->inv = NULL; -- -- if (have_iou_events != 0) -- break; /* Event loop should cycle now so don't poll again. */ -+ loop->watchers[loop->nwatchers] = NULL; -+ loop->watchers[loop->nwatchers + 1] = NULL; - - if (have_signals != 0) -- break; /* Event loop should cycle now so don't poll again. */ -+ return; /* Event loop should cycle now so don't poll again. */ - - if (nevents != 0) { - if (nfds == ARRAY_SIZE(events) && --count != 0) { -@@ -1333,11 +558,11 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - timeout = 0; - continue; - } -- break; -+ return; - } - - if (timeout == 0) -- break; -+ return; - - if (timeout == -1) - continue; -@@ -1347,14 +572,10 @@ update_timeout: - - real_timeout -= (loop->time - base); - if (real_timeout <= 0) -- break; -+ return; - - timeout = real_timeout; - } -- -- if (ctl->ringfd != -1) -- while (*ctl->sqhead != *ctl->sqtail) -- uv__epoll_ctl_flush(epollfd, ctl, &prep); - } - - uint64_t uv__hrtime(uv_clocktype_t type) { -@@ -1867,7 +1088,7 @@ static uint64_t uv__read_uint64(const char* filename) { - if (0 == uv__slurp(filename, buf, sizeof(buf))) - if (1 != sscanf(buf, "%" PRIu64, &rc)) - if (0 == strcmp(buf, "max\n")) -- rc = UINT64_MAX; -+ rc = ~0ull; - - return rc; - } -@@ -1903,7 +1124,6 @@ static void uv__get_cgroup1_memory_limits(char buf[static 1024], uint64_t* high, - char filename[4097]; - char* p; - int n; -- uint64_t cgroup1_max; - - /* Find out where the controller is mounted. */ - p = uv__cgroup1_find_memory_controller(buf, &n); -@@ -1920,22 +1140,12 @@ static void uv__get_cgroup1_memory_limits(char buf[static 1024], uint64_t* high, - * as indicated by uv__read_uint64 returning 0. - */ - if (*high != 0 && *max != 0) -- goto update_limits; -+ return; - } - - /* Fall back to the limits of the global memory controller. */ - *high = uv__read_uint64("/sys/fs/cgroup/memory/memory.soft_limit_in_bytes"); - *max = uv__read_uint64("/sys/fs/cgroup/memory/memory.limit_in_bytes"); -- -- /* uv__read_uint64 detects cgroup2's "max", so we need to separately detect -- * cgroup1's maximum value (which is derived from LONG_MAX and PAGE_SIZE). -- */ --update_limits: -- cgroup1_max = LONG_MAX & ~(sysconf(_SC_PAGESIZE) - 1); -- if (*high == cgroup1_max) -- *high = UINT64_MAX; -- if (*max == cgroup1_max) -- *max = UINT64_MAX; - } - - static void uv__get_cgroup2_memory_limits(char buf[static 1024], uint64_t* high, -diff --git a/deps/uv/src/unix/os390.c b/deps/uv/src/unix/os390.c -index a87c2d77faf..3954b2c2753 100644 ---- a/deps/uv/src/unix/os390.c -+++ b/deps/uv/src/unix/os390.c -@@ -808,7 +808,6 @@ static int os390_message_queue_handler(uv__os390_epoll* ep) { - - void uv__io_poll(uv_loop_t* loop, int timeout) { - static const int max_safe_timeout = 1789569; -- uv__loop_internal_fields_t* lfields; - struct epoll_event events[1024]; - struct epoll_event* pe; - struct epoll_event e; -@@ -831,8 +830,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - return; - } - -- lfields = uv__get_internal_fields(loop); -- - while (!QUEUE_EMPTY(&loop->watcher_queue)) { - uv_stream_t* stream; - -@@ -880,7 +877,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - int nevents = 0; - have_signals = 0; - -- if (lfields->flags & UV_METRICS_IDLE_TIME) { -+ if (uv__get_internal_fields(loop)->flags & UV_METRICS_IDLE_TIME) { - reset_timeout = 1; - user_timeout = timeout; - timeout = 0; -@@ -899,12 +896,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - if (sizeof(int32_t) == sizeof(long) && timeout >= max_safe_timeout) - timeout = max_safe_timeout; - -- /* Store the current timeout in a location that's globally accessible so -- * other locations like uv__work_done() can determine whether the queue -- * of events in the callback were waiting when poll was called. -- */ -- lfields->current_timeout = timeout; -- - nfds = epoll_wait(loop->ep, events, - ARRAY_SIZE(events), timeout); - -diff --git a/deps/uv/src/unix/posix-poll.c b/deps/uv/src/unix/posix-poll.c -index 7e7de86845d..711780ece8d 100644 ---- a/deps/uv/src/unix/posix-poll.c -+++ b/deps/uv/src/unix/posix-poll.c -@@ -132,7 +132,6 @@ static void uv__pollfds_del(uv_loop_t* loop, int fd) { - - - void uv__io_poll(uv_loop_t* loop, int timeout) { -- uv__loop_internal_fields_t* lfields; - sigset_t* pset; - sigset_t set; - uint64_t time_base; -@@ -153,8 +152,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - return; - } - -- lfields = uv__get_internal_fields(loop); -- - /* Take queued watchers and add their fds to our poll fds array. */ - while (!QUEUE_EMPTY(&loop->watcher_queue)) { - q = QUEUE_HEAD(&loop->watcher_queue); -@@ -182,7 +179,7 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - assert(timeout >= -1); - time_base = loop->time; - -- if (lfields->flags & UV_METRICS_IDLE_TIME) { -+ if (uv__get_internal_fields(loop)->flags & UV_METRICS_IDLE_TIME) { - reset_timeout = 1; - user_timeout = timeout; - timeout = 0; -@@ -201,12 +198,6 @@ void uv__io_poll(uv_loop_t* loop, int timeout) { - if (timeout != 0) - uv__metrics_set_provider_entry_time(loop); - -- /* Store the current timeout in a location that's globally accessible so -- * other locations like uv__work_done() can determine whether the queue -- * of events in the callback were waiting when poll was called. -- */ -- lfields->current_timeout = timeout; -- - if (pset != NULL) - if (pthread_sigmask(SIG_BLOCK, pset, NULL)) - abort(); -diff --git a/deps/uv/src/uv-common.h b/deps/uv/src/uv-common.h -index decde5362c8..2720121addc 100644 ---- a/deps/uv/src/uv-common.h -+++ b/deps/uv/src/uv-common.h -@@ -396,37 +396,9 @@ struct uv__loop_metrics_s { - void uv__metrics_update_idle_time(uv_loop_t* loop); - void uv__metrics_set_provider_entry_time(uv_loop_t* loop); - --#ifdef __linux__ --struct uv__iou { -- uint32_t* sqhead; -- uint32_t* sqtail; -- uint32_t* sqarray; -- uint32_t sqmask; -- uint32_t* sqflags; -- uint32_t* cqhead; -- uint32_t* cqtail; -- uint32_t cqmask; -- void* sq; /* pointer to munmap() on event loop teardown */ -- void* cqe; /* pointer to array of struct uv__io_uring_cqe */ -- void* sqe; /* pointer to array of struct uv__io_uring_sqe */ -- size_t sqlen; -- size_t cqlen; -- size_t maxlen; -- size_t sqelen; -- int ringfd; -- uint32_t in_flight; --}; --#endif /* __linux__ */ -- - struct uv__loop_internal_fields_s { - unsigned int flags; - uv__loop_metrics_t loop_metrics; -- int current_timeout; --#ifdef __linux__ -- struct uv__iou ctl; -- struct uv__iou iou; -- void* inv; /* used by uv__platform_invalidate_fd() */ --#endif /* __linux__ */ - }; - - #endif /* UV_COMMON_H_ */ -diff --git a/deps/uv/src/win/core.c b/deps/uv/src/win/core.c -index 9a3be58849a..e4041ec86a6 100644 ---- a/deps/uv/src/win/core.c -+++ b/deps/uv/src/win/core.c -@@ -424,7 +424,6 @@ int uv_backend_timeout(const uv_loop_t* loop) { - - - static void uv__poll_wine(uv_loop_t* loop, DWORD timeout) { -- uv__loop_internal_fields_t* lfields; - DWORD bytes; - ULONG_PTR key; - OVERLAPPED* overlapped; -@@ -434,10 +433,9 @@ static void uv__poll_wine(uv_loop_t* loop, DWORD timeout) { - uint64_t user_timeout; - int reset_timeout; - -- lfields = uv__get_internal_fields(loop); - timeout_time = loop->time + timeout; - -- if (lfields->flags & UV_METRICS_IDLE_TIME) { -+ if (uv__get_internal_fields(loop)->flags & UV_METRICS_IDLE_TIME) { - reset_timeout = 1; - user_timeout = timeout; - timeout = 0; -@@ -452,12 +450,6 @@ static void uv__poll_wine(uv_loop_t* loop, DWORD timeout) { - if (timeout != 0) - uv__metrics_set_provider_entry_time(loop); - -- /* Store the current timeout in a location that's globally accessible so -- * other locations like uv__work_done() can determine whether the queue -- * of events in the callback were waiting when poll was called. -- */ -- lfields->current_timeout = timeout; -- - GetQueuedCompletionStatus(loop->iocp, - &bytes, - &key, -@@ -515,7 +507,6 @@ static void uv__poll_wine(uv_loop_t* loop, DWORD timeout) { - - - static void uv__poll(uv_loop_t* loop, DWORD timeout) { -- uv__loop_internal_fields_t* lfields; - BOOL success; - uv_req_t* req; - OVERLAPPED_ENTRY overlappeds[128]; -@@ -527,10 +518,9 @@ static void uv__poll(uv_loop_t* loop, DWORD timeout) { - uint64_t actual_timeout; - int reset_timeout; - -- lfields = uv__get_internal_fields(loop); - timeout_time = loop->time + timeout; - -- if (lfields->flags & UV_METRICS_IDLE_TIME) { -+ if (uv__get_internal_fields(loop)->flags & UV_METRICS_IDLE_TIME) { - reset_timeout = 1; - user_timeout = timeout; - timeout = 0; -@@ -547,12 +537,6 @@ static void uv__poll(uv_loop_t* loop, DWORD timeout) { - if (timeout != 0) - uv__metrics_set_provider_entry_time(loop); - -- /* Store the current timeout in a location that's globally accessible so -- * other locations like uv__work_done() can determine whether the queue -- * of events in the callback were waiting when poll was called. -- */ -- lfields->current_timeout = timeout; -- - success = pGetQueuedCompletionStatusEx(loop->iocp, - overlappeds, - ARRAY_SIZE(overlappeds), diff --git a/meta-oe/recipes-devtools/nodejs/nodejs_20.11.0.bb b/meta-oe/recipes-devtools/nodejs/nodejs_20.11.1.bb similarity index 96% rename from meta-oe/recipes-devtools/nodejs/nodejs_20.11.0.bb rename to meta-oe/recipes-devtools/nodejs/nodejs_20.11.1.bb index cc5135478..e2e2f5ae8 100644 --- a/meta-oe/recipes-devtools/nodejs/nodejs_20.11.0.bb +++ b/meta-oe/recipes-devtools/nodejs/nodejs_20.11.1.bb @@ -24,8 +24,7 @@ SRC_URI = "http://nodejs.org/dist/v${PV}/node-v${PV}.tar.xz \ file://0004-v8-don-t-override-ARM-CFLAGS.patch \ file://system-c-ares.patch \ file://0001-liftoff-Correct-function-signatures.patch \ - file://0001-Revert-io_uring-changes-from-libuv-1.46.0.patch \ - file://0002-Revert-io_uring-changes-from-libuv-1.45.0.patch \ + file://0001-deps-disable-io_uring-support-in-libuv.patch \ file://run-ptest \ " @@ -38,7 +37,7 @@ SRC_URI:append:toolchain-clang:x86 = " \ SRC_URI:append:toolchain-clang:powerpc64le = " \ file://0001-ppc64-Do-not-use-mminimal-toc-with-clang.patch \ " -SRC_URI[sha256sum] = "31807ebeeeb049c53f1765e4a95aed69476a4b696dd100cb539ab668d7950b40" +SRC_URI[sha256sum] = "77813edbf3f7f16d2d35d3353443dee4e61d5ee84d9e3138c7538a3c0ca5209e" S = "${WORKDIR}/node-v${PV}"