aboutsummaryrefslogtreecommitdiff
path: root/tests/sys/aio/aio_test.c
diff options
context:
space:
mode:
Diffstat (limited to 'tests/sys/aio/aio_test.c')
-rw-r--r--tests/sys/aio/aio_test.c2101
1 files changed, 2101 insertions, 0 deletions
diff --git a/tests/sys/aio/aio_test.c b/tests/sys/aio/aio_test.c
new file mode 100644
index 000000000000..def8a9d548d6
--- /dev/null
+++ b/tests/sys/aio/aio_test.c
@@ -0,0 +1,2101 @@
+/*-
+ * Copyright (c) 2004 Robert N. M. Watson
+ * All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
+ * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
+ * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+ * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
+ * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
+ * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
+ * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
+ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
+ * SUCH DAMAGE.
+ */
+
+/*
+ * Regression test to do some very basic AIO exercising on several types of
+ * file descriptors. Currently, the tests consist of initializing a fixed
+ * size buffer with pseudo-random data, writing it to one fd using AIO, then
+ * reading it from a second descriptor using AIO. For some targets, the same
+ * fd is used for write and read (i.e., file, md device), but for others the
+ * operation is performed on a peer (pty, socket, fifo, etc). For each file
+ * descriptor type, several completion methods are tested. This test program
+ * does not attempt to exercise error cases or more subtle asynchronous
+ * behavior, just make sure that the basic operations work on some basic object
+ * types.
+ */
+
+#include <sys/param.h>
+#include <sys/event.h>
+#include <sys/mdioctl.h>
+#include <sys/module.h>
+#include <sys/resource.h>
+#include <sys/socket.h>
+#include <sys/stat.h>
+#include <sys/un.h>
+
+#include <aio.h>
+#include <err.h>
+#include <errno.h>
+#include <fcntl.h>
+#include <libutil.h>
+#include <limits.h>
+#include <semaphore.h>
+#include <signal.h>
+#include <stdint.h>
+#include <stdio.h>
+#include <stdlib.h>
+#include <string.h>
+#include <termios.h>
+#include <unistd.h>
+
+#include <atf-c.h>
+
+#include "local.h"
+
+/*
+ * GLOBAL_MAX sets the largest usable buffer size to be read and written, as
+ * it sizes ac_buffer in the aio_context structure. It is also the default
+ * size for file I/O. For other types, we use smaller blocks or we risk
+ * blocking (and we run in a single process/thread so that would be bad).
+ */
+#define GLOBAL_MAX 16384
+
+#define BUFFER_MAX GLOBAL_MAX
+
+/*
+ * A completion function will block until the aio has completed, then return
+ * the result of the aio. errno will be set appropriately.
+ */
+typedef ssize_t (*completion)(struct aiocb*);
+
+struct aio_context {
+ int ac_read_fd, ac_write_fd;
+ long ac_seed;
+ char ac_buffer[GLOBAL_MAX];
+ int ac_buflen;
+ int ac_seconds;
+};
+
+static sem_t completions;
+
+/*
+ * Fill a buffer given a seed that can be fed into srandom() to initialize
+ * the PRNG in a repeatable manner.
+ */
+static void
+aio_fill_buffer(char *buffer, int len, long seed)
+{
+ char ch;
+ int i;
+
+ srandom(seed);
+ for (i = 0; i < len; i++) {
+ ch = random() & 0xff;
+ buffer[i] = ch;
+ }
+}
+
+/*
+ * Test that a buffer matches a given seed. See aio_fill_buffer(). Return
+ * (1) on a match, (0) on a mismatch.
+ */
+static int
+aio_test_buffer(char *buffer, int len, long seed)
+{
+ char ch;
+ int i;
+
+ srandom(seed);
+ for (i = 0; i < len; i++) {
+ ch = random() & 0xff;
+ if (buffer[i] != ch)
+ return (0);
+ }
+ return (1);
+}
+
+/*
+ * Initialize a testing context given the file descriptors provided by the
+ * test setup.
+ */
+static void
+aio_context_init(struct aio_context *ac, int read_fd,
+ int write_fd, int buflen)
+{
+
+ ATF_REQUIRE_MSG(buflen <= BUFFER_MAX,
+ "aio_context_init: buffer too large (%d > %d)",
+ buflen, BUFFER_MAX);
+ bzero(ac, sizeof(*ac));
+ ac->ac_read_fd = read_fd;
+ ac->ac_write_fd = write_fd;
+ ac->ac_buflen = buflen;
+ srandomdev();
+ ac->ac_seed = random();
+ aio_fill_buffer(ac->ac_buffer, buflen, ac->ac_seed);
+ ATF_REQUIRE_MSG(aio_test_buffer(ac->ac_buffer, buflen,
+ ac->ac_seed) != 0, "aio_test_buffer: internal error");
+}
+
+static ssize_t
+poll(struct aiocb *aio)
+{
+ int error;
+
+ while ((error = aio_error(aio)) == EINPROGRESS)
+ usleep(25000);
+ if (error)
+ return (error);
+ else
+ return (aio_return(aio));
+}
+
+static void
+sigusr1_handler(int sig __unused)
+{
+ ATF_REQUIRE_EQ(0, sem_post(&completions));
+}
+
+static void
+thr_handler(union sigval sv __unused)
+{
+ ATF_REQUIRE_EQ(0, sem_post(&completions));
+}
+
+static ssize_t
+poll_signaled(struct aiocb *aio)
+{
+ int error;
+
+ ATF_REQUIRE_EQ(0, sem_wait(&completions));
+ error = aio_error(aio);
+ switch (error) {
+ case EINPROGRESS:
+ errno = EINTR;
+ return (-1);
+ case 0:
+ return (aio_return(aio));
+ default:
+ return (error);
+ }
+}
+
+/*
+ * Setup a signal handler for signal delivery tests
+ * This isn't thread safe, but it's ok since ATF runs each testcase in a
+ * separate process
+ */
+static struct sigevent*
+setup_signal(void)
+{
+ static struct sigevent sev;
+
+ ATF_REQUIRE_EQ(0, sem_init(&completions, false, 0));
+ sev.sigev_notify = SIGEV_SIGNAL;
+ sev.sigev_signo = SIGUSR1;
+ ATF_REQUIRE(SIG_ERR != signal(SIGUSR1, sigusr1_handler));
+ return (&sev);
+}
+
+/*
+ * Setup a thread for thread delivery tests
+ * This isn't thread safe, but it's ok since ATF runs each testcase in a
+ * separate process
+ */
+static struct sigevent*
+setup_thread(void)
+{
+ static struct sigevent sev;
+
+ ATF_REQUIRE_EQ(0, sem_init(&completions, false, 0));
+ sev.sigev_notify = SIGEV_THREAD;
+ sev.sigev_notify_function = thr_handler;
+ sev.sigev_notify_attributes = NULL;
+ return (&sev);
+}
+
+static ssize_t
+suspend(struct aiocb *aio)
+{
+ const struct aiocb *const iocbs[] = {aio};
+ int error;
+
+ error = aio_suspend(iocbs, 1, NULL);
+ if (error == 0)
+ return (aio_return(aio));
+ else
+ return (error);
+}
+
+static ssize_t
+waitcomplete(struct aiocb *aio)
+{
+ struct aiocb *aiop;
+ ssize_t ret;
+
+ ret = aio_waitcomplete(&aiop, NULL);
+ ATF_REQUIRE_EQ(aio, aiop);
+ return (ret);
+}
+
+/*
+ * Setup an iocb for kqueue notification. This isn't thread
+ * safe, but it's ok because ATF runs every test case in a separate process.
+ */
+static struct sigevent*
+setup_kqueue(void)
+{
+ static struct sigevent sev;
+ static int kq;
+
+ kq = kqueue();
+ ATF_REQUIRE(kq >= 0);
+
+ memset(&sev, 0, sizeof(sev));
+ sev.sigev_notify_kqueue = kq;
+ sev.sigev_value.sival_ptr = (void*)0xdeadbeef;
+ sev.sigev_notify = SIGEV_KEVENT;
+
+ return (&sev);
+}
+
+static ssize_t
+poll_kqueue(struct aiocb *aio)
+{
+ int kq, nevents;
+ struct kevent events[1];
+
+ kq = aio->aio_sigevent.sigev_notify_kqueue;
+
+ nevents = kevent(kq, NULL, 0, events, 1, NULL);
+ ATF_CHECK_EQ(1, nevents);
+ ATF_CHECK_EQ(events[0].ident, (uintptr_t) aio);
+ ATF_CHECK_EQ(events[0].filter, EVFILT_AIO);
+ ATF_CHECK_EQ(events[0].flags, EV_EOF);
+ ATF_CHECK_EQ(events[0].fflags, 0);
+ ATF_CHECK_EQ(events[0].data, 0);
+ ATF_CHECK_EQ((uintptr_t)events[0].udata, 0xdeadbeef);
+
+ return (aio_return(aio));
+}
+
+/*
+ * Perform a simple write test of our initialized data buffer to the provided
+ * file descriptor.
+ */
+static void
+aio_write_test(struct aio_context *ac, completion comp, struct sigevent *sev)
+{
+ struct aiocb aio;
+ ssize_t len;
+
+ bzero(&aio, sizeof(aio));
+ aio.aio_buf = ac->ac_buffer;
+ aio.aio_nbytes = ac->ac_buflen;
+ aio.aio_fildes = ac->ac_write_fd;
+ aio.aio_offset = 0;
+ if (sev)
+ aio.aio_sigevent = *sev;
+
+ if (aio_write(&aio) < 0)
+ atf_tc_fail("aio_write failed: %s", strerror(errno));
+
+ len = comp(&aio);
+ if (len < 0)
+ atf_tc_fail("aio failed: %s", strerror(errno));
+
+ if (len != ac->ac_buflen)
+ atf_tc_fail("aio short write (%jd)", (intmax_t)len);
+}
+
+/*
+ * Perform a vectored I/O test of our initialized data buffer to the provided
+ * file descriptor.
+ *
+ * To vectorize the linear buffer, chop it up into two pieces of dissimilar
+ * size, and swap their offsets.
+ */
+static void
+aio_writev_test(struct aio_context *ac, completion comp, struct sigevent *sev)
+{
+ struct aiocb aio;
+ struct iovec iov[2];
+ size_t len0, len1;
+ ssize_t len;
+
+ bzero(&aio, sizeof(aio));
+
+ aio.aio_fildes = ac->ac_write_fd;
+ aio.aio_offset = 0;
+ len0 = ac->ac_buflen * 3 / 4;
+ len1 = ac->ac_buflen / 4;
+ iov[0].iov_base = ac->ac_buffer + len1;
+ iov[0].iov_len = len0;
+ iov[1].iov_base = ac->ac_buffer;
+ iov[1].iov_len = len1;
+ aio.aio_iov = iov;
+ aio.aio_iovcnt = 2;
+ if (sev)
+ aio.aio_sigevent = *sev;
+
+ if (aio_writev(&aio) < 0)
+ atf_tc_fail("aio_writev failed: %s", strerror(errno));
+
+ len = comp(&aio);
+ if (len < 0)
+ atf_tc_fail("aio failed: %s", strerror(errno));
+
+ if (len != ac->ac_buflen)
+ atf_tc_fail("aio short write (%jd)", (intmax_t)len);
+}
+
+/*
+ * Perform a simple read test of our initialized data buffer from the
+ * provided file descriptor.
+ */
+static void
+aio_read_test(struct aio_context *ac, completion comp, struct sigevent *sev)
+{
+ struct aiocb aio;
+ ssize_t len;
+
+ bzero(ac->ac_buffer, ac->ac_buflen);
+ bzero(&aio, sizeof(aio));
+ aio.aio_buf = ac->ac_buffer;
+ aio.aio_nbytes = ac->ac_buflen;
+ aio.aio_fildes = ac->ac_read_fd;
+ aio.aio_offset = 0;
+ if (sev)
+ aio.aio_sigevent = *sev;
+
+ if (aio_read(&aio) < 0)
+ atf_tc_fail("aio_read failed: %s", strerror(errno));
+
+ len = comp(&aio);
+ if (len < 0)
+ atf_tc_fail("aio failed: %s", strerror(errno));
+
+ ATF_REQUIRE_EQ_MSG(len, ac->ac_buflen,
+ "aio short read (%jd)", (intmax_t)len);
+
+ if (aio_test_buffer(ac->ac_buffer, ac->ac_buflen, ac->ac_seed) == 0)
+ atf_tc_fail("buffer mismatched");
+}
+
+static void
+aio_readv_test(struct aio_context *ac, completion comp, struct sigevent *sev)
+{
+ struct aiocb aio;
+ struct iovec iov[2];
+ size_t len0, len1;
+ ssize_t len;
+
+ bzero(ac->ac_buffer, ac->ac_buflen);
+ bzero(&aio, sizeof(aio));
+ aio.aio_fildes = ac->ac_read_fd;
+ aio.aio_offset = 0;
+ len0 = ac->ac_buflen * 3 / 4;
+ len1 = ac->ac_buflen / 4;
+ iov[0].iov_base = ac->ac_buffer + len1;
+ iov[0].iov_len = len0;
+ iov[1].iov_base = ac->ac_buffer;
+ iov[1].iov_len = len1;
+ aio.aio_iov = iov;
+ aio.aio_iovcnt = 2;
+ if (sev)
+ aio.aio_sigevent = *sev;
+
+ if (aio_readv(&aio) < 0)
+ atf_tc_fail("aio_read failed: %s", strerror(errno));
+
+ len = comp(&aio);
+ if (len < 0)
+ atf_tc_fail("aio failed: %s", strerror(errno));
+
+ ATF_REQUIRE_EQ_MSG(len, ac->ac_buflen,
+ "aio short read (%jd)", (intmax_t)len);
+
+ if (aio_test_buffer(ac->ac_buffer, ac->ac_buflen, ac->ac_seed) == 0)
+ atf_tc_fail("buffer mismatched");
+}
+
+/*
+ * Series of type-specific tests for AIO. For now, we just make sure we can
+ * issue a write and then a read to each type. We assume that once a write
+ * is issued, a read can follow.
+ */
+
+/*
+ * Test with a classic file. Assumes we can create a moderate size temporary
+ * file.
+ */
+#define FILE_LEN GLOBAL_MAX
+#define FILE_PATHNAME "testfile"
+
+static void
+aio_file_test(completion comp, struct sigevent *sev, bool vectored)
+{
+ struct aio_context ac;
+ int fd;
+
+ ATF_REQUIRE_UNSAFE_AIO();
+
+ fd = open(FILE_PATHNAME, O_RDWR | O_CREAT, 0600);
+ ATF_REQUIRE_MSG(fd != -1, "open failed: %s", strerror(errno));
+
+ aio_context_init(&ac, fd, fd, FILE_LEN);
+ if (vectored) {
+ aio_writev_test(&ac, comp, sev);
+ aio_readv_test(&ac, comp, sev);
+ } else {
+ aio_write_test(&ac, comp, sev);
+ aio_read_test(&ac, comp, sev);
+ }
+ close(fd);
+}
+
+ATF_TC_WITHOUT_HEAD(file_kq);
+ATF_TC_BODY(file_kq, tc)
+{
+ aio_file_test(poll_kqueue, setup_kqueue(), false);
+}
+
+ATF_TC_WITHOUT_HEAD(file_poll);
+ATF_TC_BODY(file_poll, tc)
+{
+ aio_file_test(poll, NULL, false);
+}
+
+ATF_TC_WITHOUT_HEAD(file_signal);
+ATF_TC_BODY(file_signal, tc)
+{
+ aio_file_test(poll_signaled, setup_signal(), false);
+}
+
+ATF_TC_WITHOUT_HEAD(file_suspend);
+ATF_TC_BODY(file_suspend, tc)
+{
+ aio_file_test(suspend, NULL, false);
+}
+
+ATF_TC_WITHOUT_HEAD(file_thread);
+ATF_TC_BODY(file_thread, tc)
+{
+ aio_file_test(poll_signaled, setup_thread(), false);
+}
+
+ATF_TC_WITHOUT_HEAD(file_waitcomplete);
+ATF_TC_BODY(file_waitcomplete, tc)
+{
+ aio_file_test(waitcomplete, NULL, false);
+}
+
+#define FIFO_LEN 256
+#define FIFO_PATHNAME "testfifo"
+
+static void
+aio_fifo_test(completion comp, struct sigevent *sev)
+{
+ int error, read_fd = -1, write_fd = -1;
+ struct aio_context ac;
+
+ ATF_REQUIRE_UNSAFE_AIO();
+
+ ATF_REQUIRE_MSG(mkfifo(FIFO_PATHNAME, 0600) != -1,
+ "mkfifo failed: %s", strerror(errno));
+
+ read_fd = open(FIFO_PATHNAME, O_RDONLY | O_NONBLOCK);
+ if (read_fd == -1) {
+ error = errno;
+ errno = error;
+ atf_tc_fail("read_fd open failed: %s",
+ strerror(errno));
+ }
+
+ write_fd = open(FIFO_PATHNAME, O_WRONLY);
+ if (write_fd == -1) {
+ error = errno;
+ errno = error;
+ atf_tc_fail("write_fd open failed: %s",
+ strerror(errno));
+ }
+
+ aio_context_init(&ac, read_fd, write_fd, FIFO_LEN);
+ aio_write_test(&ac, comp, sev);
+ aio_read_test(&ac, comp, sev);
+
+ close(read_fd);
+ close(write_fd);
+}
+
+ATF_TC_WITHOUT_HEAD(fifo_kq);
+ATF_TC_BODY(fifo_kq, tc)
+{
+ aio_fifo_test(poll_kqueue, setup_kqueue());
+}
+
+ATF_TC_WITHOUT_HEAD(fifo_poll);
+ATF_TC_BODY(fifo_poll, tc)
+{
+ aio_fifo_test(poll, NULL);
+}
+
+ATF_TC_WITHOUT_HEAD(fifo_signal);
+ATF_TC_BODY(fifo_signal, tc)
+{
+ aio_fifo_test(poll_signaled, setup_signal());
+}
+
+ATF_TC_WITHOUT_HEAD(fifo_suspend);
+ATF_TC_BODY(fifo_suspend, tc)
+{
+ aio_fifo_test(suspend, NULL);
+}
+
+ATF_TC_WITHOUT_HEAD(fifo_thread);
+ATF_TC_BODY(fifo_thread, tc)
+{
+ aio_fifo_test(poll_signaled, setup_thread());
+}
+
+ATF_TC_WITHOUT_HEAD(fifo_waitcomplete);
+ATF_TC_BODY(fifo_waitcomplete, tc)
+{
+ aio_fifo_test(waitcomplete, NULL);
+}
+
+#define UNIX_SOCKETPAIR_LEN 256
+static void
+aio_unix_socketpair_test(completion comp, struct sigevent *sev, bool vectored)
+{
+ struct aio_context ac;
+ struct rusage ru_before, ru_after;
+ int sockets[2];
+
+ ATF_REQUIRE_MSG(socketpair(PF_UNIX, SOCK_STREAM, 0, sockets) != -1,
+ "socketpair failed: %s", strerror(errno));
+
+ aio_context_init(&ac, sockets[0], sockets[1], UNIX_SOCKETPAIR_LEN);
+ ATF_REQUIRE_MSG(getrusage(RUSAGE_SELF, &ru_before) != -1,
+ "getrusage failed: %s", strerror(errno));
+ if (vectored) {
+ aio_writev_test(&ac, comp, sev);
+ aio_readv_test(&ac, comp, sev);
+ } else {
+ aio_write_test(&ac, comp, sev);
+ aio_read_test(&ac, comp, sev);
+ }
+ ATF_REQUIRE_MSG(getrusage(RUSAGE_SELF, &ru_after) != -1,
+ "getrusage failed: %s", strerror(errno));
+ ATF_REQUIRE(ru_after.ru_msgsnd == ru_before.ru_msgsnd + 1);
+ ATF_REQUIRE(ru_after.ru_msgrcv == ru_before.ru_msgrcv + 1);
+
+ close(sockets[0]);
+ close(sockets[1]);
+}
+
+ATF_TC_WITHOUT_HEAD(socket_kq);
+ATF_TC_BODY(socket_kq, tc)
+{
+ aio_unix_socketpair_test(poll_kqueue, setup_kqueue(), false);
+}
+
+ATF_TC_WITHOUT_HEAD(socket_poll);
+ATF_TC_BODY(socket_poll, tc)
+{
+ aio_unix_socketpair_test(poll, NULL, false);
+}
+
+ATF_TC_WITHOUT_HEAD(socket_signal);
+ATF_TC_BODY(socket_signal, tc)
+{
+ aio_unix_socketpair_test(poll_signaled, setup_signal(), false);
+}
+
+ATF_TC_WITHOUT_HEAD(socket_suspend);
+ATF_TC_BODY(socket_suspend, tc)
+{
+ aio_unix_socketpair_test(suspend, NULL, false);
+}
+
+ATF_TC_WITHOUT_HEAD(socket_thread);
+ATF_TC_BODY(socket_thread, tc)
+{
+ aio_unix_socketpair_test(poll_signaled, setup_thread(), false);
+}
+
+ATF_TC_WITHOUT_HEAD(socket_waitcomplete);
+ATF_TC_BODY(socket_waitcomplete, tc)
+{
+ aio_unix_socketpair_test(waitcomplete, NULL, false);
+}
+
+struct aio_pty_arg {
+ int apa_read_fd;
+ int apa_write_fd;
+};
+
+#define PTY_LEN 256
+static void
+aio_pty_test(completion comp, struct sigevent *sev)
+{
+ struct aio_context ac;
+ int read_fd, write_fd;
+ struct termios ts;
+ int error;
+
+ ATF_REQUIRE_UNSAFE_AIO();
+
+ ATF_REQUIRE_MSG(openpty(&read_fd, &write_fd, NULL, NULL, NULL) == 0,
+ "openpty failed: %s", strerror(errno));
+
+
+ if (tcgetattr(write_fd, &ts) < 0) {
+ error = errno;
+ errno = error;
+ atf_tc_fail("tcgetattr failed: %s", strerror(errno));
+ }
+ cfmakeraw(&ts);
+ if (tcsetattr(write_fd, TCSANOW, &ts) < 0) {
+ error = errno;
+ errno = error;
+ atf_tc_fail("tcsetattr failed: %s", strerror(errno));
+ }
+ aio_context_init(&ac, read_fd, write_fd, PTY_LEN);
+
+ aio_write_test(&ac, comp, sev);
+ aio_read_test(&ac, comp, sev);
+
+ close(read_fd);
+ close(write_fd);
+}
+
+ATF_TC_WITHOUT_HEAD(pty_kq);
+ATF_TC_BODY(pty_kq, tc)
+{
+ aio_pty_test(poll_kqueue, setup_kqueue());
+}
+
+ATF_TC_WITHOUT_HEAD(pty_poll);
+ATF_TC_BODY(pty_poll, tc)
+{
+ aio_pty_test(poll, NULL);
+}
+
+ATF_TC_WITHOUT_HEAD(pty_signal);
+ATF_TC_BODY(pty_signal, tc)
+{
+ aio_pty_test(poll_signaled, setup_signal());
+}
+
+ATF_TC_WITHOUT_HEAD(pty_suspend);
+ATF_TC_BODY(pty_suspend, tc)
+{
+ aio_pty_test(suspend, NULL);
+}
+
+ATF_TC_WITHOUT_HEAD(pty_thread);
+ATF_TC_BODY(pty_thread, tc)
+{
+ aio_pty_test(poll_signaled, setup_thread());
+}
+
+ATF_TC_WITHOUT_HEAD(pty_waitcomplete);
+ATF_TC_BODY(pty_waitcomplete, tc)
+{
+ aio_pty_test(waitcomplete, NULL);
+}
+
+#define PIPE_LEN 256
+static void
+aio_pipe_test(completion comp, struct sigevent *sev)
+{
+ struct aio_context ac;
+ int pipes[2];
+
+ ATF_REQUIRE_UNSAFE_AIO();
+
+ ATF_REQUIRE_MSG(pipe(pipes) != -1,
+ "pipe failed: %s", strerror(errno));
+
+ aio_context_init(&ac, pipes[0], pipes[1], PIPE_LEN);
+ aio_write_test(&ac, comp, sev);
+ aio_read_test(&ac, comp, sev);
+
+ close(pipes[0]);
+ close(pipes[1]);
+}
+
+ATF_TC_WITHOUT_HEAD(pipe_kq);
+ATF_TC_BODY(pipe_kq, tc)
+{
+ aio_pipe_test(poll_kqueue, setup_kqueue());
+}
+
+ATF_TC_WITHOUT_HEAD(pipe_poll);
+ATF_TC_BODY(pipe_poll, tc)
+{
+ aio_pipe_test(poll, NULL);
+}
+
+ATF_TC_WITHOUT_HEAD(pipe_signal);
+ATF_TC_BODY(pipe_signal, tc)
+{
+ aio_pipe_test(poll_signaled, setup_signal());
+}
+
+ATF_TC_WITHOUT_HEAD(pipe_suspend);
+ATF_TC_BODY(pipe_suspend, tc)
+{
+ aio_pipe_test(suspend, NULL);
+}
+
+ATF_TC_WITHOUT_HEAD(pipe_thread);
+ATF_TC_BODY(pipe_thread, tc)
+{
+ aio_pipe_test(poll_signaled, setup_thread());
+}
+
+ATF_TC_WITHOUT_HEAD(pipe_waitcomplete);
+ATF_TC_BODY(pipe_waitcomplete, tc)
+{
+ aio_pipe_test(waitcomplete, NULL);
+}
+
+#define DEVICE_IO_LEN GLOBAL_MAX
+#define MDUNIT_LINK "mdunit_link"
+
+static int
+aio_md_setup(void)
+{
+ int error, fd, mdctl_fd, unit;
+ char pathname[PATH_MAX];
+ struct md_ioctl mdio;
+ char buf[80];
+
+ mdctl_fd = open("/dev/" MDCTL_NAME, O_RDWR, 0);
+ ATF_REQUIRE_MSG(mdctl_fd != -1,
+ "opening /dev/%s failed: %s", MDCTL_NAME, strerror(errno));
+
+ bzero(&mdio, sizeof(mdio));
+ mdio.md_version = MDIOVERSION;
+ mdio.md_type = MD_MALLOC;
+ mdio.md_options = MD_AUTOUNIT | MD_COMPRESS;
+ mdio.md_mediasize = 1024 * 1024; /* 1 MB, enough for max_buf_aio up to 2047 */
+ mdio.md_sectorsize = 512;
+ strlcpy(buf, __func__, sizeof(buf));
+ mdio.md_label = buf;
+
+ if (ioctl(mdctl_fd, MDIOCATTACH, &mdio) < 0) {
+ error = errno;
+ errno = error;
+ atf_tc_fail("ioctl MDIOCATTACH failed: %s", strerror(errno));
+ }
+ close(mdctl_fd);
+
+ /* Store the md unit number in a symlink for future cleanup */
+ unit = mdio.md_unit;
+ snprintf(buf, sizeof(buf), "%d", unit);
+ ATF_REQUIRE_EQ(0, symlink(buf, MDUNIT_LINK));
+ snprintf(pathname, PATH_MAX, "/dev/md%d", unit);
+ fd = open(pathname, O_RDWR);
+ ATF_REQUIRE_MSG(fd != -1,
+ "opening %s failed: %s", pathname, strerror(errno));
+
+ return (fd);
+}
+
+static void
+aio_md_cleanup(void)
+{
+ struct md_ioctl mdio;
+ int mdctl_fd, n, unit;
+ char buf[80];
+
+ mdctl_fd = open("/dev/" MDCTL_NAME, O_RDWR, 0);
+ if (mdctl_fd < 0) {
+ fprintf(stderr, "opening /dev/%s failed: %s\n", MDCTL_NAME,
+ strerror(errno));
+ return;
+ }
+ n = readlink(MDUNIT_LINK, buf, sizeof(buf) - 1);
+ if (n > 0) {
+ buf[n] = '\0';
+ if (sscanf(buf, "%d", &unit) == 1 && unit >= 0) {
+ bzero(&mdio, sizeof(mdio));
+ mdio.md_version = MDIOVERSION;
+ mdio.md_unit = unit;
+ if (ioctl(mdctl_fd, MDIOCDETACH, &mdio) == -1) {
+ fprintf(stderr,
+ "ioctl MDIOCDETACH unit %d failed: %s\n",
+ unit, strerror(errno));
+ }
+ }
+ }
+
+ close(mdctl_fd);
+}
+
+static void
+aio_md_test(completion comp, struct sigevent *sev, bool vectored)
+{
+ struct aio_context ac;
+ int fd;
+
+ fd = aio_md_setup();
+ aio_context_init(&ac, fd, fd, DEVICE_IO_LEN);
+ if (vectored) {
+ aio_writev_test(&ac, comp, sev);
+ aio_readv_test(&ac, comp, sev);
+ } else {
+ aio_write_test(&ac, comp, sev);
+ aio_read_test(&ac, comp, sev);
+ }
+
+ close(fd);
+}
+
+ATF_TC_WITH_CLEANUP(md_kq);
+ATF_TC_HEAD(md_kq, tc)
+{
+
+ atf_tc_set_md_var(tc, "require.user", "root");
+}
+ATF_TC_BODY(md_kq, tc)
+{
+ aio_md_test(poll_kqueue, setup_kqueue(), false);
+}
+ATF_TC_CLEANUP(md_kq, tc)
+{
+ aio_md_cleanup();
+}
+
+ATF_TC_WITH_CLEANUP(md_poll);
+ATF_TC_HEAD(md_poll, tc)
+{
+
+ atf_tc_set_md_var(tc, "require.user", "root");
+}
+ATF_TC_BODY(md_poll, tc)
+{
+ aio_md_test(poll, NULL, false);
+}
+ATF_TC_CLEANUP(md_poll, tc)
+{
+ aio_md_cleanup();
+}
+
+ATF_TC_WITH_CLEANUP(md_signal);
+ATF_TC_HEAD(md_signal, tc)
+{
+
+ atf_tc_set_md_var(tc, "require.user", "root");
+}
+ATF_TC_BODY(md_signal, tc)
+{
+ aio_md_test(poll_signaled, setup_signal(), false);
+}
+ATF_TC_CLEANUP(md_signal, tc)
+{
+ aio_md_cleanup();
+}
+
+ATF_TC_WITH_CLEANUP(md_suspend);
+ATF_TC_HEAD(md_suspend, tc)
+{
+
+ atf_tc_set_md_var(tc, "require.user", "root");
+}
+ATF_TC_BODY(md_suspend, tc)
+{
+ aio_md_test(suspend, NULL, false);
+}
+ATF_TC_CLEANUP(md_suspend, tc)
+{
+ aio_md_cleanup();
+}
+
+ATF_TC_WITH_CLEANUP(md_thread);
+ATF_TC_HEAD(md_thread, tc)
+{
+
+ atf_tc_set_md_var(tc, "require.user", "root");
+}
+ATF_TC_BODY(md_thread, tc)
+{
+ aio_md_test(poll_signaled, setup_thread(), false);
+}
+ATF_TC_CLEANUP(md_thread, tc)
+{
+ aio_md_cleanup();
+}
+
+ATF_TC_WITH_CLEANUP(md_waitcomplete);
+ATF_TC_HEAD(md_waitcomplete, tc)
+{
+
+ atf_tc_set_md_var(tc, "require.user", "root");
+}
+ATF_TC_BODY(md_waitcomplete, tc)
+{
+ aio_md_test(waitcomplete, NULL, false);
+}
+ATF_TC_CLEANUP(md_waitcomplete, tc)
+{
+ aio_md_cleanup();
+}
+
+#define ZVOL_VDEV_PATHNAME "test_vdev"
+#define POOL_SIZE (1 << 28) /* 256 MB */
+#define ZVOL_SIZE "64m"
+#define POOL_NAME "aio_testpool"
+#define ZVOL_NAME "aio_testvol"
+
+static int
+aio_zvol_setup(const char *unique)
+{
+ FILE *pidfile;
+ int fd;
+ pid_t pid;
+ char vdev_name[160];
+ char pool_name[80];
+ char cmd[160];
+ char zvol_name[160];
+ char devname[160];
+
+ pid = getpid();
+ snprintf(vdev_name, sizeof(vdev_name), "%s", ZVOL_VDEV_PATHNAME);
+ snprintf(pool_name, sizeof(pool_name), "%s_%s.%d", POOL_NAME, unique,
+ pid);
+ snprintf(zvol_name, sizeof(zvol_name), "%s/%s_%s", pool_name, ZVOL_NAME,
+ unique);
+
+ fd = open(vdev_name, O_RDWR | O_CREAT, 0600);
+ ATF_REQUIRE_MSG(fd != -1, "open failed: %s", strerror(errno));
+ ATF_REQUIRE_EQ_MSG(0,
+ ftruncate(fd, POOL_SIZE), "ftruncate failed: %s", strerror(errno));
+ close(fd);
+
+ pidfile = fopen("pidfile", "w");
+ ATF_REQUIRE_MSG(NULL != pidfile, "fopen: %s", strerror(errno));
+ fprintf(pidfile, "%d", pid);
+ fclose(pidfile);
+
+ snprintf(cmd, sizeof(cmd), "zpool create %s $PWD/%s", pool_name,
+ vdev_name);
+ ATF_REQUIRE_EQ_MSG(0, system(cmd),
+ "zpool create failed: %s", strerror(errno));
+ snprintf(cmd, sizeof(cmd),
+ "zfs create -o volblocksize=8192 -o volmode=dev -V %s %s",
+ ZVOL_SIZE, zvol_name);
+ ATF_REQUIRE_EQ_MSG(0, system(cmd),
+ "zfs create failed: %s", strerror(errno));
+
+ snprintf(devname, sizeof(devname), "/dev/zvol/%s", zvol_name);
+ do {
+ fd = open(devname, O_RDWR);
+ } while (fd == -1 && errno == EINTR);
+ ATF_REQUIRE_MSG(fd != -1, "open failed: %s", strerror(errno));
+ return (fd);
+}
+
+static void
+aio_zvol_cleanup(const char *unique)
+{
+ FILE *pidfile;
+ pid_t testpid;
+ char cmd[160];
+
+ pidfile = fopen("pidfile", "r");
+ if (pidfile == NULL && errno == ENOENT) {
+ /* Setup probably failed */
+ return;
+ }
+ ATF_REQUIRE_MSG(NULL != pidfile, "fopen: %s", strerror(errno));
+ ATF_REQUIRE_EQ(1, fscanf(pidfile, "%d", &testpid));
+ fclose(pidfile);
+
+ snprintf(cmd, sizeof(cmd), "zpool destroy %s_%s.%d", POOL_NAME, unique,
+ testpid);
+ system(cmd);
+}
+
+
+ATF_TC_WITHOUT_HEAD(aio_large_read_test);
+ATF_TC_BODY(aio_large_read_test, tc)
+{
+ struct aiocb cb, *cbp;
+ ssize_t nread;
+ size_t len;
+ int fd;
+#ifdef __LP64__
+ int clamped;
+#endif
+
+ ATF_REQUIRE_UNSAFE_AIO();
+
+#ifdef __LP64__
+ len = sizeof(clamped);
+ if (sysctlbyname("debug.iosize_max_clamp", &clamped, &len, NULL, 0) ==
+ -1)
+ atf_libc_error(errno, "Failed to read debug.iosize_max_clamp");
+#endif
+
+ /* Determine the maximum supported read(2) size. */
+ len = SSIZE_MAX;
+#ifdef __LP64__
+ if (clamped)
+ len = INT_MAX;
+#endif
+
+ fd = open(FILE_PATHNAME, O_RDWR | O_CREAT, 0600);
+ ATF_REQUIRE_MSG(fd != -1, "open failed: %s", strerror(errno));
+
+ unlink(FILE_PATHNAME);
+
+ memset(&cb, 0, sizeof(cb));
+ cb.aio_nbytes = len;
+ cb.aio_fildes = fd;
+ cb.aio_buf = NULL;
+ if (aio_read(&cb) == -1)
+ atf_tc_fail("aio_read() of maximum read size failed: %s",
+ strerror(errno));
+
+ nread = aio_waitcomplete(&cbp, NULL);
+ if (nread == -1)
+ atf_tc_fail("aio_waitcomplete() failed: %s", strerror(errno));
+ if (nread != 0)
+ atf_tc_fail("aio_read() from empty file returned data: %zd",
+ nread);
+
+ memset(&cb, 0, sizeof(cb));
+ cb.aio_nbytes = len + 1;
+ cb.aio_fildes = fd;
+ cb.aio_buf = NULL;
+ if (aio_read(&cb) == -1) {
+ if (errno == EINVAL)
+ goto finished;
+ atf_tc_fail("aio_read() of too large read size failed: %s",
+ strerror(errno));
+ }
+
+ nread = aio_waitcomplete(&cbp, NULL);
+ if (nread == -1) {
+ if (errno == EINVAL)
+ goto finished;
+ atf_tc_fail("aio_waitcomplete() failed: %s", strerror(errno));
+ }
+ atf_tc_fail("aio_read() of too large read size returned: %zd", nread);
+
+finished:
+ close(fd);
+}
+
+/*
+ * This tests for a bug where arriving socket data can wakeup multiple
+ * AIO read requests resulting in an uncancellable request.
+ */
+ATF_TC_WITHOUT_HEAD(aio_socket_two_reads);
+ATF_TC_BODY(aio_socket_two_reads, tc)
+{
+ struct ioreq {
+ struct aiocb iocb;
+ char buffer[1024];
+ } ioreq[2];
+ struct aiocb *iocb;
+ unsigned i;
+ int s[2];
+ char c;
+
+#if __FreeBSD_version < 1100101
+ aft_tc_skip("kernel version %d is too old (%d required)",
+ __FreeBSD_version, 1100101);
+#endif
+
+ ATF_REQUIRE(socketpair(PF_UNIX, SOCK_STREAM, 0, s) != -1);
+
+ /* Queue two read requests. */
+ memset(&ioreq, 0, sizeof(ioreq));
+ for (i = 0; i < nitems(ioreq); i++) {
+ ioreq[i].iocb.aio_nbytes = sizeof(ioreq[i].buffer);
+ ioreq[i].iocb.aio_fildes = s[0];
+ ioreq[i].iocb.aio_buf = ioreq[i].buffer;
+ ATF_REQUIRE(aio_read(&ioreq[i].iocb) == 0);
+ }
+
+ /* Send a single byte. This should complete one request. */
+ c = 0xc3;
+ ATF_REQUIRE(write(s[1], &c, sizeof(c)) == 1);
+
+ ATF_REQUIRE(aio_waitcomplete(&iocb, NULL) == 1);
+
+ /* Determine which request completed and verify the data was read. */
+ if (iocb == &ioreq[0].iocb)
+ i = 0;
+ else
+ i = 1;
+ ATF_REQUIRE(ioreq[i].buffer[0] == c);
+
+ i ^= 1;
+
+ /*
+ * Try to cancel the other request. On broken systems this
+ * will fail and the process will hang on exit.
+ */
+ ATF_REQUIRE(aio_error(&ioreq[i].iocb) == EINPROGRESS);
+ ATF_REQUIRE(aio_cancel(s[0], &ioreq[i].iocb) == AIO_CANCELED);
+
+ close(s[1]);
+ close(s[0]);
+}
+
+static void
+aio_socket_blocking_short_write_test(bool vectored)
+{
+ struct aiocb iocb, *iocbp;
+ struct iovec iov[2];
+ char *buffer[2];
+ ssize_t done, r;
+ int buffer_size, sb_size;
+ socklen_t len;
+ int s[2];
+
+ ATF_REQUIRE(socketpair(PF_UNIX, SOCK_STREAM, 0, s) != -1);
+
+ len = sizeof(sb_size);
+ ATF_REQUIRE(getsockopt(s[0], SOL_SOCKET, SO_RCVBUF, &sb_size, &len) !=
+ -1);
+ ATF_REQUIRE(len == sizeof(sb_size));
+ buffer_size = sb_size;
+
+ ATF_REQUIRE(getsockopt(s[1], SOL_SOCKET, SO_SNDBUF, &sb_size, &len) !=
+ -1);
+ ATF_REQUIRE(len == sizeof(sb_size));
+ if (sb_size > buffer_size)
+ buffer_size = sb_size;
+
+ /*
+ * Use twice the size of the MAX(receive buffer, send buffer)
+ * to ensure that the write is split up into multiple writes
+ * internally.
+ */
+ buffer_size *= 2;
+
+ buffer[0] = malloc(buffer_size);
+ ATF_REQUIRE(buffer[0] != NULL);
+ buffer[1] = malloc(buffer_size);
+ ATF_REQUIRE(buffer[1] != NULL);
+
+ srandomdev();
+ aio_fill_buffer(buffer[1], buffer_size, random());
+
+ memset(&iocb, 0, sizeof(iocb));
+ iocb.aio_fildes = s[1];
+ if (vectored) {
+ iov[0].iov_base = buffer[1];
+ iov[0].iov_len = buffer_size / 2 + 1;
+ iov[1].iov_base = buffer[1] + buffer_size / 2 + 1;
+ iov[1].iov_len = buffer_size / 2 - 1;
+ iocb.aio_iov = iov;
+ iocb.aio_iovcnt = 2;
+ r = aio_writev(&iocb);
+ ATF_CHECK_EQ_MSG(0, r, "aio_writev returned %zd", r);
+ } else {
+ iocb.aio_buf = buffer[1];
+ iocb.aio_nbytes = buffer_size;
+ r = aio_write(&iocb);
+ ATF_CHECK_EQ_MSG(0, r, "aio_writev returned %zd", r);
+ }
+
+ done = recv(s[0], buffer[0], buffer_size, MSG_WAITALL);
+ ATF_REQUIRE(done == buffer_size);
+
+ done = aio_waitcomplete(&iocbp, NULL);
+ ATF_REQUIRE(iocbp == &iocb);
+ ATF_REQUIRE(done == buffer_size);
+
+ ATF_REQUIRE(memcmp(buffer[0], buffer[1], buffer_size) == 0);
+
+ close(s[1]);
+ close(s[0]);
+}
+
+/*
+ * This test ensures that aio_write() on a blocking socket of a "large"
+ * buffer does not return a short completion.
+ */
+ATF_TC_WITHOUT_HEAD(aio_socket_blocking_short_write);
+ATF_TC_BODY(aio_socket_blocking_short_write, tc)
+{
+ aio_socket_blocking_short_write_test(false);
+}
+
+/*
+ * Like aio_socket_blocking_short_write, but also tests that partially
+ * completed vectored sends can be retried correctly.
+ */
+ATF_TC_WITHOUT_HEAD(aio_socket_blocking_short_write_vectored);
+ATF_TC_BODY(aio_socket_blocking_short_write_vectored, tc)
+{
+ aio_socket_blocking_short_write_test(true);
+}
+
+/*
+ * Verify that AIO requests fail when applied to a listening socket.
+ */
+ATF_TC_WITHOUT_HEAD(aio_socket_listen_fail);
+ATF_TC_BODY(aio_socket_listen_fail, tc)
+{
+ struct aiocb iocb;
+ struct sockaddr_un sun;
+ char buf[16];
+ int s;
+
+ s = socket(AF_LOCAL, SOCK_STREAM, 0);
+ ATF_REQUIRE(s != -1);
+
+ memset(&sun, 0, sizeof(sun));
+ snprintf(sun.sun_path, sizeof(sun.sun_path), "%s", "listen.XXXXXX");
+ mktemp(sun.sun_path);
+ sun.sun_family = AF_LOCAL;
+ sun.sun_len = SUN_LEN(&sun);
+
+ ATF_REQUIRE(bind(s, (struct sockaddr *)&sun, SUN_LEN(&sun)) == 0);
+ ATF_REQUIRE(listen(s, 5) == 0);
+
+ memset(buf, 0, sizeof(buf));
+ memset(&iocb, 0, sizeof(iocb));
+ iocb.aio_fildes = s;
+ iocb.aio_buf = buf;
+ iocb.aio_nbytes = sizeof(buf);
+
+ ATF_REQUIRE_ERRNO(EINVAL, aio_read(&iocb) == -1);
+ ATF_REQUIRE_ERRNO(EINVAL, aio_write(&iocb) == -1);
+
+ ATF_REQUIRE(unlink(sun.sun_path) == 0);
+ close(s);
+}
+
+/*
+ * Verify that listen(2) fails if a socket has pending AIO requests.
+ */
+ATF_TC_WITHOUT_HEAD(aio_socket_listen_pending);
+ATF_TC_BODY(aio_socket_listen_pending, tc)
+{
+ struct aiocb iocb;
+ struct sockaddr_un sun;
+ char buf[16];
+ int s;
+
+ s = socket(AF_LOCAL, SOCK_STREAM, 0);
+ ATF_REQUIRE(s != -1);
+
+ memset(&sun, 0, sizeof(sun));
+ snprintf(sun.sun_path, sizeof(sun.sun_path), "%s", "listen.XXXXXX");
+ mktemp(sun.sun_path);
+ sun.sun_family = AF_LOCAL;
+ sun.sun_len = SUN_LEN(&sun);
+
+ ATF_REQUIRE(bind(s, (struct sockaddr *)&sun, SUN_LEN(&sun)) == 0);
+
+ memset(buf, 0, sizeof(buf));
+ memset(&iocb, 0, sizeof(iocb));
+ iocb.aio_fildes = s;
+ iocb.aio_buf = buf;
+ iocb.aio_nbytes = sizeof(buf);
+ ATF_REQUIRE(aio_read(&iocb) == 0);
+
+ ATF_REQUIRE_ERRNO(EINVAL, listen(s, 5) == -1);
+
+ ATF_REQUIRE(aio_cancel(s, &iocb) != -1);
+
+ ATF_REQUIRE(unlink(sun.sun_path) == 0);
+ close(s);
+}
+
+/*
+ * This test verifies that cancelling a partially completed socket write
+ * returns a short write rather than ECANCELED.
+ */
+ATF_TC_WITHOUT_HEAD(aio_socket_short_write_cancel);
+ATF_TC_BODY(aio_socket_short_write_cancel, tc)
+{
+ struct aiocb iocb, *iocbp;
+ char *buffer[2];
+ ssize_t done;
+ int buffer_size, sb_size;
+ socklen_t len;
+ int s[2];
+
+ ATF_REQUIRE(socketpair(PF_UNIX, SOCK_STREAM, 0, s) != -1);
+
+ len = sizeof(sb_size);
+ ATF_REQUIRE(getsockopt(s[0], SOL_SOCKET, SO_RCVBUF, &sb_size, &len) !=
+ -1);
+ ATF_REQUIRE(len == sizeof(sb_size));
+ buffer_size = sb_size;
+
+ ATF_REQUIRE(getsockopt(s[1], SOL_SOCKET, SO_SNDBUF, &sb_size, &len) !=
+ -1);
+ ATF_REQUIRE(len == sizeof(sb_size));
+ if (sb_size > buffer_size)
+ buffer_size = sb_size;
+
+ /*
+ * Use three times the size of the MAX(receive buffer, send
+ * buffer) for the write to ensure that the write is split up
+ * into multiple writes internally. The recv() ensures that
+ * the write has partially completed, but a remaining size of
+ * two buffers should ensure that the write has not completed
+ * fully when it is cancelled.
+ */
+ buffer[0] = malloc(buffer_size);
+ ATF_REQUIRE(buffer[0] != NULL);
+ buffer[1] = malloc(buffer_size * 3);
+ ATF_REQUIRE(buffer[1] != NULL);
+
+ srandomdev();
+ aio_fill_buffer(buffer[1], buffer_size * 3, random());
+
+ memset(&iocb, 0, sizeof(iocb));
+ iocb.aio_fildes = s[1];
+ iocb.aio_buf = buffer[1];
+ iocb.aio_nbytes = buffer_size * 3;
+ ATF_REQUIRE(aio_write(&iocb) == 0);
+
+ done = recv(s[0], buffer[0], buffer_size, MSG_WAITALL);
+ ATF_REQUIRE(done == buffer_size);
+
+ ATF_REQUIRE(aio_error(&iocb) == EINPROGRESS);
+ ATF_REQUIRE(aio_cancel(s[1], &iocb) == AIO_NOTCANCELED);
+
+ done = aio_waitcomplete(&iocbp, NULL);
+ ATF_REQUIRE(iocbp == &iocb);
+ ATF_REQUIRE(done >= buffer_size && done <= buffer_size * 2);
+
+ ATF_REQUIRE(memcmp(buffer[0], buffer[1], buffer_size) == 0);
+
+ close(s[1]);
+ close(s[0]);
+}
+
+/*
+ * Test handling of aio_read() and aio_write() on shut-down sockets.
+ */
+ATF_TC_WITHOUT_HEAD(aio_socket_shutdown);
+ATF_TC_BODY(aio_socket_shutdown, tc)
+{
+ struct aiocb iocb;
+ sigset_t set;
+ char *buffer;
+ ssize_t len;
+ size_t bsz;
+ int error, s[2];
+
+ ATF_REQUIRE(socketpair(PF_UNIX, SOCK_STREAM, 0, s) != -1);
+
+ bsz = 1024;
+ buffer = malloc(bsz);
+ memset(buffer, 0, bsz);
+
+ /* Put some data in s[0]'s recv buffer. */
+ ATF_REQUIRE(send(s[1], buffer, bsz, 0) == (ssize_t)bsz);
+
+ /* No more reading from s[0]. */
+ ATF_REQUIRE(shutdown(s[0], SHUT_RD) != -1);
+
+ ATF_REQUIRE(buffer != NULL);
+
+ memset(&iocb, 0, sizeof(iocb));
+ iocb.aio_fildes = s[0];
+ iocb.aio_buf = buffer;
+ iocb.aio_nbytes = bsz;
+ ATF_REQUIRE(aio_read(&iocb) == 0);
+
+ /* Expect to see zero bytes, analogous to recv(2). */
+ while ((error = aio_error(&iocb)) == EINPROGRESS)
+ usleep(25000);
+ ATF_REQUIRE_MSG(error == 0, "aio_error() returned %d", error);
+ len = aio_return(&iocb);
+ ATF_REQUIRE_MSG(len == 0, "read job returned %zd bytes", len);
+
+ /* No more writing to s[1]. */
+ ATF_REQUIRE(shutdown(s[1], SHUT_WR) != -1);
+
+ /* Block SIGPIPE so that we can detect the error in-band. */
+ sigemptyset(&set);
+ sigaddset(&set, SIGPIPE);
+ ATF_REQUIRE(sigprocmask(SIG_BLOCK, &set, NULL) == 0);
+
+ memset(&iocb, 0, sizeof(iocb));
+ iocb.aio_fildes = s[1];
+ iocb.aio_buf = buffer;
+ iocb.aio_nbytes = bsz;
+ ATF_REQUIRE(aio_write(&iocb) == 0);
+
+ /* Expect an error, analogous to send(2). */
+ while ((error = aio_error(&iocb)) == EINPROGRESS)
+ usleep(25000);
+ ATF_REQUIRE_MSG(error == EPIPE, "aio_error() returned %d", error);
+
+ ATF_REQUIRE(close(s[0]) != -1);
+ ATF_REQUIRE(close(s[1]) != -1);
+ free(buffer);
+}
+
+/*
+ * test aio_fsync's behavior with bad inputs
+ */
+ATF_TC_WITHOUT_HEAD(aio_fsync_errors);
+ATF_TC_BODY(aio_fsync_errors, tc)
+{
+ int fd;
+ struct aiocb iocb;
+
+ ATF_REQUIRE_UNSAFE_AIO();
+
+ fd = open(FILE_PATHNAME, O_RDWR | O_CREAT, 0600);
+ ATF_REQUIRE_MSG(fd != -1, "open failed: %s", strerror(errno));
+ unlink(FILE_PATHNAME);
+
+ /* aio_fsync should return EINVAL unless op is O_SYNC or O_DSYNC */
+ memset(&iocb, 0, sizeof(iocb));
+ iocb.aio_fildes = fd;
+ ATF_CHECK_EQ(-1, aio_fsync(666, &iocb));
+ ATF_CHECK_EQ(EINVAL, errno);
+
+ /* aio_fsync should return EBADF if fd is not a valid descriptor */
+ memset(&iocb, 0, sizeof(iocb));
+ iocb.aio_fildes = 666;
+ ATF_CHECK_EQ(-1, aio_fsync(O_SYNC, &iocb));
+ ATF_CHECK_EQ(EBADF, errno);
+
+ /* aio_fsync should return EINVAL if sigev_notify is invalid */
+ memset(&iocb, 0, sizeof(iocb));
+ iocb.aio_fildes = fd;
+ iocb.aio_sigevent.sigev_notify = 666;
+ ATF_CHECK_EQ(-1, aio_fsync(666, &iocb));
+ ATF_CHECK_EQ(EINVAL, errno);
+}
+
+/*
+ * This test just performs a basic test of aio_fsync().
+ */
+static void
+aio_fsync_test(int op)
+{
+ struct aiocb synccb, *iocbp;
+ struct {
+ struct aiocb iocb;
+ bool done;
+ char *buffer;
+ } buffers[16];
+ struct stat sb;
+ ssize_t rval;
+ unsigned i;
+ int fd;
+
+ ATF_REQUIRE_UNSAFE_AIO();
+
+ fd = open(FILE_PATHNAME, O_RDWR | O_CREAT, 0600);
+ ATF_REQUIRE_MSG(fd != -1, "open failed: %s", strerror(errno));
+ unlink(FILE_PATHNAME);
+
+ ATF_REQUIRE(fstat(fd, &sb) == 0);
+ ATF_REQUIRE(sb.st_blksize != 0);
+ ATF_REQUIRE(ftruncate(fd, sb.st_blksize * nitems(buffers)) == 0);
+
+ /*
+ * Queue several asynchronous write requests. Hopefully this
+ * forces the aio_fsync() request to be deferred. There is no
+ * reliable way to guarantee that however.
+ */
+ srandomdev();
+ for (i = 0; i < nitems(buffers); i++) {
+ buffers[i].done = false;
+ memset(&buffers[i].iocb, 0, sizeof(buffers[i].iocb));
+ buffers[i].buffer = malloc(sb.st_blksize);
+ aio_fill_buffer(buffers[i].buffer, sb.st_blksize, random());
+ buffers[i].iocb.aio_fildes = fd;
+ buffers[i].iocb.aio_buf = buffers[i].buffer;
+ buffers[i].iocb.aio_nbytes = sb.st_blksize;
+ buffers[i].iocb.aio_offset = sb.st_blksize * i;
+ ATF_REQUIRE(aio_write(&buffers[i].iocb) == 0);
+ }
+
+ /* Queue the aio_fsync request. */
+ memset(&synccb, 0, sizeof(synccb));
+ synccb.aio_fildes = fd;
+ ATF_REQUIRE(aio_fsync(op, &synccb) == 0);
+
+ /* Wait for requests to complete. */
+ for (;;) {
+ next:
+ rval = aio_waitcomplete(&iocbp, NULL);
+ ATF_REQUIRE(iocbp != NULL);
+ if (iocbp == &synccb) {
+ ATF_REQUIRE(rval == 0);
+ break;
+ }
+
+ for (i = 0; i < nitems(buffers); i++) {
+ if (iocbp == &buffers[i].iocb) {
+ ATF_REQUIRE(buffers[i].done == false);
+ ATF_REQUIRE(rval == sb.st_blksize);
+ buffers[i].done = true;
+ goto next;
+ }
+ }
+
+ ATF_REQUIRE_MSG(false, "unmatched AIO request");
+ }
+
+ for (i = 0; i < nitems(buffers); i++)
+ ATF_REQUIRE_MSG(buffers[i].done,
+ "AIO request %u did not complete", i);
+
+ close(fd);
+}
+
+ATF_TC_WITHOUT_HEAD(aio_fsync_sync_test);
+ATF_TC_BODY(aio_fsync_sync_test, tc)
+{
+ aio_fsync_test(O_SYNC);
+}
+
+ATF_TC_WITHOUT_HEAD(aio_fsync_dsync_test);
+ATF_TC_BODY(aio_fsync_dsync_test, tc)
+{
+ aio_fsync_test(O_DSYNC);
+}
+
+/*
+ * We shouldn't be able to DoS the system by setting iov_len to an insane
+ * value
+ */
+ATF_TC_WITHOUT_HEAD(aio_writev_dos_iov_len);
+ATF_TC_BODY(aio_writev_dos_iov_len, tc)
+{
+ struct aiocb aio;
+ const struct aiocb *const iocbs[] = {&aio};
+ const char *wbuf = "Hello, world!";
+ struct iovec iov[1];
+ ssize_t r;
+ int fd;
+
+ ATF_REQUIRE_UNSAFE_AIO();
+
+ fd = open("testfile", O_RDWR | O_CREAT, 0600);
+ ATF_REQUIRE_MSG(fd != -1, "open failed: %s", strerror(errno));
+
+ iov[0].iov_base = __DECONST(void*, wbuf);
+ iov[0].iov_len = 1 << 30;
+ bzero(&aio, sizeof(aio));
+ aio.aio_fildes = fd;
+ aio.aio_offset = 0;
+ aio.aio_iov = iov;
+ aio.aio_iovcnt = 1;
+
+ r = aio_writev(&aio);
+ ATF_CHECK_EQ_MSG(0, r, "aio_writev returned %zd", r);
+ ATF_REQUIRE_EQ(0, aio_suspend(iocbs, 1, NULL));
+ r = aio_return(&aio);
+ ATF_CHECK_EQ_MSG(-1, r, "aio_return returned %zd", r);
+ ATF_CHECK_MSG(errno == EFAULT || errno == EINVAL,
+ "aio_writev: %s", strerror(errno));
+
+ close(fd);
+}
+
+/*
+ * We shouldn't be able to DoS the system by setting aio_iovcnt to an insane
+ * value
+ */
+ATF_TC_WITHOUT_HEAD(aio_writev_dos_iovcnt);
+ATF_TC_BODY(aio_writev_dos_iovcnt, tc)
+{
+ struct aiocb aio;
+ const char *wbuf = "Hello, world!";
+ struct iovec iov[1];
+ ssize_t len;
+ int fd;
+
+ ATF_REQUIRE_UNSAFE_AIO();
+
+ fd = open("testfile", O_RDWR | O_CREAT, 0600);
+ ATF_REQUIRE_MSG(fd != -1, "open failed: %s", strerror(errno));
+
+ len = strlen(wbuf);
+ iov[0].iov_base = __DECONST(void*, wbuf);
+ iov[0].iov_len = len;
+ bzero(&aio, sizeof(aio));
+ aio.aio_fildes = fd;
+ aio.aio_offset = 0;
+ aio.aio_iov = iov;
+ aio.aio_iovcnt = 1 << 30;
+
+ ATF_REQUIRE_EQ(-1, aio_writev(&aio));
+ ATF_CHECK_EQ(EINVAL, errno);
+
+ close(fd);
+}
+
+ATF_TC_WITH_CLEANUP(aio_writev_efault);
+ATF_TC_HEAD(aio_writev_efault, tc)
+{
+ atf_tc_set_md_var(tc, "descr",
+ "Vectored AIO should gracefully handle invalid addresses");
+ atf_tc_set_md_var(tc, "require.user", "root");
+}
+ATF_TC_BODY(aio_writev_efault, tc)
+{
+ struct aiocb aio;
+ ssize_t buflen;
+ char *buffer;
+ struct iovec iov[2];
+ long seed;
+ int fd;
+
+ ATF_REQUIRE_UNSAFE_AIO();
+
+ fd = aio_md_setup();
+
+ seed = random();
+ buflen = 4096;
+ buffer = malloc(buflen);
+ aio_fill_buffer(buffer, buflen, seed);
+ iov[0].iov_base = buffer;
+ iov[0].iov_len = buflen;
+ iov[1].iov_base = (void*)-1; /* Invalid! */
+ iov[1].iov_len = buflen;
+ bzero(&aio, sizeof(aio));
+ aio.aio_fildes = fd;
+ aio.aio_offset = 0;
+ aio.aio_iov = iov;
+ aio.aio_iovcnt = nitems(iov);
+
+ ATF_REQUIRE_EQ(-1, aio_writev(&aio));
+ ATF_CHECK_EQ(EFAULT, errno);
+
+ close(fd);
+}
+ATF_TC_CLEANUP(aio_writev_efault, tc)
+{
+ aio_md_cleanup();
+}
+
+ATF_TC_WITHOUT_HEAD(aio_writev_empty_file_poll);
+ATF_TC_BODY(aio_writev_empty_file_poll, tc)
+{
+ struct aiocb aio;
+ int fd;
+
+ ATF_REQUIRE_UNSAFE_AIO();
+
+ fd = open("testfile", O_RDWR | O_CREAT, 0600);
+ ATF_REQUIRE_MSG(fd != -1, "open failed: %s", strerror(errno));
+
+ bzero(&aio, sizeof(aio));
+ aio.aio_fildes = fd;
+ aio.aio_offset = 0;
+ aio.aio_iovcnt = 0;
+
+ ATF_REQUIRE_EQ(0, aio_writev(&aio));
+ ATF_REQUIRE_EQ(0, suspend(&aio));
+
+ close(fd);
+}
+
+ATF_TC_WITHOUT_HEAD(aio_writev_empty_file_signal);
+ATF_TC_BODY(aio_writev_empty_file_signal, tc)
+{
+ struct aiocb aio;
+ int fd;
+
+ ATF_REQUIRE_UNSAFE_AIO();
+
+ fd = open("testfile", O_RDWR | O_CREAT, 0600);
+ ATF_REQUIRE_MSG(fd != -1, "open failed: %s", strerror(errno));
+
+ bzero(&aio, sizeof(aio));
+ aio.aio_fildes = fd;
+ aio.aio_offset = 0;
+ aio.aio_iovcnt = 0;
+ aio.aio_sigevent = *setup_signal();
+
+ ATF_REQUIRE_EQ(0, aio_writev(&aio));
+ ATF_REQUIRE_EQ(0, poll_signaled(&aio));
+
+ close(fd);
+}
+
+/*
+ * Use an aiocb with kqueue and EV_ONESHOT. kqueue should deliver the event
+ * only once, even if the user doesn't promptly call aio_return.
+ */
+ATF_TC_WITHOUT_HEAD(ev_oneshot);
+ATF_TC_BODY(ev_oneshot, tc)
+{
+ int fd, kq, nevents;
+ struct aiocb iocb;
+ struct kevent events[1];
+ struct timespec timeout;
+
+ kq = kqueue();
+ ATF_REQUIRE(kq >= 0);
+
+ fd = open(FILE_PATHNAME, O_RDWR | O_CREAT, 0600);
+ ATF_REQUIRE_MSG(fd != -1, "open failed: %s", strerror(errno));
+
+ memset(&iocb, 0, sizeof(iocb));
+ iocb.aio_fildes = fd;
+ iocb.aio_sigevent.sigev_notify_kqueue = kq;
+ iocb.aio_sigevent.sigev_value.sival_ptr = (void*)0xdeadbeef;
+ iocb.aio_sigevent.sigev_notify_kevent_flags = EV_ONESHOT;
+ iocb.aio_sigevent.sigev_notify = SIGEV_KEVENT;
+
+ ATF_CHECK_EQ(0, aio_fsync(O_SYNC, &iocb));
+
+ nevents = kevent(kq, NULL, 0, events, 1, NULL);
+ ATF_CHECK_EQ(1, nevents);
+ ATF_CHECK_EQ(events[0].ident, (uintptr_t) &iocb);
+ ATF_CHECK_EQ(events[0].filter, EVFILT_AIO);
+ ATF_CHECK_EQ(events[0].flags, EV_EOF | EV_ONESHOT);
+ ATF_CHECK_EQ(events[0].fflags, 0);
+ ATF_CHECK_EQ(events[0].data, 0);
+ ATF_CHECK_EQ((uintptr_t)events[0].udata, 0xdeadbeef);
+
+ /*
+ * Even though we haven't called aio_return, kevent will not return the
+ * event again due to EV_ONESHOT.
+ */
+ timeout.tv_sec = 0;
+ timeout.tv_nsec = 100000000;
+ nevents = kevent(kq, NULL, 0, events, 1, &timeout);
+ ATF_CHECK_EQ(0, nevents);
+
+ ATF_CHECK_EQ(0, aio_return(&iocb));
+ close(fd);
+ close(kq);
+}
+
+
+// aio_writev and aio_readv should still work even if the iovcnt is greater
+// than the number of buffered AIO operations permitted per process.
+ATF_TC_WITH_CLEANUP(vectored_big_iovcnt);
+ATF_TC_HEAD(vectored_big_iovcnt, tc)
+{
+ atf_tc_set_md_var(tc, "descr",
+ "Vectored AIO should still work even if the iovcnt is greater than "
+ "the number of buffered AIO operations permitted by the process");
+ atf_tc_set_md_var(tc, "require.user", "root");
+}
+ATF_TC_BODY(vectored_big_iovcnt, tc)
+{
+ struct aiocb aio;
+ struct iovec *iov;
+ ssize_t len, buflen;
+ char *buffer;
+ const char *oid = "vfs.aio.max_buf_aio";
+ long seed;
+ int max_buf_aio;
+ int fd, i;
+ ssize_t sysctl_len = sizeof(max_buf_aio);
+
+ ATF_REQUIRE_UNSAFE_AIO();
+
+ if (sysctlbyname(oid, &max_buf_aio, &sysctl_len, NULL, 0) == -1)
+ atf_libc_error(errno, "Failed to read %s", oid);
+
+ seed = random();
+ buflen = 512 * (max_buf_aio + 1);
+ buffer = malloc(buflen);
+ aio_fill_buffer(buffer, buflen, seed);
+ iov = calloc(max_buf_aio + 1, sizeof(struct iovec));
+
+ fd = aio_md_setup();
+
+ bzero(&aio, sizeof(aio));
+ aio.aio_fildes = fd;
+ aio.aio_offset = 0;
+ for (i = 0; i < max_buf_aio + 1; i++) {
+ iov[i].iov_base = &buffer[i * 512];
+ iov[i].iov_len = 512;
+ }
+ aio.aio_iov = iov;
+ aio.aio_iovcnt = max_buf_aio + 1;
+
+ if (aio_writev(&aio) < 0)
+ atf_tc_fail("aio_writev failed: %s", strerror(errno));
+
+ len = poll(&aio);
+ if (len < 0)
+ atf_tc_fail("aio failed: %s", strerror(errno));
+
+ if (len != buflen)
+ atf_tc_fail("aio short write: got %jd, expected: %jd "
+ "(max_buf_aio=%d, iovcnt=%zu)",
+ (intmax_t)len, (intmax_t)buflen, max_buf_aio, aio.aio_iovcnt);
+
+ bzero(&aio, sizeof(aio));
+ aio.aio_fildes = fd;
+ aio.aio_offset = 0;
+ aio.aio_iov = iov;
+ aio.aio_iovcnt = max_buf_aio + 1;
+
+ if (aio_readv(&aio) < 0)
+ atf_tc_fail("aio_readv failed: %s", strerror(errno));
+
+ len = poll(&aio);
+ if (len < 0)
+ atf_tc_fail("aio failed: %s", strerror(errno));
+
+ if (len != buflen)
+ atf_tc_fail("aio short read (%jd)", (intmax_t)len);
+
+ if (aio_test_buffer(buffer, buflen, seed) == 0)
+ atf_tc_fail("buffer mismatched");
+
+ close(fd);
+}
+ATF_TC_CLEANUP(vectored_big_iovcnt, tc)
+{
+ aio_md_cleanup();
+}
+
+ATF_TC_WITHOUT_HEAD(vectored_file_poll);
+ATF_TC_BODY(vectored_file_poll, tc)
+{
+ aio_file_test(poll, NULL, true);
+}
+
+ATF_TC_WITHOUT_HEAD(vectored_thread);
+ATF_TC_BODY(vectored_thread, tc)
+{
+ aio_file_test(poll_signaled, setup_thread(), true);
+}
+
+ATF_TC_WITH_CLEANUP(vectored_md_poll);
+ATF_TC_HEAD(vectored_md_poll, tc)
+{
+ atf_tc_set_md_var(tc, "require.user", "root");
+}
+ATF_TC_BODY(vectored_md_poll, tc)
+{
+ aio_md_test(poll, NULL, true);
+}
+ATF_TC_CLEANUP(vectored_md_poll, tc)
+{
+ aio_md_cleanup();
+}
+
+ATF_TC_WITHOUT_HEAD(vectored_socket_poll);
+ATF_TC_BODY(vectored_socket_poll, tc)
+{
+ aio_unix_socketpair_test(poll, NULL, true);
+}
+
+// aio_writev and aio_readv should still work even if the iov contains elements
+// that aren't a multiple of the device's sector size, and even if the total
+// amount if I/O _is_ a multiple of the device's sector size.
+ATF_TC_WITH_CLEANUP(vectored_unaligned);
+ATF_TC_HEAD(vectored_unaligned, tc)
+{
+ atf_tc_set_md_var(tc, "descr",
+ "Vectored AIO should still work even if the iov contains elements "
+ "that aren't a multiple of the sector size.");
+ atf_tc_set_md_var(tc, "require.user", "root");
+ atf_tc_set_md_var(tc, "require.kmods", "zfs");
+}
+ATF_TC_BODY(vectored_unaligned, tc)
+{
+ struct aio_context ac;
+ struct aiocb aio;
+ struct iovec iov[3];
+ ssize_t len, total_len;
+ int fd;
+
+ if (atf_tc_get_config_var_as_bool_wd(tc, "ci", false))
+ atf_tc_skip("https://bugs.freebsd.org/258766");
+
+ ATF_REQUIRE_UNSAFE_AIO();
+
+ /*
+ * Use a zvol with volmode=dev, so it will allow .d_write with
+ * unaligned uio. geom devices use physio, which doesn't allow that.
+ */
+ fd = aio_zvol_setup(atf_tc_get_ident(tc));
+ aio_context_init(&ac, fd, fd, FILE_LEN);
+
+ /* Break the buffer into 3 parts:
+ * * A 4kB part, aligned to 4kB
+ * * Two other parts that add up to 4kB:
+ * - 256B
+ * - 4kB - 256B
+ */
+ iov[0].iov_base = ac.ac_buffer;
+ iov[0].iov_len = 4096;
+ iov[1].iov_base = (void*)((uintptr_t)iov[0].iov_base + iov[0].iov_len);
+ iov[1].iov_len = 256;
+ iov[2].iov_base = (void*)((uintptr_t)iov[1].iov_base + iov[1].iov_len);
+ iov[2].iov_len = 4096 - iov[1].iov_len;
+ total_len = iov[0].iov_len + iov[1].iov_len + iov[2].iov_len;
+ bzero(&aio, sizeof(aio));
+ aio.aio_fildes = ac.ac_write_fd;
+ aio.aio_offset = 0;
+ aio.aio_iov = iov;
+ aio.aio_iovcnt = 3;
+
+ if (aio_writev(&aio) < 0)
+ atf_tc_fail("aio_writev failed: %s", strerror(errno));
+
+ len = poll(&aio);
+ if (len < 0)
+ atf_tc_fail("aio failed: %s", strerror(errno));
+
+ if (len != total_len)
+ atf_tc_fail("aio short write (%jd)", (intmax_t)len);
+
+ bzero(&aio, sizeof(aio));
+ aio.aio_fildes = ac.ac_read_fd;
+ aio.aio_offset = 0;
+ aio.aio_iov = iov;
+ aio.aio_iovcnt = 3;
+
+ if (aio_readv(&aio) < 0)
+ atf_tc_fail("aio_readv failed: %s", strerror(errno));
+ len = poll(&aio);
+
+ ATF_REQUIRE_MSG(aio_test_buffer(ac.ac_buffer, total_len,
+ ac.ac_seed) != 0, "aio_test_buffer: internal error");
+
+ close(fd);
+}
+ATF_TC_CLEANUP(vectored_unaligned, tc)
+{
+ aio_zvol_cleanup(atf_tc_get_ident(tc));
+}
+
+static void
+aio_zvol_test(completion comp, struct sigevent *sev, bool vectored,
+ const char *unique)
+{
+ struct aio_context ac;
+ int fd;
+
+ fd = aio_zvol_setup(unique);
+ aio_context_init(&ac, fd, fd, DEVICE_IO_LEN);
+ if (vectored) {
+ aio_writev_test(&ac, comp, sev);
+ aio_readv_test(&ac, comp, sev);
+ } else {
+ aio_write_test(&ac, comp, sev);
+ aio_read_test(&ac, comp, sev);
+ }
+
+ close(fd);
+}
+
+/*
+ * Note that unlike md, the zvol is not a geom device, does not allow unmapped
+ * buffers, and does not use physio.
+ */
+ATF_TC_WITH_CLEANUP(vectored_zvol_poll);
+ATF_TC_HEAD(vectored_zvol_poll, tc)
+{
+ atf_tc_set_md_var(tc, "require.user", "root");
+ atf_tc_set_md_var(tc, "require.kmods", "zfs");
+}
+ATF_TC_BODY(vectored_zvol_poll, tc)
+{
+ if (atf_tc_get_config_var_as_bool_wd(tc, "ci", false))
+ atf_tc_skip("https://bugs.freebsd.org/258766");
+ aio_zvol_test(poll, NULL, true, atf_tc_get_ident(tc));
+}
+ATF_TC_CLEANUP(vectored_zvol_poll, tc)
+{
+ aio_zvol_cleanup(atf_tc_get_ident(tc));
+}
+
+ATF_TP_ADD_TCS(tp)
+{
+
+ /* Test every file type with every completion method */
+ ATF_TP_ADD_TC(tp, file_kq);
+ ATF_TP_ADD_TC(tp, file_poll);
+ ATF_TP_ADD_TC(tp, file_signal);
+ ATF_TP_ADD_TC(tp, file_suspend);
+ ATF_TP_ADD_TC(tp, file_thread);
+ ATF_TP_ADD_TC(tp, file_waitcomplete);
+ ATF_TP_ADD_TC(tp, fifo_kq);
+ ATF_TP_ADD_TC(tp, fifo_poll);
+ ATF_TP_ADD_TC(tp, fifo_signal);
+ ATF_TP_ADD_TC(tp, fifo_suspend);
+ ATF_TP_ADD_TC(tp, fifo_thread);
+ ATF_TP_ADD_TC(tp, fifo_waitcomplete);
+ ATF_TP_ADD_TC(tp, socket_kq);
+ ATF_TP_ADD_TC(tp, socket_poll);
+ ATF_TP_ADD_TC(tp, socket_signal);
+ ATF_TP_ADD_TC(tp, socket_suspend);
+ ATF_TP_ADD_TC(tp, socket_thread);
+ ATF_TP_ADD_TC(tp, socket_waitcomplete);
+ ATF_TP_ADD_TC(tp, pty_kq);
+ ATF_TP_ADD_TC(tp, pty_poll);
+ ATF_TP_ADD_TC(tp, pty_signal);
+ ATF_TP_ADD_TC(tp, pty_suspend);
+ ATF_TP_ADD_TC(tp, pty_thread);
+ ATF_TP_ADD_TC(tp, pty_waitcomplete);
+ ATF_TP_ADD_TC(tp, pipe_kq);
+ ATF_TP_ADD_TC(tp, pipe_poll);
+ ATF_TP_ADD_TC(tp, pipe_signal);
+ ATF_TP_ADD_TC(tp, pipe_suspend);
+ ATF_TP_ADD_TC(tp, pipe_thread);
+ ATF_TP_ADD_TC(tp, pipe_waitcomplete);
+ ATF_TP_ADD_TC(tp, md_kq);
+ ATF_TP_ADD_TC(tp, md_poll);
+ ATF_TP_ADD_TC(tp, md_signal);
+ ATF_TP_ADD_TC(tp, md_suspend);
+ ATF_TP_ADD_TC(tp, md_thread);
+ ATF_TP_ADD_TC(tp, md_waitcomplete);
+
+ /* Various special cases */
+ ATF_TP_ADD_TC(tp, aio_fsync_errors);
+ ATF_TP_ADD_TC(tp, aio_fsync_sync_test);
+ ATF_TP_ADD_TC(tp, aio_fsync_dsync_test);
+ ATF_TP_ADD_TC(tp, aio_large_read_test);
+ ATF_TP_ADD_TC(tp, aio_socket_two_reads);
+ ATF_TP_ADD_TC(tp, aio_socket_blocking_short_write);
+ ATF_TP_ADD_TC(tp, aio_socket_blocking_short_write_vectored);
+ ATF_TP_ADD_TC(tp, aio_socket_listen_fail);
+ ATF_TP_ADD_TC(tp, aio_socket_listen_pending);
+ ATF_TP_ADD_TC(tp, aio_socket_short_write_cancel);
+ ATF_TP_ADD_TC(tp, aio_socket_shutdown);
+ ATF_TP_ADD_TC(tp, aio_writev_dos_iov_len);
+ ATF_TP_ADD_TC(tp, aio_writev_dos_iovcnt);
+ ATF_TP_ADD_TC(tp, aio_writev_efault);
+ ATF_TP_ADD_TC(tp, aio_writev_empty_file_poll);
+ ATF_TP_ADD_TC(tp, aio_writev_empty_file_signal);
+ ATF_TP_ADD_TC(tp, ev_oneshot);
+ ATF_TP_ADD_TC(tp, vectored_big_iovcnt);
+ ATF_TP_ADD_TC(tp, vectored_file_poll);
+ ATF_TP_ADD_TC(tp, vectored_md_poll);
+ ATF_TP_ADD_TC(tp, vectored_zvol_poll);
+ ATF_TP_ADD_TC(tp, vectored_unaligned);
+ ATF_TP_ADD_TC(tp, vectored_socket_poll);
+ ATF_TP_ADD_TC(tp, vectored_thread);
+
+ return (atf_no_error());
+}