#include "liburing.h"
extern inline int
rust_io_uring_opcode_supported(struct io_uring_probe* p, int op)
{
return io_uring_opcode_supported(p, op);
}
extern inline void
rust_io_uring_cq_advance(struct io_uring* ring, unsigned nr)
{
io_uring_cq_advance(ring, nr);
}
extern inline void
rust_io_uring_cqe_seen(struct io_uring* ring, struct io_uring_cqe* cqe)
{
io_uring_cqe_seen(ring, cqe);
}
extern inline void
rust_io_uring_sqe_set_data(struct io_uring_sqe* sqe, void* data)
{
io_uring_sqe_set_data(sqe, data);
}
extern inline void*
rust_io_uring_cqe_get_data(struct io_uring_cqe* cqe)
{
return io_uring_cqe_get_data(cqe);
}
extern inline void
rust_io_uring_sqe_set_data64(struct io_uring_sqe* sqe, __u64 data)
{
io_uring_sqe_set_data64(sqe, data);
}
extern inline __u64
rust_io_uring_cqe_get_data64(struct io_uring_cqe* cqe)
{
return io_uring_cqe_get_data64(cqe);
}
extern inline void
rust_io_uring_sqe_set_flags(struct io_uring_sqe* sqe, unsigned flags)
{
io_uring_sqe_set_flags(sqe, flags);
}
extern inline void
rust_io_uring_prep_rw(int op,
struct io_uring_sqe* sqe,
int fd,
const void* addr,
unsigned len,
__u64 offset)
{
io_uring_prep_rw(op, sqe, fd, addr, len, offset);
}
extern inline void
rust_io_uring_prep_splice(struct io_uring_sqe* sqe,
int fd_in,
int64_t off_in,
int fd_out,
int64_t off_out,
unsigned int nbytes,
unsigned int splice_flags)
{
io_uring_prep_splice(
sqe, fd_in, off_in, fd_out, off_out, nbytes, splice_flags);
}
extern inline void
rust_io_uring_prep_tee(struct io_uring_sqe* sqe,
int fd_in,
int fd_out,
unsigned int nbytes,
unsigned int splice_flags)
{
io_uring_prep_tee(sqe, fd_in, fd_out, nbytes, splice_flags);
}
extern inline void
rust_io_uring_prep_readv(struct io_uring_sqe* sqe,
int fd,
const struct iovec* iovecs,
unsigned nr_vecs,
off_t offset)
{
io_uring_prep_readv(sqe, fd, iovecs, nr_vecs, offset);
}
extern inline void
rust_io_uring_prep_readv2(struct io_uring_sqe* sqe,
int fd,
const struct iovec* iovecs,
unsigned nr_vecs,
__u64 offset,
int flags)
{
io_uring_prep_readv2(sqe, fd, iovecs, nr_vecs, offset, flags);
}
extern inline void
rust_io_uring_prep_read_fixed(struct io_uring_sqe* sqe,
int fd,
void* buf,
unsigned nbytes,
off_t offset,
int buf_index)
{
io_uring_prep_read_fixed(sqe, fd, buf, nbytes, offset, buf_index);
}
extern inline void
rust_io_uring_prep_writev(struct io_uring_sqe* sqe,
int fd,
const struct iovec* iovecs,
unsigned nr_vecs,
off_t offset)
{
io_uring_prep_writev(sqe, fd, iovecs, nr_vecs, offset);
}
extern inline void
rust_io_uring_prep_writev2(struct io_uring_sqe* sqe,
int fd,
const struct iovec* iovecs,
unsigned nr_vecs,
__u64 offset,
int flags)
{
io_uring_prep_writev2(sqe, fd, iovecs, nr_vecs, offset, flags);
}
extern inline void
rust_io_uring_prep_write_fixed(struct io_uring_sqe* sqe,
int fd,
const void* buf,
unsigned nbytes,
off_t offset,
int buf_index)
{
io_uring_prep_write_fixed(sqe, fd, buf, nbytes, offset, buf_index);
}
extern inline void
rust_io_uring_prep_recvmsg(struct io_uring_sqe* sqe,
int fd,
struct msghdr* msg,
unsigned flags)
{
io_uring_prep_recvmsg(sqe, fd, msg, flags);
}
extern inline void
rust_io_uring_prep_sendmsg(struct io_uring_sqe* sqe,
int fd,
const struct msghdr* msg,
unsigned flags)
{
io_uring_prep_sendmsg(sqe, fd, msg, flags);
}
extern inline void
rust_io_uring_prep_poll_add(struct io_uring_sqe* sqe, int fd, short poll_mask)
{
io_uring_prep_poll_add(sqe, fd, poll_mask);
}
extern inline void
rust_io_uring_prep_poll_multishot(struct io_uring_sqe* sqe,
int fd,
short poll_mask)
{
io_uring_prep_poll_multishot(sqe, fd, poll_mask);
}
extern inline void
rust_io_uring_prep_poll_remove(struct io_uring_sqe* sqe, __u64 user_data)
{
io_uring_prep_poll_remove(sqe, user_data);
}
extern inline void
rust_io_uring_prep_poll_update(struct io_uring_sqe* sqe,
__u64 old_user_data,
__u64 new_user_data,
unsigned poll_mask,
unsigned flags)
{
io_uring_prep_poll_update(
sqe, old_user_data, new_user_data, poll_mask, flags);
}
extern inline void
rust_io_uring_prep_fsync(struct io_uring_sqe* sqe, int fd, unsigned fsync_flags)
{
io_uring_prep_fsync(sqe, fd, fsync_flags);
}
extern inline void
rust_io_uring_prep_nop(struct io_uring_sqe* sqe)
{
io_uring_prep_nop(sqe);
}
extern inline void
rust_io_uring_prep_timeout(struct io_uring_sqe* sqe,
struct __kernel_timespec* ts,
unsigned count,
unsigned flags)
{
io_uring_prep_timeout(sqe, ts, count, flags);
}
extern inline void
rust_io_uring_prep_timeout_remove(struct io_uring_sqe* sqe,
__u64 user_data,
unsigned flags)
{
io_uring_prep_timeout_remove(sqe, user_data, flags);
}
extern inline void
rust_io_uring_prep_timeout_update(struct io_uring_sqe* sqe,
struct __kernel_timespec* ts,
__u64 user_data,
unsigned flags)
{
io_uring_prep_timeout_update(sqe, ts, user_data, flags);
}
extern inline void
rust_io_uring_prep_accept(struct io_uring_sqe* sqe,
int fd,
struct sockaddr* addr,
socklen_t* addrlen,
int flags)
{
io_uring_prep_accept(sqe, fd, addr, addrlen, flags);
}
extern inline void
rust_io_uring_prep_accept_direct(struct io_uring_sqe* sqe,
int fd,
struct sockaddr* addr,
socklen_t* addrlen,
int flags,
unsigned file_index)
{
io_uring_prep_accept_direct(sqe, fd, addr, addrlen, flags, file_index);
}
extern inline void
rust_io_uring_prep_cancel(struct io_uring_sqe* sqe, __u64 user_data, int flags)
{
io_uring_prep_cancel(sqe, user_data, flags);
}
extern inline void
rust_io_uring_prep_link_timeout(struct io_uring_sqe* sqe,
struct __kernel_timespec* ts,
unsigned flags)
{
io_uring_prep_link_timeout(sqe, ts, flags);
}
extern inline void
rust_io_uring_prep_connect(struct io_uring_sqe* sqe,
int fd,
struct sockaddr* addr,
socklen_t addrlen)
{
io_uring_prep_connect(sqe, fd, addr, addrlen);
}
extern inline void
rust_io_uring_prep_files_update(struct io_uring_sqe* sqe,
int* fds,
unsigned nr_fds,
int offset)
{
io_uring_prep_files_update(sqe, fds, nr_fds, offset);
}
extern inline void
rust_io_uring_prep_fallocate(struct io_uring_sqe* sqe,
int fd,
int mode,
off_t offset,
off_t len)
{
io_uring_prep_fallocate(sqe, fd, mode, offset, len);
}
extern inline void
rust_io_uring_prep_openat(struct io_uring_sqe* sqe,
int dfd,
const char* path,
int flags,
mode_t mode)
{
io_uring_prep_openat(sqe, dfd, path, flags, mode);
}
extern inline void
rust_io_uring_prep_openat_direct(struct io_uring_sqe* sqe,
int dfd,
const char* path,
int flags,
mode_t mode,
unsigned file_index)
{
io_uring_prep_openat_direct(sqe, dfd, path, flags, mode, file_index);
}
extern inline void
rust_io_uring_prep_close(struct io_uring_sqe* sqe, int fd)
{
io_uring_prep_close(sqe, fd);
}
extern inline void
rust_io_uring_prep_close_direct(struct io_uring_sqe* sqe, int file_index)
{
io_uring_prep_close_direct(sqe, file_index);
}
extern inline void
rust_io_uring_prep_read(struct io_uring_sqe* sqe,
int fd,
void* buf,
unsigned nbytes,
off_t offset)
{
io_uring_prep_read(sqe, fd, buf, nbytes, offset);
}
extern inline void
rust_io_uring_prep_write(struct io_uring_sqe* sqe,
int fd,
const void* buf,
unsigned nbytes,
off_t offset)
{
io_uring_prep_write(sqe, fd, buf, nbytes, offset);
}
extern inline void
rust_io_uring_prep_statx(struct io_uring_sqe* sqe,
int dfd,
const char* path,
int flags,
unsigned mask,
struct statx* statxbuf)
{
io_uring_prep_statx(sqe, dfd, path, flags, mask, statxbuf);
}
extern inline void
rust_io_uring_prep_fadvise(struct io_uring_sqe* sqe,
int fd,
off_t offset,
off_t len,
int advice)
{
io_uring_prep_fadvise(sqe, fd, offset, len, advice);
}
extern inline void
rust_io_uring_prep_madvise(struct io_uring_sqe* sqe,
void* addr,
off_t length,
int advice)
{
io_uring_prep_madvise(sqe, addr, length, advice);
}
extern inline void
rust_io_uring_prep_send(struct io_uring_sqe* sqe,
int sockfd,
const void* buf,
size_t len,
int flags)
{
io_uring_prep_send(sqe, sockfd, buf, len, flags);
}
extern inline void
rust_io_uring_prep_recv(struct io_uring_sqe* sqe,
int sockfd,
void* buf,
size_t len,
int flags)
{
io_uring_prep_recv(sqe, sockfd, buf, len, flags);
}
extern inline void
rust_io_uring_prep_openat2(struct io_uring_sqe* sqe,
int dfd,
const char* path,
struct open_how* how)
{
io_uring_prep_openat2(sqe, dfd, path, how);
}
extern inline void
rust_io_uring_prep_openat2_direct(struct io_uring_sqe* sqe,
int dfd,
const char* path,
struct open_how* how,
unsigned file_index)
{
io_uring_prep_openat2_direct(sqe, dfd, path, how, file_index);
}
extern inline void
rust_io_uring_prep_epoll_ctl(struct io_uring_sqe* sqe,
int epfd,
int fd,
int op,
struct epoll_event* ev)
{
io_uring_prep_epoll_ctl(sqe, epfd, fd, op, ev);
}
extern inline void
rust_io_uring_prep_provide_buffers(struct io_uring_sqe* sqe,
void* addr,
int len,
int nr,
int bgid,
int bid)
{
io_uring_prep_provide_buffers(sqe, addr, len, nr, bgid, bid);
}
extern inline void
rust_io_uring_prep_remove_buffers(struct io_uring_sqe* sqe, int nr, int bgid)
{
io_uring_prep_remove_buffers(sqe, nr, bgid);
}
extern inline void
rust_io_uring_prep_shutdown(struct io_uring_sqe* sqe, int fd, int how)
{
io_uring_prep_shutdown(sqe, fd, how);
}
extern inline void
rust_io_uring_prep_unlinkat(struct io_uring_sqe* sqe,
int dfd,
const char* path,
int flags)
{
io_uring_prep_unlinkat(sqe, dfd, path, flags);
}
extern inline void
rust_io_uring_prep_renameat(struct io_uring_sqe* sqe,
int olddfd,
const char* oldpath,
int newdfd,
const char* newpath,
int flags)
{
io_uring_prep_renameat(sqe, olddfd, oldpath, newdfd, newpath, flags);
}
extern inline void
rust_io_uring_prep_sync_file_range(struct io_uring_sqe* sqe,
int fd,
unsigned len,
__u64 offset,
int flags)
{
io_uring_prep_sync_file_range(sqe, fd, len, offset, flags);
}
extern inline void
rust_io_uring_prep_mkdirat(struct io_uring_sqe* sqe,
int dfd,
const char* path,
mode_t mode)
{
io_uring_prep_mkdirat(sqe, dfd, path, mode);
}
extern inline void
rust_io_uring_prep_symlinkat(struct io_uring_sqe* sqe,
const char* target,
int newdirfd,
const char* linkpath)
{
io_uring_prep_symlinkat(sqe, target, newdirfd, linkpath);
}
extern inline void
rust_io_uring_prep_linkat(struct io_uring_sqe* sqe,
int olddfd,
const char* oldpath,
int newdfd,
const char* newpath,
int flags)
{
io_uring_prep_linkat(sqe, olddfd, oldpath, newdfd, newpath, flags);
}
extern inline unsigned
rust_io_uring_sq_ready(struct io_uring* ring)
{
return io_uring_sq_ready(ring);
}
extern inline unsigned
rust_io_uring_sq_space_left(struct io_uring* ring)
{
return io_uring_sq_space_left(ring);
}
extern inline int
rust_io_uring_sqring_wait(struct io_uring* ring)
{
return io_uring_sqring_wait(ring);
}
extern inline unsigned
rust_io_uring_cq_ready(struct io_uring* ring)
{
return io_uring_cq_ready(ring);
}
extern inline bool
rust_io_uring_cq_eventfd_enabled(struct io_uring* ring)
{
return io_uring_cq_eventfd_enabled(ring);
}
extern inline int
rust_io_uring_cq_eventfd_toggle(struct io_uring* ring, bool enabled)
{
return io_uring_cq_eventfd_toggle(ring, enabled);
}
extern inline int
rust_io_uring_wait_cqe_nr(struct io_uring* ring,
struct io_uring_cqe** cqe_ptr,
unsigned wait_nr)
{
return io_uring_wait_cqe_nr(ring, cqe_ptr, wait_nr);
}
extern inline int
rust_io_uring_peek_cqe(struct io_uring* ring, struct io_uring_cqe** cqe_ptr)
{
return io_uring_peek_cqe(ring, cqe_ptr);
}
extern inline int
rust_io_uring_wait_cqe(struct io_uring* ring, struct io_uring_cqe** cqe_ptr)
{
return io_uring_wait_cqe(ring, cqe_ptr);
}
extern inline struct io_uring_sqe*
rust_io_uring_get_sqe(struct io_uring* ring)
{
return io_uring_get_sqe(ring);
}