use std::alloc::{self, Layout};
use std::borrow::Borrow;
use std::fmt;
const HEAP_FLAG: usize = 1 << (usize::BITS - 1);
const LEN_MASK: usize = !HEAP_FLAG;
const ALIGN: usize = std::mem::align_of::<u8>();
pub const INIT_CAPACITY: usize = std::mem::size_of::<Buffer>() - std::mem::size_of::<u8>();
#[repr(C)]
pub struct Buffer {
data_: *mut u8,
cap_: usize,
len_: usize,
}
unsafe impl Send for Buffer {}
unsafe impl Sync for Buffer {}
impl Drop for Buffer {
fn drop(&mut self) {
if self.is_stack() {
return;
}
unsafe {
let layout = Layout::from_size_align_unchecked(self.cap_, ALIGN);
alloc::dealloc(self.data_, layout);
}
}
}
impl Clone for Buffer {
fn clone(&self) -> Self {
let mut ret = Self::with_capacity(self.len());
unsafe { ret.extend_from_slice(self.as_slice()) };
ret
}
}
impl Buffer {
pub const fn new() -> Self {
Self {
data_: std::ptr::null_mut(),
cap_: 0,
len_: 0,
}
}
pub fn with_capacity(cap: usize) -> Self {
let mut buffer = Self::new();
buffer.reserve(cap);
buffer
}
pub fn from_vec(mut vals: Vec<u8>) -> Self {
if vals.capacity() == 0 {
Self::new()
} else {
let ret = Self {
data_: vals.as_mut_ptr(),
cap_: vals.capacity(),
len_: vals.len() | HEAP_FLAG,
};
std::mem::forget(vals);
ret
}
}
pub unsafe fn push(&mut self, v: u8) {
let old_len = self.len();
debug_assert!(old_len < self.capacity());
self.set_len(old_len + 1);
self.as_mut_slice()[old_len] = v;
}
pub unsafe fn extend_from_slice(&mut self, vals: &[u8]) {
let ptr = self.as_mut_ptr().add(self.len());
ptr.copy_from_nonoverlapping(vals.as_ptr(), vals.len());
self.set_len(self.len() + vals.len());
}
pub const fn len(&self) -> usize {
if self.is_stack() {
self.len_ >> (usize::BITS - u8::BITS)
} else {
self.len_ & LEN_MASK
}
}
pub fn capacity(&self) -> usize {
if self.is_stack() {
INIT_CAPACITY
} else {
self.cap_
}
}
pub fn reserve(&mut self, additional: usize) {
let new_cap = self.len() + additional;
if new_cap <= self.capacity() {
return;
}
if self.is_stack() {
unsafe {
let layout = Layout::array::<u8>(new_cap).unwrap();
let ptr = alloc::alloc(layout);
if ptr.is_null() {
alloc::handle_alloc_error(layout);
}
ptr.copy_from_nonoverlapping(self.as_ptr(), self.len());
self.data_ = ptr;
self.cap_ = new_cap;
self.len_ = self.len() | HEAP_FLAG;
}
} else {
unsafe {
let layout = Layout::array::<u8>(self.capacity()).unwrap();
let ptr = alloc::realloc(self.data_, layout, new_cap);
if ptr.is_null() {
let layout = Layout::array::<u8>(new_cap).unwrap();
alloc::handle_alloc_error(layout);
}
self.data_ = ptr;
self.cap_ = new_cap;
}
}
}
pub unsafe fn set_len(&mut self, len: usize) {
if self.is_stack() {
const MASK: usize = (u8::MAX as usize) << (usize::BITS - u8::BITS);
self.len_ &= !MASK;
self.len_ |= len << (usize::BITS - u8::BITS);
} else {
self.len_ = len | HEAP_FLAG;
}
}
pub fn as_ptr(&self) -> *const u8 {
if self.is_stack() {
let ptr = self as *const Self;
ptr as *const u8
} else {
self.data_
}
}
pub fn as_mut_ptr(&mut self) -> *mut u8 {
let ptr = self.as_ptr();
ptr as *mut u8
}
const fn is_stack(&self) -> bool {
self.len_ & HEAP_FLAG == 0
}
pub fn as_slice(&self) -> &[u8] {
unsafe { std::slice::from_raw_parts(self.as_ptr(), self.len()) }
}
pub fn as_mut_slice(&mut self) -> &mut [u8] {
unsafe { std::slice::from_raw_parts_mut(self.as_mut_ptr(), self.len()) }
}
pub fn into_vec(self) -> Vec<u8> {
if self.is_stack() {
Vec::from(self.as_slice())
} else {
unsafe {
let vec = Vec::from_raw_parts(self.data_, self.len(), self.cap_);
std::mem::forget(self);
vec
}
}
}
}
impl fmt::Debug for Buffer {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
let contents: &[u8] = self.as_slice();
f.debug_tuple("Buffer").field(&contents).finish()
}
}
impl Borrow<[u8]> for Buffer {
fn borrow(&self) -> &[u8] {
self.as_slice()
}
}
impl<T> PartialEq<T> for Buffer
where
T: Borrow<[u8]>,
{
fn eq(&self, other: &T) -> bool {
self.as_slice() == other.borrow()
}
}
impl Eq for Buffer {}