0
0
Fork 0
mirror of https://github.com/denoland/rusty_v8.git synced 2024-12-26 09:13:46 -05:00
denoland-rusty-v8/src/shared_array_buffer.rs

231 lines
7.4 KiB
Rust

// Copyright 2019-2021 the Deno authors. All rights reserved. MIT license.
use std::ffi::c_void;
use crate::support::SharedRef;
use crate::support::UniqueRef;
use crate::BackingStore;
use crate::BackingStoreDeleterCallback;
use crate::HandleScope;
use crate::Isolate;
use crate::Local;
use crate::SharedArrayBuffer;
extern "C" {
fn v8__SharedArrayBuffer__New__with_byte_length(
isolate: *mut Isolate,
byte_length: usize,
) -> *const SharedArrayBuffer;
fn v8__SharedArrayBuffer__New__with_backing_store(
isolate: *mut Isolate,
backing_store: *const SharedRef<BackingStore>,
) -> *const SharedArrayBuffer;
fn v8__SharedArrayBuffer__ByteLength(this: *const SharedArrayBuffer)
-> usize;
fn v8__SharedArrayBuffer__GetBackingStore(
this: *const SharedArrayBuffer,
) -> SharedRef<BackingStore>;
fn v8__SharedArrayBuffer__NewBackingStore__with_byte_length(
isolate: *mut Isolate,
byte_length: usize,
) -> *mut BackingStore;
fn v8__SharedArrayBuffer__NewBackingStore__with_data(
data: *mut c_void,
byte_length: usize,
deleter: BackingStoreDeleterCallback,
deleter_data: *mut c_void,
) -> *mut BackingStore;
fn v8__BackingStore__EmptyBackingStore(shared: bool) -> *mut BackingStore;
}
impl SharedArrayBuffer {
/// Create a new SharedArrayBuffer. Allocate |byte_length| bytes.
/// Allocated memory will be owned by a created SharedArrayBuffer and
/// will be deallocated when it is garbage-collected,
/// unless the object is externalized.
#[inline(always)]
pub fn new<'s>(
scope: &mut HandleScope<'s>,
byte_length: usize,
) -> Option<Local<'s, SharedArrayBuffer>> {
unsafe {
scope.cast_local(|sd| {
v8__SharedArrayBuffer__New__with_byte_length(
sd.get_isolate_ptr(),
byte_length,
)
})
}
}
#[inline(always)]
pub fn with_backing_store<'s>(
scope: &mut HandleScope<'s>,
backing_store: &SharedRef<BackingStore>,
) -> Local<'s, SharedArrayBuffer> {
unsafe {
scope.cast_local(|sd| {
v8__SharedArrayBuffer__New__with_backing_store(
sd.get_isolate_ptr(),
backing_store,
)
})
}
.unwrap()
}
/// Create a new, empty SharedArrayBuffer.
#[inline(always)]
pub fn empty<'s>(
scope: &mut HandleScope<'s>,
) -> Local<'s, SharedArrayBuffer> {
// SAFETY: This is a v8-provided empty backing store
let backing_store =
unsafe { UniqueRef::from_raw(v8__BackingStore__EmptyBackingStore(true)) };
Self::with_backing_store(scope, &backing_store.make_shared())
}
/// Data length in bytes.
#[inline(always)]
pub fn byte_length(&self) -> usize {
unsafe { v8__SharedArrayBuffer__ByteLength(self) }
}
/// Get a shared pointer to the backing store of this array buffer. This
/// pointer coordinates the lifetime management of the internal storage
/// with any live ArrayBuffers on the heap, even across isolates. The embedder
/// should not attempt to manage lifetime of the storage through other means.
#[inline(always)]
pub fn get_backing_store(&self) -> SharedRef<BackingStore> {
unsafe { v8__SharedArrayBuffer__GetBackingStore(self) }
}
/// Returns a new standalone BackingStore that is allocated using the array
/// buffer allocator of the isolate. The result can be later passed to
/// ArrayBuffer::New.
///
/// If the allocator returns nullptr, then the function may cause GCs in the
/// given isolate and re-try the allocation. If GCs do not help, then the
/// function will crash with an out-of-memory error.
#[inline(always)]
pub fn new_backing_store(
scope: &mut Isolate,
byte_length: usize,
) -> UniqueRef<BackingStore> {
unsafe {
UniqueRef::from_raw(
v8__SharedArrayBuffer__NewBackingStore__with_byte_length(
scope,
byte_length,
),
)
}
}
/// Returns a new standalone BackingStore that takes over the ownership of
/// the given buffer.
///
/// The destructor of the BackingStore frees owned buffer memory.
///
/// The result can be later passed to SharedArrayBuffer::New. The raw pointer
/// to the buffer must not be passed again to any V8 API function.
#[inline(always)]
pub fn new_backing_store_from_boxed_slice(
data: Box<[u8]>,
) -> UniqueRef<BackingStore> {
Self::new_backing_store_from_bytes(data)
}
/// Returns a new standalone BackingStore that takes over the ownership of
/// the given buffer.
///
/// The destructor of the BackingStore frees owned buffer memory.
///
/// The result can be later passed to SharedArrayBuffer::New. The raw pointer
/// to the buffer must not be passed again to any V8 API function.
#[inline(always)]
pub fn new_backing_store_from_vec(data: Vec<u8>) -> UniqueRef<BackingStore> {
Self::new_backing_store_from_bytes(data)
}
/// Returns a new standalone BackingStore backed by a container that dereferences
/// to a mutable slice of bytes. The object is dereferenced once, and the resulting slice's
/// memory is used for the lifetime of the buffer.
///
/// This method may be called with most single-ownership containers that implement `AsMut<[u8]>`, including
/// `Box<[u8]>`, and `Vec<u8>`. This will also support most other mutable bytes containers (including `bytes::BytesMut`),
/// though these buffers will need to be boxed to manage ownership of memory.
///
/// ```
/// // Vector of bytes
/// let backing_store = v8::ArrayBuffer::new_backing_store_from_bytes(vec![1, 2, 3]);
/// // Boxes slice of bytes
/// let boxed_slice: Box<[u8]> = vec![1, 2, 3].into_boxed_slice();
/// let backing_store = v8::ArrayBuffer::new_backing_store_from_bytes(boxed_slice);
/// // BytesMut from bytes crate
/// let backing_store = v8::ArrayBuffer::new_backing_store_from_bytes(Box::new(bytes::BytesMut::new()));
/// ```
#[inline(always)]
pub fn new_backing_store_from_bytes<T, U>(
mut bytes: T,
) -> UniqueRef<BackingStore>
where
U: ?Sized,
U: AsMut<[u8]>,
T: AsMut<U>,
T: crate::array_buffer::sealed::Rawable<U>,
{
let len = bytes.as_mut().as_mut().len();
if len == 0 {
return unsafe {
UniqueRef::from_raw(v8__BackingStore__EmptyBackingStore(false))
};
}
let (ptr, slice) = T::into_raw(bytes);
extern "C" fn drop_rawable<
T: crate::array_buffer::sealed::Rawable<U>,
U: ?Sized,
>(
_ptr: *mut c_void,
len: usize,
data: *mut c_void,
) {
// SAFETY: We know that data is a raw T from above
unsafe {
<T as crate::array_buffer::sealed::Rawable<U>>::drop_raw(data as _, len)
}
}
// SAFETY: We are extending the lifetime of a slice, but we're locking away the box that we
// derefed from so there's no way to get another mutable reference.
unsafe {
Self::new_backing_store_from_ptr(
slice as _,
len,
drop_rawable::<T, U>,
ptr as _,
)
}
}
/// Returns a new standalone shared BackingStore backed by given ptr.
///
/// SAFETY: This API consumes raw pointers so is inherently
/// unsafe. Usually you should use new_backing_store_from_boxed_slice.
#[inline(always)]
pub unsafe fn new_backing_store_from_ptr(
data_ptr: *mut c_void,
byte_length: usize,
deleter_callback: BackingStoreDeleterCallback,
deleter_data: *mut c_void,
) -> UniqueRef<BackingStore> {
UniqueRef::from_raw(v8__SharedArrayBuffer__NewBackingStore__with_data(
data_ptr,
byte_length,
deleter_callback,
deleter_data,
))
}
}