1
0
Fork 0
mirror of https://github.com/denoland/deno.git synced 2024-11-22 15:06:54 -05:00
denoland-deno/core/ops.rs

174 lines
4.5 KiB
Rust
Raw Normal View History

2019-09-30 14:59:44 -04:00
// Copyright 2018-2019 the Deno authors. All rights reserved. MIT license.
pub use crate::libdeno::OpId;
use crate::PinnedBuf;
use futures::Future;
use std::collections::HashMap;
2019-11-16 19:17:47 -05:00
use std::pin::Pin;
use std::sync::Arc;
use std::sync::RwLock;
2019-09-30 14:59:44 -04:00
pub type Buf = Box<[u8]>;
2019-11-16 19:17:47 -05:00
pub type OpAsyncFuture<E> =
Pin<Box<dyn Future<Output = Result<Buf, E>> + Send>>;
2019-09-30 14:59:44 -04:00
pub(crate) type PendingOpFuture =
2019-11-16 19:17:47 -05:00
Pin<Box<dyn Future<Output = Result<(OpId, Buf), CoreError>> + Send>>;
2019-09-30 14:59:44 -04:00
pub type OpResult<E> = Result<Op<E>, E>;
pub enum Op<E> {
Sync(Buf),
Async(OpAsyncFuture<E>),
}
pub type CoreError = ();
pub type CoreOp = Op<CoreError>;
/// Main type describing op
pub type OpDispatcher =
dyn Fn(&[u8], Option<PinnedBuf>) -> CoreOp + Send + Sync + 'static;
2019-09-30 14:59:44 -04:00
#[derive(Default)]
pub struct OpRegistry {
dispatchers: RwLock<Vec<Arc<Box<OpDispatcher>>>>,
name_to_id: RwLock<HashMap<String, OpId>>,
2019-09-30 14:59:44 -04:00
}
impl OpRegistry {
pub fn new() -> Self {
let registry = Self::default();
2019-09-30 14:59:44 -04:00
let op_id = registry.register("ops", |_, _| {
// ops is a special op which is handled in call.
unreachable!()
});
assert_eq!(op_id, 0);
registry
}
pub fn register<F>(&self, name: &str, op: F) -> OpId
2019-09-30 14:59:44 -04:00
where
F: Fn(&[u8], Option<PinnedBuf>) -> CoreOp + Send + Sync + 'static,
{
let mut lock = self.dispatchers.write().unwrap();
let op_id = lock.len() as u32;
2019-09-30 14:59:44 -04:00
let mut name_lock = self.name_to_id.write().unwrap();
let existing = name_lock.insert(name.to_string(), op_id);
2019-09-30 14:59:44 -04:00
assert!(
existing.is_none(),
format!("Op already registered: {}", name)
);
lock.push(Arc::new(Box::new(op)));
drop(name_lock);
drop(lock);
2019-09-30 14:59:44 -04:00
op_id
}
fn json_map(&self) -> Buf {
let lock = self.name_to_id.read().unwrap();
let op_map_json = serde_json::to_string(&*lock).unwrap();
2019-09-30 14:59:44 -04:00
op_map_json.as_bytes().to_owned().into_boxed_slice()
}
/// This function returns None only if op with given id doesn't exist in registry.
2019-09-30 14:59:44 -04:00
pub fn call(
&self,
op_id: OpId,
control: &[u8],
zero_copy_buf: Option<PinnedBuf>,
) -> Option<CoreOp> {
2019-09-30 14:59:44 -04:00
// Op with id 0 has special meaning - it's a special op that is always
// provided to retrieve op id map. The map consists of name to `OpId`
// mappings.
if op_id == 0 {
return Some(Op::Sync(self.json_map()));
2019-09-30 14:59:44 -04:00
}
let lock = self.dispatchers.read().unwrap();
if let Some(op) = lock.get(op_id as usize) {
let op_ = Arc::clone(&op);
// This should allow for changes to the dispatcher list during a call.
drop(lock);
Some(op_(control, zero_copy_buf))
} else {
None
}
2019-09-30 14:59:44 -04:00
}
}
#[test]
fn test_op_registry() {
use std::sync::atomic;
use std::sync::Arc;
let op_registry = OpRegistry::new();
2019-09-30 14:59:44 -04:00
let c = Arc::new(atomic::AtomicUsize::new(0));
let c_ = c.clone();
let test_id = op_registry.register("test", move |_, _| {
c_.fetch_add(1, atomic::Ordering::SeqCst);
CoreOp::Sync(Box::new([]))
});
assert!(test_id != 0);
let mut expected = HashMap::new();
expected.insert("ops".to_string(), 0);
expected.insert("test".to_string(), 1);
let name_to_id = op_registry.name_to_id.read().unwrap();
assert_eq!(*name_to_id, expected);
2019-09-30 14:59:44 -04:00
let res = op_registry.call(test_id, &[], None).unwrap();
2019-09-30 14:59:44 -04:00
if let Op::Sync(buf) = res {
assert_eq!(buf.len(), 0);
} else {
unreachable!();
}
assert_eq!(c.load(atomic::Ordering::SeqCst), 1);
let res = op_registry.call(100, &[], None);
assert!(res.is_none());
2019-09-30 14:59:44 -04:00
}
#[test]
fn register_op_during_call() {
use std::sync::atomic;
use std::sync::Arc;
let op_registry = Arc::new(OpRegistry::new());
let c = Arc::new(atomic::AtomicUsize::new(0));
let c_ = c.clone();
let op_registry_ = op_registry.clone();
let test_id = op_registry.register("dynamic_register_op", move |_, _| {
let c__ = c_.clone();
op_registry_.register("test", move |_, _| {
c__.fetch_add(1, atomic::Ordering::SeqCst);
CoreOp::Sync(Box::new([]))
});
CoreOp::Sync(Box::new([]))
});
assert!(test_id != 0);
op_registry.call(test_id, &[], None);
let mut expected = HashMap::new();
expected.insert("ops".to_string(), 0);
expected.insert("dynamic_register_op".to_string(), 1);
expected.insert("test".to_string(), 2);
let name_to_id = op_registry.name_to_id.read().unwrap();
assert_eq!(*name_to_id, expected);
let res = op_registry.call(2, &[], None).unwrap();
if let Op::Sync(buf) = res {
assert_eq!(buf.len(), 0);
} else {
unreachable!();
}
assert_eq!(c.load(atomic::Ordering::SeqCst), 1);
let res = op_registry.call(100, &[], None);
assert!(res.is_none());
}