use std::collections::HashMap;
use std::future::Future;
use std::num::NonZeroUsize;
use std::ops::ControlFlow;
use std::pin::Pin;
use std::sync::{Arc, Weak};
use std::task::{Context, Poll};
use std::thread::available_parallelism;
use futures::stream::{AbortHandle, Abortable};
use futures::task::AtomicWaker;
use lsp_types::notification::{self, Notification};
use pin_project_lite::pin_project;
use tower_layer::Layer;
use tower_service::Service;
use crate::{
AnyEvent, AnyNotification, AnyRequest, ErrorCode, LspService, RequestId, ResponseError, Result,
};
pub struct Concurrency<S> {
service: S,
max_concurrency: NonZeroUsize,
semaphore: Arc<AtomicWaker>,
ongoing: HashMap<RequestId, AbortHandle>,
}
define_getters!(impl[S] Concurrency<S>, service: S);
impl<S: LspService> Service<AnyRequest> for Concurrency<S>
where
S::Error: From<ResponseError>,
{
type Response = S::Response;
type Error = S::Error;
type Future = ResponseFuture<S::Future>;
fn poll_ready(&mut self, cx: &mut Context<'_>) -> Poll<Result<(), Self::Error>> {
if Arc::weak_count(&self.semaphore) >= self.max_concurrency.get() {
self.semaphore.register(cx.waker());
if Arc::weak_count(&self.semaphore) >= self.max_concurrency.get() {
return Poll::Pending;
}
}
Poll::Ready(Ok(()))
}
fn call(&mut self, req: AnyRequest) -> Self::Future {
let guard = SemaphoreGuard(Arc::downgrade(&self.semaphore));
debug_assert!(
Arc::weak_count(&self.semaphore) <= self.max_concurrency.get(),
"`poll_ready` is not called before `call`",
);
let (handle, registration) = AbortHandle::new_pair();
if self.ongoing.len() >= self.max_concurrency.get() * 2 {
self.ongoing.retain(|_, handle| !handle.is_aborted());
}
self.ongoing.insert(req.id.clone(), handle.clone());
let fut = self.service.call(req);
let fut = Abortable::new(fut, registration);
ResponseFuture {
fut,
_abort_on_drop: AbortOnDrop(handle),
_guard: guard,
}
}
}
struct SemaphoreGuard(Weak<AtomicWaker>);
impl Drop for SemaphoreGuard {
fn drop(&mut self) {
if let Some(sema) = self.0.upgrade() {
if let Some(waker) = sema.take() {
drop(sema);
waker.wake();
}
}
}
}
struct AbortOnDrop(AbortHandle);
impl Drop for AbortOnDrop {
fn drop(&mut self) {
self.0.abort();
}
}
pin_project! {
pub struct ResponseFuture<Fut> {
#[pin]
fut: Abortable<Fut>,
_abort_on_drop: AbortOnDrop,
_guard: SemaphoreGuard,
}
}
impl<Fut, Response, Error> Future for ResponseFuture<Fut>
where
Fut: Future<Output = Result<Response, Error>>,
Error: From<ResponseError>,
{
type Output = Fut::Output;
fn poll(self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Self::Output> {
match self.project().fut.poll(cx) {
Poll::Pending => Poll::Pending,
Poll::Ready(Ok(inner_ret)) => Poll::Ready(inner_ret),
Poll::Ready(Err(_aborted)) => Poll::Ready(Err(ResponseError {
code: ErrorCode::REQUEST_CANCELLED,
message: "Client cancelled the request".into(),
data: None,
}
.into())),
}
}
}
impl<S: LspService> LspService for Concurrency<S>
where
S::Error: From<ResponseError>,
{
fn notify(&mut self, notif: AnyNotification) -> ControlFlow<Result<()>> {
if notif.method == notification::Cancel::METHOD {
if let Ok(params) = serde_json::from_value::<lsp_types::CancelParams>(notif.params) {
self.ongoing.remove(¶ms.id);
}
return ControlFlow::Continue(());
}
self.service.notify(notif)
}
fn emit(&mut self, event: AnyEvent) -> ControlFlow<Result<()>> {
self.service.emit(event)
}
}
#[derive(Clone, Debug)]
#[must_use]
pub struct ConcurrencyBuilder {
max_concurrency: NonZeroUsize,
}
impl Default for ConcurrencyBuilder {
fn default() -> Self {
Self::new(available_parallelism().unwrap_or(NonZeroUsize::new(1).unwrap()))
}
}
impl ConcurrencyBuilder {
pub fn new(max_concurrency: NonZeroUsize) -> Self {
Self { max_concurrency }
}
}
pub type ConcurrencyLayer = ConcurrencyBuilder;
impl<S> Layer<S> for ConcurrencyBuilder {
type Service = Concurrency<S>;
fn layer(&self, inner: S) -> Self::Service {
Concurrency {
service: inner,
max_concurrency: self.max_concurrency,
semaphore: Arc::new(AtomicWaker::new()),
ongoing: HashMap::with_capacity(
self.max_concurrency
.get()
.checked_mul(2)
.expect("max_concurrency overflow"),
),
}
}
}