docs: re-documented remaining members

This commit is contained in:
əlemi 2024-09-04 21:37:35 +02:00 committed by zaaarf
parent f0a6a6b730
commit bfe84c45e0
No known key found for this signature in database
GPG key ID: 102E445F4C3F829B
18 changed files with 156 additions and 595 deletions

View file

@ -1,7 +1,5 @@
//! # Cursor //! ### Cursor
//! //! Represents the position of a remote user's cursor, with their display name
//! represents the position of an user's cursor, with
//! information about their identity
use codemp_proto as proto; use codemp_proto as proto;
use uuid::Uuid; use uuid::Uuid;
@ -9,15 +7,18 @@ use uuid::Uuid;
#[cfg(feature = "python")] #[cfg(feature = "python")]
use pyo3::prelude::*; use pyo3::prelude::*;
/// user cursor position in a buffer /// User cursor position in a buffer
#[derive(Clone, Debug, Default)] #[derive(Clone, Debug, Default)]
#[cfg_attr(feature = "python", pyclass)] #[cfg_attr(feature = "python", pyclass)]
// #[cfg_attr(feature = "python", pyo3(crate = "reexported::pyo3"))] // #[cfg_attr(feature = "python", pyo3(crate = "reexported::pyo3"))]
pub struct Cursor { pub struct Cursor {
/// range of text change, as char indexes in buffer previous state /// Cursor start position in buffer, as 0-indexed row-column tuple
pub start: (i32, i32), pub start: (i32, i32),
/// Cursor end position in buffer, as 0-indexed row-column tuple
pub end: (i32, i32), pub end: (i32, i32),
/// Path of buffer this cursor is on
pub buffer: String, pub buffer: String,
/// User display name, if provided
pub user: Option<Uuid>, pub user: Option<Uuid>,
} }

View file

@ -1,10 +1,17 @@
//! # Event
//! Real time notification of changes in a workspace, to either users or buffers
use codemp_proto::workspace::workspace_event::Event as WorkspaceEventInner; use codemp_proto::workspace::workspace_event::Event as WorkspaceEventInner;
/// Event in a [crate::Workspace]
#[derive(Debug, Clone)] #[derive(Debug, Clone)]
#[cfg_attr(feature = "python", pyo3::pyclass)] #[cfg_attr(feature = "python", pyo3::pyclass)]
pub enum Event { pub enum Event {
/// Fired when the file tree changes
/// containes the modified buffer path (deleted or created or renamed)
FileTreeUpdated(String), FileTreeUpdated(String),
/// Fired when an user joins the current workspace
UserJoin(String), UserJoin(String),
/// Fired when an user leaves the current workspace
UserLeave(String), UserLeave(String),
} }

View file

@ -1,8 +1,5 @@
//! # api //! # API
//! //! These traits and structs represent the main `codemp` library API.
//! these traits represent the internal api for the codemp library.
//! more methods and structs are provided but these are the core interfaces to
//! interact with the client.
/// a generic async provider for bidirectional communication /// a generic async provider for bidirectional communication
pub mod controller; pub mod controller;
@ -13,10 +10,10 @@ pub mod change;
/// representation for an user's cursor /// representation for an user's cursor
pub mod cursor; pub mod cursor;
/// workspace events /// live events in workspaces
pub mod event; pub mod event;
/// data structure for service users /// data structure for remote users
pub mod user; pub mod user;
pub use controller::Controller; pub use controller::Controller;

View file

@ -1,15 +1,15 @@
//! # User //! # User
//! //! An user is identified by an UUID, which should never change.
//! data structures for our service users //! Each user has an username, which can change but should be unique.
use uuid::Uuid; use uuid::Uuid;
/// represents a service user /// Represents a service user
///
/// all users are identified uniquely with UUIDs
#[derive(Debug, Clone)] #[derive(Debug, Clone)]
pub struct User { pub struct User {
/// User unique identifier, should never change
pub id: Uuid, pub id: Uuid,
/// User name, can change but should be unique
pub name: String, pub name: String,
} }

View file

@ -1,6 +1,5 @@
//! ### controller //! ### Buffer Controller
//! //! A [Controller] implementation for buffer actions
//! a controller implementation for buffer actions
use std::sync::Arc; use std::sync::Arc;
@ -15,25 +14,22 @@ use crate::ext::InternallyMutable;
use super::worker::DeltaRequest; use super::worker::DeltaRequest;
/// the buffer controller implementation /// A [Controller] to asyncrhonously interact with remote buffers
/// ///
/// for each controller a worker exists, managing outgoing and inbound /// Each buffer controller internally tracks the last acknowledged state, remaining always in sync
/// queues, transforming outbound delayed ops and applying remote changes /// with the server while allowing to procedurally receiving changes while still sending new ones.
/// to the local buffer
///
/// upon dropping this handle will stop the associated worker
#[derive(Debug, Clone)] #[derive(Debug, Clone)]
#[cfg_attr(feature = "python", pyo3::pyclass)] #[cfg_attr(feature = "python", pyo3::pyclass)]
#[cfg_attr(feature = "js", napi_derive::napi)] #[cfg_attr(feature = "js", napi_derive::napi)]
pub struct BufferController(pub(crate) Arc<BufferControllerInner>); pub struct BufferController(pub(crate) Arc<BufferControllerInner>);
impl BufferController { impl BufferController {
/// unique identifier of buffer /// Get the buffer path
pub fn name(&self) -> &str { pub fn path(&self) -> &str {
&self.0.name &self.0.name
} }
/// return buffer whole content, updating internal buffer previous state /// Return buffer whole content, updating internal acknowledgement tracker
pub async fn content(&self) -> ControllerResult<String> { pub async fn content(&self) -> ControllerResult<String> {
let (tx, rx) = oneshot::channel(); let (tx, rx) = oneshot::channel();
self.0.content_request.send(tx).await?; self.0.content_request.send(tx).await?;
@ -60,8 +56,6 @@ pub(crate) struct BufferControllerInner {
#[async_trait] #[async_trait]
impl Controller<TextChange> for BufferController { impl Controller<TextChange> for BufferController {
/// block until a text change is available
/// this returns immediately if one is already available
async fn poll(&self) -> ControllerResult<()> { async fn poll(&self) -> ControllerResult<()> {
if self.0.last_update.get() != *self.0.latest_version.borrow() { if self.0.last_update.get() != *self.0.latest_version.borrow() {
return Ok(()); return Ok(());
@ -73,7 +67,6 @@ impl Controller<TextChange> for BufferController {
Ok(()) Ok(())
} }
/// if a text change is available, return it immediately
async fn try_recv(&self) -> ControllerResult<Option<TextChange>> { async fn try_recv(&self) -> ControllerResult<Option<TextChange>> {
let last_update = self.0.last_update.get(); let last_update = self.0.last_update.get();
let latest_version = self.0.latest_version.borrow().clone(); let latest_version = self.0.latest_version.borrow().clone();
@ -89,8 +82,6 @@ impl Controller<TextChange> for BufferController {
Ok(change) Ok(change)
} }
/// enqueue a text change for processing
/// this also updates internal buffer previous state
async fn send(&self, op: TextChange) -> ControllerResult<()> { async fn send(&self, op: TextChange) -> ControllerResult<()> {
// we let the worker do the updating to the last version and send it back. // we let the worker do the updating to the last version and send it back.
let (tx, rx) = oneshot::channel(); let (tx, rx) = oneshot::channel();

View file

@ -1,10 +1,7 @@
//! ### buffer //! ### Buffer
//! //! A buffer is a container of text, modifiable in sync by users.
//! ![demo gif of early buffer sync in action](https://cdn.alemi.dev/codemp/demo-vscode.gif) //! It is built on top of [diamond_types] CRDT, guaranteeing that all peers which have received the
//! //! same set of operations will converge to the same content.
//! a buffer is a container fo text edited by users.
//! this module contains buffer-related operations and helpers to create Operation Sequences
//! (the underlying chunks of changes sent over the wire)
/// buffer controller implementation /// buffer controller implementation
pub mod controller; pub mod controller;

View file

@ -19,7 +19,7 @@ use pyo3::prelude::*;
/// ///
/// It generates a new UUID and stores user credentials upon connecting. /// It generates a new UUID and stores user credentials upon connecting.
/// ///
/// A new [Client] can be obtained with [Client::connect]. /// A new [`Client`] can be obtained with [`Client::connect`].
#[derive(Debug, Clone)] #[derive(Debug, Clone)]
#[cfg_attr(feature = "js", napi_derive::napi)] #[cfg_attr(feature = "js", napi_derive::napi)]
#[cfg_attr(feature = "python", pyclass)] #[cfg_attr(feature = "python", pyclass)]
@ -36,7 +36,7 @@ struct ClientInner {
} }
impl Client { impl Client {
/// Connect to the server, authenticate and instantiate a new [Client]. /// Connect to the server, authenticate and instantiate a new [`Client`].
pub async fn connect( pub async fn connect(
host: impl AsRef<str>, host: impl AsRef<str>,
username: impl AsRef<str>, username: impl AsRef<str>,
@ -74,7 +74,7 @@ impl Client {
} }
/// Refresh session token. /// Refresh session token.
pub async fn refresh(&self) -> tonic::Result<()> { pub async fn refresh(&self) -> RemoteResult<()> {
let new_token = self.0.auth.clone().refresh(self.0.claims.get()) let new_token = self.0.auth.clone().refresh(self.0.claims.get())
.await? .await?
.into_inner(); .into_inner();
@ -128,7 +128,7 @@ impl Client {
Ok(out) Ok(out)
} }
/// Join and return a [Workspace]. /// Join and return a [`Workspace`].
pub async fn join_workspace(&self, workspace: impl AsRef<str>) -> ConnectionResult<Workspace> { pub async fn join_workspace(&self, workspace: impl AsRef<str>) -> ConnectionResult<Workspace> {
let token = self.0.session let token = self.0.session
.clone() .clone()
@ -152,17 +152,17 @@ impl Client {
Ok(ws) Ok(ws)
} }
/// Leave the [Workspace] with the given name. /// Leave the [`Workspace`] with the given name.
pub fn leave_workspace(&self, id: &str) -> bool { pub fn leave_workspace(&self, id: &str) -> bool {
self.0.workspaces.remove(id).is_some() self.0.workspaces.remove(id).is_some()
} }
/// Gets a [Workspace] handle by name. /// Gets a [`Workspace`] handle by name.
pub fn get_workspace(&self, id: &str) -> Option<Workspace> { pub fn get_workspace(&self, id: &str) -> Option<Workspace> {
self.0.workspaces.get(id).map(|x| x.clone()) self.0.workspaces.get(id).map(|x| x.clone())
} }
/// Get the names of all active [Workspace]s. /// Get the names of all active [`Workspace`]s.
pub fn active_workspaces(&self) -> Vec<String> { pub fn active_workspaces(&self) -> Vec<String> {
self.0 self.0
.workspaces .workspaces

View file

@ -1,6 +1,6 @@
//! ### controller //! ### Cursor Controller
//! //! A [Controller] implementation for [Cursor] actions in a [crate::Workspace]
//! a controller implementation for cursor actions
use std::sync::Arc; use std::sync::Arc;
use tokio::sync::{mpsc, oneshot, watch}; use tokio::sync::{mpsc, oneshot, watch};
@ -8,17 +8,10 @@ use tonic::async_trait;
use crate::{api::{controller::ControllerCallback, Controller, Cursor}, errors::ControllerResult}; use crate::{api::{controller::ControllerCallback, Controller, Cursor}, errors::ControllerResult};
use codemp_proto::cursor::CursorPosition; use codemp_proto::cursor::CursorPosition;
/// the cursor controller implementation
/// A [Controller] for asynchronously sending and receiving [Cursor] events
/// ///
/// this contains /// An unique [CursorController] exists for each active [crate::Workspace].
/// * the unique identifier of current user
/// * a sink to send movements into
/// * a mutex over a stream of inbound cursor events
/// * a channel to stop the associated worker
///
/// for each controller a worker exists , managing outgoing and inbound event queues
///
/// upon dropping this handle will stop the associated worker
#[derive(Debug, Clone)] #[derive(Debug, Clone)]
#[cfg_attr(feature = "python", pyo3::pyclass)] #[cfg_attr(feature = "python", pyo3::pyclass)]
#[cfg_attr(feature = "js", napi_derive::napi)] #[cfg_attr(feature = "js", napi_derive::napi)]
@ -35,8 +28,6 @@ pub(crate) struct CursorControllerInner {
#[async_trait] #[async_trait]
impl Controller<Cursor> for CursorController { impl Controller<Cursor> for CursorController {
/// enqueue a cursor event to be broadcast to current workspace
/// will automatically invert cursor start/end if they are inverted
async fn send(&self, mut cursor: Cursor) -> ControllerResult<()> { async fn send(&self, mut cursor: Cursor) -> ControllerResult<()> {
if cursor.start > cursor.end { if cursor.start > cursor.end {
std::mem::swap(&mut cursor.start, &mut cursor.end); std::mem::swap(&mut cursor.start, &mut cursor.end);
@ -44,14 +35,12 @@ impl Controller<Cursor> for CursorController {
Ok(self.0.op.send(cursor.into()).await?) Ok(self.0.op.send(cursor.into()).await?)
} }
/// try to receive without blocking, but will still block on stream mutex
async fn try_recv(&self) -> ControllerResult<Option<Cursor>> { async fn try_recv(&self) -> ControllerResult<Option<Cursor>> {
let (tx, rx) = oneshot::channel(); let (tx, rx) = oneshot::channel();
self.0.stream.send(tx).await?; self.0.stream.send(tx).await?;
Ok(rx.await?) Ok(rx.await?)
} }
/// await for changed mutex and then next op change
async fn poll(&self) -> ControllerResult<()> { async fn poll(&self) -> ControllerResult<()> {
let (tx, rx) = oneshot::channel(); let (tx, rx) = oneshot::channel();
self.0.poll.send(tx)?; self.0.poll.send(tx)?;

View file

@ -1,9 +1,6 @@
//! ### cursor //! ### Cursor
//! //! Each user in a [crate::Workspace] holds a cursor and can move it across multiple buffers.
//! ![demo gif of early cursor sync in action](https://cdn.alemi.dev/codemp/demo-nvim.gif) //! A cursor spans zero or more characters across one or more lines.
//!
//! each user holds a cursor, which consists of multiple highlighted region
//! on a specific buffer
pub(crate) mod worker; pub(crate) mod worker;

View file

@ -1,3 +1,11 @@
//! ### java
//! Since for java it is necessary to deal with the JNI and no complete FFI library is available,
//! java glue directly writes JNI functions leveraging [jni] rust bindings.
//!
//! To have a runnable `jar`, some extra Java code must be compiled (available under `dist/java`)
//! and bundled together with the shared object. Such extra wrapper provides classes and methods
//! loading the native extension and invoking the underlying native functions.
pub mod client; pub mod client;
pub mod workspace; pub mod workspace;
pub mod cursor; pub mod cursor;

View file

@ -1,3 +1,7 @@
//! ### javascript
//! Using [napi] it's possible to map perfectly the entirety of `codemp` API.
//! Async operations run on a dedicated [tokio] runtime and the result is sent back to main thread
pub mod client; pub mod client;
pub mod workspace; pub mod workspace;
pub mod cursor; pub mod cursor;

View file

@ -1,10 +1,23 @@
//! ### lua
//! Using [mlua] it's possible to map almost perfectly the entirety of `codemp` API.
//! Notable outliers are functions that receive `codemp` objects: these instead receive arguments
//! to build the object instead (such as [`crate::api::Controller::send`])
//!
//! Note that async operations are carried out on a [tokio] current_thread runtime, so it is
//! necessary to drive it. A separate driver thread can be spawned with `spawn_runtime_driver`
//! function.
//!
//! To make callbacks work, the main lua thread must periodically stop and poll for callbacks via
//! `poll_callback`, otherwise those will never run. This is necessary to allow safe concurrent
//! access to the global Lua state, so minimize runtime inside callbacks as much as possile.
use std::io::Write; use std::io::Write;
use std::sync::Mutex; use std::sync::Mutex;
use crate::api::Cursor; use crate::api::Cursor;
use crate::ext::IgnorableError; use crate::ext::IgnorableError;
use crate::prelude::*; use crate::prelude::*;
use crate::workspace::worker::DetachResult; use crate::workspace::DetachResult;
use mlua::prelude::*; use mlua::prelude::*;
use tokio::sync::mpsc; use tokio::sync::mpsc;
use tokio::sync::mpsc::error::TryRecvError; use tokio::sync::mpsc::error::TryRecvError;
@ -15,8 +28,8 @@ impl From::<crate::errors::ConnectionError> for LuaError {
} }
} }
impl From::<crate::errors::ProcedureError> for LuaError { impl From::<crate::errors::RemoteError> for LuaError {
fn from(value: crate::errors::ProcedureError) -> Self { fn from(value: crate::errors::RemoteError) -> Self {
LuaError::runtime(value.to_string()) LuaError::runtime(value.to_string())
} }
} }
@ -520,7 +533,7 @@ fn codemp_native(lua: &Lua) -> LuaResult<LuaTable> {
// utils // utils
exports.set("hash", lua.create_function(|_, (txt,):(String,)| exports.set("hash", lua.create_function(|_, (txt,):(String,)|
Ok(crate::hash(txt)) Ok(crate::ext::hash(txt))
)?)?; )?)?;
// runtime // runtime

View file

@ -1,11 +1,21 @@
//! ### FFI
//! Foreign-Function-Interface glue code, each gated behind feature flags
//!
//! For all except java, the resulting shared object is ready to use, but external packages are
//! available to simplify the dependancy and provide type hints in editor.
/// java bindings, built with [jni]
#[cfg(feature = "java")] #[cfg(feature = "java")]
pub mod java; pub mod java;
/// lua bindings, built with [mlua]
#[cfg(feature = "lua")] #[cfg(feature = "lua")]
pub mod lua; pub mod lua;
/// javascript bindings, built with [napi]
#[cfg(feature = "js")] #[cfg(feature = "js")]
pub mod js; pub mod js;
/// python bindings, built with [pyo3]
#[cfg(feature = "python")] #[cfg(feature = "python")]
pub mod python; pub mod python;

View file

@ -1,3 +1,7 @@
//! ### python
//! Using [pyo3] it's possible to map perfectly the entirety of `codemp` API.
//! Async operations run on a dedicated [tokio] runtime
pub mod client; pub mod client;
pub mod controllers; pub mod controllers;
pub mod workspace; pub mod workspace;

View file

@ -96,60 +96,6 @@ impl Workspace {
Ok(ws) Ok(ws)
} }
pub(crate) fn run_actor(
&self,
mut stream: Streaming<WorkspaceEvent>,
tx: mpsc::UnboundedSender<crate::api::Event>,
) {
// TODO for buffer and cursor controller we invoke the tokio::spawn outside, but here inside..?
let inner = self.0.clone();
let name = self.id();
tokio::spawn(async move {
loop {
match stream.message().await {
Err(e) => break tracing::error!("workspace '{}' stream closed: {}", name, e),
Ok(None) => break tracing::info!("leaving workspace {}", name),
Ok(Some(WorkspaceEvent { event: None })) => {
tracing::warn!("workspace {} received empty event", name)
}
Ok(Some(WorkspaceEvent { event: Some(ev) })) => {
let update = crate::api::Event::from(&ev);
match ev {
// user
WorkspaceEventInner::Join(UserJoin { user }) => {
inner
.users
.insert(user.id.uuid(), user.into());
}
WorkspaceEventInner::Leave(UserLeave { user }) => {
inner.users.remove(&user.id.uuid());
}
// buffer
WorkspaceEventInner::Create(FileCreate { path }) => {
inner.filetree.insert(path);
}
WorkspaceEventInner::Rename(FileRename { before, after }) => {
inner.filetree.remove(&before);
inner.filetree.insert(after);
}
WorkspaceEventInner::Delete(FileDelete { path }) => {
inner.filetree.remove(&path);
if let Some((_name, controller)) = inner.buffers.remove(&path) {
controller.stop();
}
}
}
if tx.send(update).is_err() {
tracing::warn!("no active controller to receive workspace event");
}
}
}
}
});
}
}
impl Workspace {
/// create a new buffer in current workspace /// create a new buffer in current workspace
pub async fn create(&self, path: &str) -> RemoteResult<()> { pub async fn create(&self, path: &str) -> RemoteResult<()> {
let mut workspace_client = self.0.services.ws(); let mut workspace_client = self.0.services.ws();
@ -207,11 +153,6 @@ impl Workspace {
/// this option will be carried in background: [buffer::worker::BufferWorker] will be stopped and dropped. there /// this option will be carried in background: [buffer::worker::BufferWorker] will be stopped and dropped. there
/// may still be some events enqueued in buffers to poll, but the [buffer::Controller] itself won't be /// may still be some events enqueued in buffers to poll, but the [buffer::Controller] itself won't be
/// accessible anymore from [Workspace]. /// accessible anymore from [Workspace].
///
/// ### returns
/// [DetachResult::NotAttached] if buffer wasn't attached in the first place
/// [DetachResult::Detaching] if detach was correctly requested
/// [DetachResult::AlreadyDetached] if worker is already stopped
pub fn detach(&self, path: &str) -> DetachResult { pub fn detach(&self, path: &str) -> DetachResult {
match self.0.buffers.remove(path) { match self.0.buffers.remove(path) {
None => DetachResult::NotAttached, None => DetachResult::NotAttached,
@ -225,7 +166,7 @@ impl Workspace {
} }
} }
/// await next workspace [crate::api::Event] and return it /// Await next workspace [Event] and return it
// TODO this method is weird and ugly, can we make it more standard? // TODO this method is weird and ugly, can we make it more standard?
pub async fn event(&self) -> ControllerResult<Event> { pub async fn event(&self) -> ControllerResult<Event> {
self.0 self.0
@ -237,7 +178,7 @@ impl Workspace {
.ok_or(crate::errors::ControllerError::Unfulfilled) .ok_or(crate::errors::ControllerError::Unfulfilled)
} }
/// fetch a list of all buffers in a workspace /// Re-fetch list of all buffers in a workspace
pub async fn fetch_buffers(&self) -> RemoteResult<()> { pub async fn fetch_buffers(&self) -> RemoteResult<()> {
let mut workspace_client = self.0.services.ws(); let mut workspace_client = self.0.services.ws();
let buffers = workspace_client let buffers = workspace_client
@ -254,7 +195,7 @@ impl Workspace {
Ok(()) Ok(())
} }
/// fetch a list of all users in a workspace /// Re-fetch list of all users in a workspace
pub async fn fetch_users(&self) -> RemoteResult<()> { pub async fn fetch_users(&self) -> RemoteResult<()> {
let mut workspace_client = self.0.services.ws(); let mut workspace_client = self.0.services.ws();
let users = BTreeSet::from_iter( let users = BTreeSet::from_iter(
@ -275,9 +216,7 @@ impl Workspace {
Ok(()) Ok(())
} }
/// get a list of the users attached to a specific buffer /// Get a list of the [User]s attached to a specific buffer
///
/// TODO: discuss implementation details
pub async fn list_buffer_users(&self, path: &str) -> RemoteResult<Vec<User>> { pub async fn list_buffer_users(&self, path: &str) -> RemoteResult<Vec<User>> {
let mut workspace_client = self.0.services.ws(); let mut workspace_client = self.0.services.ws();
let buffer_users = workspace_client let buffer_users = workspace_client
@ -294,7 +233,7 @@ impl Workspace {
Ok(buffer_users) Ok(buffer_users)
} }
/// delete a buffer /// Delete a buffer
pub async fn delete(&self, path: &str) -> RemoteResult<()> { pub async fn delete(&self, path: &str) -> RemoteResult<()> {
let mut workspace_client = self.0.services.ws(); let mut workspace_client = self.0.services.ws();
workspace_client workspace_client
@ -312,25 +251,25 @@ impl Workspace {
Ok(()) Ok(())
} }
/// get the id of the workspace /// Get the workspace unique id
// #[cfg_attr(feature = "js", napi)] // https://github.com/napi-rs/napi-rs/issues/1120 // #[cfg_attr(feature = "js", napi)] // https://github.com/napi-rs/napi-rs/issues/1120
pub fn id(&self) -> String { pub fn id(&self) -> String {
self.0.name.clone() self.0.name.clone()
} }
/// return a reference to current cursor controller, if currently in a workspace /// Return a handle to workspace cursor controller
// #[cfg_attr(feature = "js", napi)] // https://github.com/napi-rs/napi-rs/issues/1120 // #[cfg_attr(feature = "js", napi)] // https://github.com/napi-rs/napi-rs/issues/1120
pub fn cursor(&self) -> cursor::Controller { pub fn cursor(&self) -> cursor::Controller {
self.0.cursor.clone() self.0.cursor.clone()
} }
/// get a new reference to a buffer controller, if any is active to given path /// Get a [buffer::Controller] by path, if any is active on given path
// #[cfg_attr(feature = "js", napi)] // https://github.com/napi-rs/napi-rs/issues/1120 // #[cfg_attr(feature = "js", napi)] // https://github.com/napi-rs/napi-rs/issues/1120
pub fn buffer_by_name(&self, path: &str) -> Option<buffer::Controller> { pub fn buffer_by_name(&self, path: &str) -> Option<buffer::Controller> {
self.0.buffers.get(path).map(|x| x.clone()) self.0.buffers.get(path).map(|x| x.clone())
} }
/// get a list of all the currently attached to buffers /// Get a list of all the currently attached buffers
// #[cfg_attr(feature = "js", napi)] // https://github.com/napi-rs/napi-rs/issues/1120 // #[cfg_attr(feature = "js", napi)] // https://github.com/napi-rs/napi-rs/issues/1120
pub fn buffer_list(&self) -> Vec<String> { pub fn buffer_list(&self) -> Vec<String> {
self.0 self.0
@ -348,6 +287,58 @@ impl Workspace {
.map(|f| f.clone()) .map(|f| f.clone())
.collect() .collect()
} }
pub(crate) fn run_actor(
&self,
mut stream: Streaming<WorkspaceEvent>,
tx: mpsc::UnboundedSender<crate::api::Event>,
) {
// TODO for buffer and cursor controller we invoke the tokio::spawn outside, but here inside..?
let inner = self.0.clone();
let name = self.id();
tokio::spawn(async move {
loop {
match stream.message().await {
Err(e) => break tracing::error!("workspace '{}' stream closed: {}", name, e),
Ok(None) => break tracing::info!("leaving workspace {}", name),
Ok(Some(WorkspaceEvent { event: None })) => {
tracing::warn!("workspace {} received empty event", name)
}
Ok(Some(WorkspaceEvent { event: Some(ev) })) => {
let update = crate::api::Event::from(&ev);
match ev {
// user
WorkspaceEventInner::Join(UserJoin { user }) => {
inner
.users
.insert(user.id.uuid(), user.into());
}
WorkspaceEventInner::Leave(UserLeave { user }) => {
inner.users.remove(&user.id.uuid());
}
// buffer
WorkspaceEventInner::Create(FileCreate { path }) => {
inner.filetree.insert(path);
}
WorkspaceEventInner::Rename(FileRename { before, after }) => {
inner.filetree.remove(&before);
inner.filetree.insert(after);
}
WorkspaceEventInner::Delete(FileDelete { path }) => {
inner.filetree.remove(&path);
if let Some((_name, controller)) = inner.buffers.remove(&path) {
controller.stop();
}
}
}
if tx.send(update).is_err() {
tracing::warn!("no active controller to receive workspace event");
}
}
}
}
});
}
} }
impl Drop for WorkspaceInner { impl Drop for WorkspaceInner {
@ -356,7 +347,7 @@ impl Drop for WorkspaceInner {
if !entry.value().stop() { if !entry.value().stop() {
tracing::warn!( tracing::warn!(
"could not stop buffer worker {} for workspace {}", "could not stop buffer worker {} for workspace {}",
entry.value().name(), entry.value().path(),
self.name self.name
); );
} }

View file

@ -1,4 +0,0 @@
pub mod service;
pub mod worker;
pub use worker::Workspace;

View file

@ -1,73 +0,0 @@
use codemp_proto::{
common::Token, buffer::buffer_client::BufferClient, cursor::cursor_client::CursorClient,
workspace::workspace_client::WorkspaceClient,
};
use tonic::{
service::{interceptor::InterceptedService, Interceptor},
transport::{Channel, Endpoint},
};
use crate::errors::ConnectionResult;
type AuthedService = InterceptedService<Channel, WorkspaceInterceptor>;
#[derive(Debug)]
pub struct Services {
workspace: WorkspaceClient<AuthedService>,
buffer: BufferClient<AuthedService>,
cursor: CursorClient<AuthedService>,
}
impl Services {
pub async fn try_new(
dest: &str,
session: tokio::sync::watch::Receiver<codemp_proto::common::Token>,
workspace: tokio::sync::watch::Receiver<codemp_proto::common::Token>,
) -> ConnectionResult<Self> {
let channel = Endpoint::from_shared(dest.to_string())?.connect().await?;
let inter = WorkspaceInterceptor { session, workspace };
Ok(Self {
cursor: CursorClient::with_interceptor(channel.clone(), inter.clone()),
workspace: WorkspaceClient::with_interceptor(channel.clone(), inter.clone()),
// TODO technically we could keep buffers on separate servers, and thus manage buffer
// connections separately, but for now it's more convenient to bundle them with workspace
buffer: BufferClient::with_interceptor(channel.clone(), inter.clone()),
})
}
// TODO just make fields pub(crate) ?? idk
pub fn ws(&self) -> WorkspaceClient<AuthedService> {
self.workspace.clone()
}
pub fn buf(&self) -> BufferClient<AuthedService> {
self.buffer.clone()
}
pub fn cur(&self) -> CursorClient<AuthedService> {
self.cursor.clone()
}
}
#[derive(Clone)]
pub struct WorkspaceInterceptor {
session: tokio::sync::watch::Receiver<Token>,
workspace: tokio::sync::watch::Receiver<Token>,
}
impl Interceptor for WorkspaceInterceptor {
fn call(
&mut self,
mut request: tonic::Request<()>,
) -> Result<tonic::Request<()>, tonic::Status> {
if let Ok(token) = self.session.borrow().token.parse() {
request.metadata_mut().insert("session", token);
}
if let Ok(token) = self.workspace.borrow().token.parse() {
request.metadata_mut().insert("workspace", token);
}
Ok(request)
}
}

View file

@ -1,371 +0,0 @@
use crate::{
api::{controller::ControllerWorker, Controller, Event, User},
buffer::{self, worker::BufferWorker},
cursor::{self, worker::CursorWorker},
errors::{ConnectionResult, ControllerResult, RemoteResult},
ext::InternallyMutable,
workspace::service::Services
};
use codemp_proto::{
common::{Empty, Token},
files::BufferNode,
workspace::{
workspace_event::{
Event as WorkspaceEventInner, FileCreate, FileDelete, FileRename, UserJoin, UserLeave,
},
WorkspaceEvent,
},
};
use dashmap::{DashMap, DashSet};
use std::{collections::BTreeSet, sync::Arc};
use tokio::sync::mpsc;
use tonic::Streaming;
use uuid::Uuid;
#[cfg(feature = "js")]
use napi_derive::napi;
#[derive(Debug, Clone)]
#[cfg_attr(feature = "python", pyo3::pyclass)]
#[cfg_attr(feature = "js", napi)]
pub struct Workspace(Arc<WorkspaceInner>);
#[derive(Debug)]
struct WorkspaceInner {
name: String,
user: User, // TODO back-reference to global user id... needed for buffer controllers
cursor: cursor::Controller,
buffers: DashMap<String, buffer::Controller>,
filetree: DashSet<String>,
users: DashMap<Uuid, User>,
services: Services,
// TODO can we drop the mutex?
events: tokio::sync::Mutex<mpsc::UnboundedReceiver<crate::api::Event>>,
}
impl Workspace {
pub(crate) async fn try_new(
name: String,
user: User,
dest: &str,
token: Token,
claims: tokio::sync::watch::Receiver<codemp_proto::common::Token>, // TODO ughh receiving this
) -> ConnectionResult<Self> {
let workspace_claim = InternallyMutable::new(token);
let services = Services::try_new(dest, claims, workspace_claim.channel()).await?;
let ws_stream = services.ws().attach(Empty {}).await?.into_inner();
let (tx, rx) = mpsc::channel(128);
let (ev_tx, ev_rx) = mpsc::unbounded_channel();
let cur_stream = services
.cur()
.attach(tokio_stream::wrappers::ReceiverStream::new(rx))
.await?
.into_inner();
let worker = CursorWorker::default();
let controller = worker.controller();
tokio::spawn(async move {
tracing::debug!("controller worker started");
worker.work(tx, cur_stream).await;
tracing::debug!("controller worker stopped");
});
let ws = Self(Arc::new(WorkspaceInner {
name,
user,
cursor: controller,
buffers: DashMap::default(),
filetree: DashSet::default(),
users: DashMap::default(),
events: tokio::sync::Mutex::new(ev_rx),
services,
}));
ws.fetch_users().await?;
ws.fetch_buffers().await?;
ws.run_actor(ws_stream, ev_tx);
Ok(ws)
}
pub(crate) fn run_actor(
&self,
mut stream: Streaming<WorkspaceEvent>,
tx: mpsc::UnboundedSender<crate::api::Event>,
) {
// TODO for buffer and cursor controller we invoke the tokio::spawn outside, but here inside..?
let inner = self.0.clone();
let name = self.id();
tokio::spawn(async move {
loop {
match stream.message().await {
Err(e) => break tracing::error!("workspace '{}' stream closed: {}", name, e),
Ok(None) => break tracing::info!("leaving workspace {}", name),
Ok(Some(WorkspaceEvent { event: None })) => {
tracing::warn!("workspace {} received empty event", name)
}
Ok(Some(WorkspaceEvent { event: Some(ev) })) => {
let update = crate::api::Event::from(&ev);
match ev {
// user
WorkspaceEventInner::Join(UserJoin { user }) => {
inner
.users
.insert(user.id.uuid(), user.into());
}
WorkspaceEventInner::Leave(UserLeave { user }) => {
inner.users.remove(&user.id.uuid());
}
// buffer
WorkspaceEventInner::Create(FileCreate { path }) => {
inner.filetree.insert(path);
}
WorkspaceEventInner::Rename(FileRename { before, after }) => {
inner.filetree.remove(&before);
inner.filetree.insert(after);
}
WorkspaceEventInner::Delete(FileDelete { path }) => {
inner.filetree.remove(&path);
if let Some((_name, controller)) = inner.buffers.remove(&path) {
controller.stop();
}
}
}
if tx.send(update).is_err() {
tracing::warn!("no active controller to receive workspace event");
}
}
}
}
});
}
}
impl Workspace {
/// create a new buffer in current workspace
pub async fn create(&self, path: &str) -> RemoteResult<()> {
let mut workspace_client = self.0.services.ws();
workspace_client
.create_buffer(tonic::Request::new(BufferNode {
path: path.to_string(),
}))
.await?;
// add to filetree
self.0.filetree.insert(path.to_string());
// fetch buffers
self.fetch_buffers().await?;
Ok(())
}
/// attach to a buffer, starting a buffer controller and returning a new reference to it
///
/// to interact with such buffer use [crate::api::Controller::send] or
/// [crate::api::Controller::recv] to exchange [crate::api::TextChange]
pub async fn attach(&self, path: &str) -> ConnectionResult<buffer::Controller> {
let mut worskspace_client = self.0.services.ws();
let request = tonic::Request::new(BufferNode {
path: path.to_string(),
});
let credentials = worskspace_client.access_buffer(request).await?.into_inner();
let (tx, rx) = mpsc::channel(256);
let mut req = tonic::Request::new(tokio_stream::wrappers::ReceiverStream::new(rx));
req.metadata_mut()
.insert(
"buffer",
tonic::metadata::MetadataValue::try_from(credentials.token)
.map_err(|e| tonic::Status::internal(format!("failed representing token to string: {e}")))?,
);
let stream = self.0.services.buf().attach(req).await?.into_inner();
let worker = BufferWorker::new(self.0.user.id, path);
let controller = worker.controller();
tokio::spawn(async move {
tracing::debug!("controller worker started");
worker.work(tx, stream).await;
tracing::debug!("controller worker stopped");
});
self.0.buffers.insert(path.to_string(), controller.clone());
Ok(controller)
}
/// detach from an active buffer
///
/// this option will be carried in background: [buffer::worker::BufferWorker] will be stopped and dropped. there
/// may still be some events enqueued in buffers to poll, but the [buffer::Controller] itself won't be
/// accessible anymore from [Workspace].
///
/// ### returns
/// [DetachResult::NotAttached] if buffer wasn't attached in the first place
/// [DetachResult::Detaching] if detach was correctly requested
/// [DetachResult::AlreadyDetached] if worker is already stopped
pub fn detach(&self, path: &str) -> DetachResult {
match self.0.buffers.remove(path) {
None => DetachResult::NotAttached,
Some((_name, controller)) => {
if controller.stop() {
DetachResult::Detaching
} else {
DetachResult::AlreadyDetached
}
}
}
}
/// await next workspace [crate::api::Event] and return it
// TODO this method is weird and ugly, can we make it more standard?
pub async fn event(&self) -> ControllerResult<Event> {
self.0
.events
.lock()
.await
.recv()
.await
.ok_or(crate::errors::ControllerError::Unfulfilled)
}
/// fetch a list of all buffers in a workspace
pub async fn fetch_buffers(&self) -> RemoteResult<()> {
let mut workspace_client = self.0.services.ws();
let buffers = workspace_client
.list_buffers(tonic::Request::new(Empty {}))
.await?
.into_inner()
.buffers;
self.0.filetree.clear();
for b in buffers {
self.0.filetree.insert(b.path);
}
Ok(())
}
/// fetch a list of all users in a workspace
pub async fn fetch_users(&self) -> RemoteResult<()> {
let mut workspace_client = self.0.services.ws();
let users = BTreeSet::from_iter(
workspace_client
.list_users(tonic::Request::new(Empty {}))
.await?
.into_inner()
.users
.into_iter()
.map(User::from),
);
self.0.users.clear();
for u in users {
self.0.users.insert(u.id, u);
}
Ok(())
}
/// get a list of the users attached to a specific buffer
///
/// TODO: discuss implementation details
pub async fn list_buffer_users(&self, path: &str) -> RemoteResult<Vec<User>> {
let mut workspace_client = self.0.services.ws();
let buffer_users = workspace_client
.list_buffer_users(tonic::Request::new(BufferNode {
path: path.to_string(),
}))
.await?
.into_inner()
.users
.into_iter()
.map(|id| id.into())
.collect();
Ok(buffer_users)
}
/// delete a buffer
pub async fn delete(&self, path: &str) -> RemoteResult<()> {
let mut workspace_client = self.0.services.ws();
workspace_client
.delete_buffer(tonic::Request::new(BufferNode {
path: path.to_string(),
}))
.await?;
if let Some((_name, controller)) = self.0.buffers.remove(path) {
controller.stop();
}
self.0.filetree.remove(path);
Ok(())
}
/// get the id of the workspace
// #[cfg_attr(feature = "js", napi)] // https://github.com/napi-rs/napi-rs/issues/1120
pub fn id(&self) -> String {
self.0.name.clone()
}
/// return a reference to current cursor controller, if currently in a workspace
// #[cfg_attr(feature = "js", napi)] // https://github.com/napi-rs/napi-rs/issues/1120
pub fn cursor(&self) -> cursor::Controller {
self.0.cursor.clone()
}
/// get a new reference to a buffer controller, if any is active to given path
// #[cfg_attr(feature = "js", napi)] // https://github.com/napi-rs/napi-rs/issues/1120
pub fn buffer_by_name(&self, path: &str) -> Option<buffer::Controller> {
self.0.buffers.get(path).map(|x| x.clone())
}
/// get a list of all the currently attached to buffers
// #[cfg_attr(feature = "js", napi)] // https://github.com/napi-rs/napi-rs/issues/1120
pub fn buffer_list(&self) -> Vec<String> {
self.0
.buffers
.iter()
.map(|elem| elem.key().clone())
.collect()
}
/// get the currently cached "filetree"
// #[cfg_attr(feature = "js", napi)] // https://github.com/napi-rs/napi-rs/issues/1120
pub fn filetree(&self, filter: Option<&str>) -> Vec<String> {
self.0.filetree.iter()
.filter(|f| filter.map_or(true, |flt| f.starts_with(flt)))
.map(|f| f.clone())
.collect()
}
}
impl Drop for WorkspaceInner {
fn drop(&mut self) {
for entry in self.buffers.iter() {
if !entry.value().stop() {
tracing::warn!(
"could not stop buffer worker {} for workspace {}",
entry.value().name(),
self.name
);
}
}
if !self.cursor.stop() {
tracing::warn!("could not stop cursor worker for workspace {}", self.name);
}
}
}
#[cfg_attr(feature = "python", pyo3::pyclass(eq, eq_int))]
#[cfg_attr(feature = "python", derive(PartialEq))]
pub enum DetachResult {
NotAttached,
Detaching,
AlreadyDetached,
}