chore: rename some struct

This commit is contained in:
appflowy 2022-02-18 23:04:55 +08:00
parent 6593855f8f
commit dd832b7482
16 changed files with 115 additions and 125 deletions

View file

@ -1,9 +0,0 @@
mod editor;
mod queue;
mod web_socket;
pub use editor::*;
pub(crate) use queue::*;
pub(crate) use web_socket::*;
pub const DOCUMENT_SYNC_INTERVAL_IN_MILLIS: u64 = 1000;

View file

@ -1,7 +1,8 @@
use crate::core::DocumentRevisionCompact; use crate::queue::DocumentRevisionCompact;
use crate::web_socket::{make_document_ws_manager, EditorCommandSender};
use crate::{ use crate::{
core::{make_document_ws_manager, EditorCommand, EditorCommandQueue, EditorCommandSender},
errors::FlowyError, errors::FlowyError,
queue::{EditorCommand, EditorCommandQueue},
DocumentUser, DocumentWSReceiver, DocumentUser, DocumentWSReceiver,
}; };
use bytes::Bytes; use bytes::Bytes;
@ -34,11 +35,11 @@ impl ClientDocumentEditor {
doc_id: &str, doc_id: &str,
user: Arc<dyn DocumentUser>, user: Arc<dyn DocumentUser>,
mut rev_manager: RevisionManager, mut rev_manager: RevisionManager,
web_socket: Arc<dyn RevisionWebSocket>, rev_web_socket: Arc<dyn RevisionWebSocket>,
server: Arc<dyn RevisionCloudService>, cloud_service: Arc<dyn RevisionCloudService>,
) -> FlowyResult<Arc<Self>> { ) -> FlowyResult<Arc<Self>> {
let document_info = rev_manager let document_info = rev_manager
.load::<DocumentInfoBuilder, DocumentRevisionCompact>(server) .load::<DocumentInfoBuilder, DocumentRevisionCompact>(cloud_service)
.await?; .await?;
let delta = document_info.delta()?; let delta = document_info.delta()?;
let rev_manager = Arc::new(rev_manager); let rev_manager = Arc::new(rev_manager);
@ -51,7 +52,7 @@ impl ClientDocumentEditor {
user_id.clone(), user_id.clone(),
edit_cmd_tx.clone(), edit_cmd_tx.clone(),
rev_manager.clone(), rev_manager.clone(),
web_socket, rev_web_socket,
) )
.await; .await;
let editor = Arc::new(Self { let editor = Arc::new(Self {

View file

@ -1,10 +1,15 @@
pub mod controller; pub mod editor;
pub mod core; pub mod manager;
pub use controller::*; mod queue;
mod web_socket;
pub use manager::*;
pub mod errors { pub mod errors {
pub use flowy_error::{internal_error, ErrorCode, FlowyError}; pub use flowy_error::{internal_error, ErrorCode, FlowyError};
} }
pub const DOCUMENT_SYNC_INTERVAL_IN_MILLIS: u64 = 1000;
use crate::errors::FlowyError; use crate::errors::FlowyError;
use flowy_collaboration::entities::document_info::{CreateDocParams, DocumentId, DocumentInfo, ResetDocumentParams}; use flowy_collaboration::entities::document_info::{CreateDocParams, DocumentId, DocumentInfo, ResetDocumentParams};
use lib_infra::future::FutureResult; use lib_infra::future::FutureResult;

View file

@ -1,4 +1,4 @@
use crate::{core::ClientDocumentEditor, errors::FlowyError, DocumentCloudService}; use crate::{editor::ClientDocumentEditor, errors::FlowyError, DocumentCloudService};
use async_trait::async_trait; use async_trait::async_trait;
use bytes::Bytes; use bytes::Bytes;
use dashmap::DashMap; use dashmap::DashMap;
@ -29,31 +29,31 @@ pub(crate) trait DocumentWSReceiver: Send + Sync {
type WebSocketDataReceivers = Arc<DashMap<String, Arc<dyn DocumentWSReceiver>>>; type WebSocketDataReceivers = Arc<DashMap<String, Arc<dyn DocumentWSReceiver>>>;
pub struct FlowyDocumentManager { pub struct FlowyDocumentManager {
cloud_service: Arc<dyn DocumentCloudService>, cloud_service: Arc<dyn DocumentCloudService>,
ws_receivers: WebSocketDataReceivers, ws_data_receivers: WebSocketDataReceivers,
web_socket: Arc<dyn RevisionWebSocket>, rev_web_socket: Arc<dyn RevisionWebSocket>,
open_cache: Arc<OpenDocCache>, document_handlers: Arc<DocumentEditorHandlers>,
user: Arc<dyn DocumentUser>, document_user: Arc<dyn DocumentUser>,
} }
impl FlowyDocumentManager { impl FlowyDocumentManager {
pub fn new( pub fn new(
cloud_service: Arc<dyn DocumentCloudService>, cloud_service: Arc<dyn DocumentCloudService>,
user: Arc<dyn DocumentUser>, document_user: Arc<dyn DocumentUser>,
web_socket: Arc<dyn RevisionWebSocket>, rev_web_socket: Arc<dyn RevisionWebSocket>,
) -> Self { ) -> Self {
let ws_receivers = Arc::new(DashMap::new()); let ws_data_receivers = Arc::new(DashMap::new());
let open_cache = Arc::new(OpenDocCache::new()); let document_handlers = Arc::new(DocumentEditorHandlers::new());
Self { Self {
cloud_service, cloud_service,
ws_receivers, ws_data_receivers,
web_socket, rev_web_socket,
open_cache, document_handlers,
user, document_user,
} }
} }
pub fn init(&self) -> FlowyResult<()> { pub fn init(&self) -> FlowyResult<()> {
listen_ws_state_changed(self.web_socket.clone(), self.ws_receivers.clone()); listen_ws_state_changed(self.rev_web_socket.clone(), self.ws_data_receivers.clone());
Ok(()) Ok(())
} }
@ -69,8 +69,8 @@ impl FlowyDocumentManager {
pub fn close_document<T: AsRef<str>>(&self, doc_id: T) -> Result<(), FlowyError> { pub fn close_document<T: AsRef<str>>(&self, doc_id: T) -> Result<(), FlowyError> {
let doc_id = doc_id.as_ref(); let doc_id = doc_id.as_ref();
tracing::Span::current().record("doc_id", &doc_id); tracing::Span::current().record("doc_id", &doc_id);
self.open_cache.remove(doc_id); self.document_handlers.remove(doc_id);
self.remove_ws_receiver(doc_id); self.ws_data_receivers.remove(doc_id);
Ok(()) Ok(())
} }
@ -78,8 +78,8 @@ impl FlowyDocumentManager {
pub fn delete<T: AsRef<str>>(&self, doc_id: T) -> Result<(), FlowyError> { pub fn delete<T: AsRef<str>>(&self, doc_id: T) -> Result<(), FlowyError> {
let doc_id = doc_id.as_ref(); let doc_id = doc_id.as_ref();
tracing::Span::current().record("doc_id", &doc_id); tracing::Span::current().record("doc_id", &doc_id);
self.open_cache.remove(doc_id); self.document_handlers.remove(doc_id);
self.remove_ws_receiver(doc_id); self.ws_data_receivers.remove(doc_id);
Ok(()) Ok(())
} }
@ -94,18 +94,18 @@ impl FlowyDocumentManager {
}) })
} }
pub async fn save_document<T: AsRef<str>>(&self, doc_id: T, revisions: RepeatedRevision) -> FlowyResult<()> { pub async fn receive_revisions<T: AsRef<str>>(&self, doc_id: T, revisions: RepeatedRevision) -> FlowyResult<()> {
let doc_id = doc_id.as_ref().to_owned(); let doc_id = doc_id.as_ref().to_owned();
let db_pool = self.user.db_pool()?; let db_pool = self.document_user.db_pool()?;
let rev_manager = self.make_rev_manager(&doc_id, db_pool)?; let rev_manager = self.make_rev_manager(&doc_id, db_pool)?;
let _ = rev_manager.reset_object(revisions).await?; let _ = rev_manager.reset_object(revisions).await?;
Ok(()) Ok(())
} }
pub async fn did_receive_ws_data(&self, data: Bytes) { pub async fn receive_ws_data(&self, data: Bytes) {
let result: Result<ServerRevisionWSData, protobuf::ProtobufError> = data.try_into(); let result: Result<ServerRevisionWSData, protobuf::ProtobufError> = data.try_into();
match result { match result {
Ok(data) => match self.ws_receivers.get(&data.object_id) { Ok(data) => match self.ws_data_receivers.get(&data.object_id) {
None => tracing::error!("Can't find any source handler for {:?}-{:?}", data.object_id, data.ty), None => tracing::error!("Can't find any source handler for {:?}-{:?}", data.object_id, data.ty),
Some(handler) => match handler.receive_ws_data(data).await { Some(handler) => match handler.receive_ws_data(data).await {
Ok(_) => {} Ok(_) => {}
@ -117,19 +117,13 @@ impl FlowyDocumentManager {
} }
} }
} }
pub async fn ws_connect_state_changed(&self, state: &WSConnectState) {
for receiver in self.ws_receivers.iter() {
receiver.value().connect_state_changed(state.clone());
}
}
} }
impl FlowyDocumentManager { impl FlowyDocumentManager {
async fn get_editor(&self, doc_id: &str) -> FlowyResult<Arc<ClientDocumentEditor>> { async fn get_editor(&self, doc_id: &str) -> FlowyResult<Arc<ClientDocumentEditor>> {
match self.open_cache.get(doc_id) { match self.document_handlers.get(doc_id) {
None => { None => {
let db_pool = self.user.db_pool()?; let db_pool = self.document_user.db_pool()?;
self.make_editor(doc_id, db_pool).await self.make_editor(doc_id, db_pool).await
} }
Some(editor) => Ok(editor), Some(editor) => Ok(editor),
@ -141,35 +135,26 @@ impl FlowyDocumentManager {
doc_id: &str, doc_id: &str,
pool: Arc<ConnectionPool>, pool: Arc<ConnectionPool>,
) -> Result<Arc<ClientDocumentEditor>, FlowyError> { ) -> Result<Arc<ClientDocumentEditor>, FlowyError> {
let user = self.user.clone(); let user = self.document_user.clone();
let token = self.user.token()?; let token = self.document_user.token()?;
let rev_manager = self.make_rev_manager(doc_id, pool.clone())?; let rev_manager = self.make_rev_manager(doc_id, pool.clone())?;
let server = Arc::new(DocumentRevisionCloudServiceImpl { let cloud_service = Arc::new(DocumentRevisionCloudServiceImpl {
token, token,
server: self.cloud_service.clone(), server: self.cloud_service.clone(),
}); });
let doc_editor = ClientDocumentEditor::new(doc_id, user, rev_manager, self.web_socket.clone(), server).await?; let doc_editor =
self.add_ws_receiver(doc_id, doc_editor.ws_handler()); ClientDocumentEditor::new(doc_id, user, rev_manager, self.rev_web_socket.clone(), cloud_service).await?;
self.open_cache.insert(doc_id, &doc_editor); self.ws_data_receivers
.insert(doc_id.to_string(), doc_editor.ws_handler());
self.document_handlers.insert(doc_id, &doc_editor);
Ok(doc_editor) Ok(doc_editor)
} }
fn make_rev_manager(&self, doc_id: &str, pool: Arc<ConnectionPool>) -> Result<RevisionManager, FlowyError> { fn make_rev_manager(&self, doc_id: &str, pool: Arc<ConnectionPool>) -> Result<RevisionManager, FlowyError> {
let user_id = self.user.user_id()?; let user_id = self.document_user.user_id()?;
let cache = Arc::new(RevisionCache::new(&user_id, doc_id, pool)); let cache = Arc::new(RevisionCache::new(&user_id, doc_id, pool));
Ok(RevisionManager::new(&user_id, doc_id, cache)) Ok(RevisionManager::new(&user_id, doc_id, cache))
} }
fn add_ws_receiver(&self, object_id: &str, receiver: Arc<dyn DocumentWSReceiver>) {
if self.ws_receivers.contains_key(object_id) {
log::error!("Duplicate handler registered for {:?}", object_id);
}
self.ws_receivers.insert(object_id.to_string(), receiver);
}
fn remove_ws_receiver(&self, id: &str) {
self.ws_receivers.remove(id);
}
} }
struct DocumentRevisionCloudServiceImpl { struct DocumentRevisionCloudServiceImpl {
@ -202,11 +187,11 @@ impl RevisionCloudService for DocumentRevisionCloudServiceImpl {
} }
} }
pub struct OpenDocCache { pub struct DocumentEditorHandlers {
inner: DashMap<String, Arc<ClientDocumentEditor>>, inner: DashMap<String, Arc<ClientDocumentEditor>>,
} }
impl OpenDocCache { impl DocumentEditorHandlers {
fn new() -> Self { fn new() -> Self {
Self { inner: DashMap::new() } Self { inner: DashMap::new() }
} }

View file

@ -1,4 +1,5 @@
use crate::{core::web_socket::EditorCommandReceiver, DocumentUser}; use crate::web_socket::EditorCommandReceiver;
use crate::DocumentUser;
use async_stream::stream; use async_stream::stream;
use flowy_collaboration::util::make_delta_from_revisions; use flowy_collaboration::util::make_delta_from_revisions;
use flowy_collaboration::{ use flowy_collaboration::{

View file

@ -1,7 +1,4 @@
use crate::{ use crate::{queue::EditorCommand, DocumentWSReceiver, DOCUMENT_SYNC_INTERVAL_IN_MILLIS};
core::{EditorCommand, DOCUMENT_SYNC_INTERVAL_IN_MILLIS},
DocumentWSReceiver,
};
use async_trait::async_trait; use async_trait::async_trait;
use bytes::Bytes; use bytes::Bytes;
use flowy_collaboration::{ use flowy_collaboration::{
@ -31,7 +28,7 @@ pub(crate) async fn make_document_ws_manager(
user_id: String, user_id: String,
edit_cmd_tx: EditorCommandSender, edit_cmd_tx: EditorCommandSender,
rev_manager: Arc<RevisionManager>, rev_manager: Arc<RevisionManager>,
web_socket: Arc<dyn RevisionWebSocket>, rev_web_socket: Arc<dyn RevisionWebSocket>,
) -> Arc<RevisionWebSocketManager> { ) -> Arc<RevisionWebSocketManager> {
let composite_sink_provider = Arc::new(CompositeWSSinkDataProvider::new(&doc_id, rev_manager.clone())); let composite_sink_provider = Arc::new(CompositeWSSinkDataProvider::new(&doc_id, rev_manager.clone()));
let resolve_target = Arc::new(DocumentRevisionResolveTarget { edit_cmd_tx }); let resolve_target = Arc::new(DocumentRevisionResolveTarget { edit_cmd_tx });
@ -50,7 +47,7 @@ pub(crate) async fn make_document_ws_manager(
let ws_manager = Arc::new(RevisionWebSocketManager::new( let ws_manager = Arc::new(RevisionWebSocketManager::new(
"Document", "Document",
&doc_id, &doc_id,
web_socket, rev_web_socket,
sink_provider, sink_provider,
ws_stream_consumer, ws_stream_consumer,
ping_duration, ping_duration,

View file

@ -1,5 +1,6 @@
use flowy_collaboration::entities::revision::RevisionState; use flowy_collaboration::entities::revision::RevisionState;
use flowy_document::core::{ClientDocumentEditor, DOCUMENT_SYNC_INTERVAL_IN_MILLIS}; use flowy_document::editor::ClientDocumentEditor;
use flowy_document::DOCUMENT_SYNC_INTERVAL_IN_MILLIS;
use flowy_test::{helper::ViewTest, FlowySDKTest}; use flowy_test::{helper::ViewTest, FlowySDKTest};
use lib_ot::{core::Interval, rich_text::RichTextDelta}; use lib_ot::{core::Interval, rich_text::RichTextDelta};
use std::sync::Arc; use std::sync::Arc;
@ -67,8 +68,8 @@ impl EditorTest {
return; return;
} }
let next_revision = next_revision.unwrap(); let next_revision = next_revision.unwrap();
let mut receiver = rev_manager.revision_ack_receiver(); let mut notify = rev_manager.ack_notify();
let _ = receiver.recv().await; let _ = notify.recv().await;
assert_eq!(next_revision.rev_id, rev_id.unwrap()); assert_eq!(next_revision.rev_id, rev_id.unwrap());
} }
EditorScript::AssertJson(expected) => { EditorScript::AssertJson(expected) => {

View file

@ -73,7 +73,7 @@ impl ViewController {
Revision::initial_revision(&user_id, &params.view_id, delta_data).into(); Revision::initial_revision(&user_id, &params.view_id, delta_data).into();
let _ = self let _ = self
.document_manager .document_manager
.save_document(&params.view_id, repeated_revision) .receive_revisions(&params.view_id, repeated_revision)
.await?; .await?;
let view = self.create_view_on_server(params).await?; let view = self.create_view_on_server(params).await?;
let _ = self.create_view_on_local(view.clone()).await?; let _ = self.create_view_on_local(view.clone()).await?;
@ -94,7 +94,10 @@ impl ViewController {
let delta_data = Bytes::from(view_data); let delta_data = Bytes::from(view_data);
let user_id = self.user.user_id()?; let user_id = self.user.user_id()?;
let repeated_revision: RepeatedRevision = Revision::initial_revision(&user_id, view_id, delta_data).into(); let repeated_revision: RepeatedRevision = Revision::initial_revision(&user_id, view_id, delta_data).into();
let _ = self.document_manager.save_document(view_id, repeated_revision).await?; let _ = self
.document_manager
.receive_revisions(view_id, repeated_revision)
.await?;
Ok(()) Ok(())
} }

View file

@ -203,8 +203,8 @@ impl FolderTest {
} }
let next_revision = next_revision let next_revision = next_revision
.unwrap_or_else(|| panic!("Expected Next revision is {}, but receive None", rev_id.unwrap())); .unwrap_or_else(|| panic!("Expected Next revision is {}, but receive None", rev_id.unwrap()));
let mut receiver = rev_manager.revision_ack_receiver(); let mut notify = rev_manager.ack_notify();
let _ = receiver.recv().await; let _ = notify.recv().await;
assert_eq!(next_revision.rev_id, rev_id.unwrap()); assert_eq!(next_revision.rev_id, rev_id.unwrap());
} }
} }

View file

@ -98,7 +98,7 @@ impl WSMessageReceiver for DocumentWSMessageReceiverImpl {
fn receive_message(&self, msg: WebSocketRawMessage) { fn receive_message(&self, msg: WebSocketRawMessage) {
let handler = self.0.clone(); let handler = self.0.clone();
tokio::spawn(async move { tokio::spawn(async move {
handler.did_receive_ws_data(Bytes::from(msg.data)).await; handler.receive_ws_data(Bytes::from(msg.data)).await;
}); });
} }
} }

View file

@ -26,20 +26,16 @@ pub struct RevisionManager {
pub object_id: String, pub object_id: String,
user_id: String, user_id: String,
rev_id_counter: RevIdCounter, rev_id_counter: RevIdCounter,
cache: Arc<RwLock<RevisionCacheCompact>>, rev_compressor: Arc<RwLock<RevisionCompressor>>,
#[cfg(feature = "flowy_unit_test")] #[cfg(feature = "flowy_unit_test")]
revision_ack_notifier: tokio::sync::broadcast::Sender<i64>, rev_ack_notifier: tokio::sync::broadcast::Sender<i64>,
} }
impl RevisionManager { impl RevisionManager {
pub fn new(user_id: &str, object_id: &str, revision_cache: Arc<RevisionCache>) -> Self { pub fn new(user_id: &str, object_id: &str, revision_cache: Arc<RevisionCache>) -> Self {
let rev_id_counter = RevIdCounter::new(0); let rev_id_counter = RevIdCounter::new(0);
let cache = Arc::new(RwLock::new(RevisionCacheCompact::new( let rev_compressor = Arc::new(RwLock::new(RevisionCompressor::new(object_id, user_id, revision_cache)));
object_id,
user_id,
revision_cache,
)));
#[cfg(feature = "flowy_unit_test")] #[cfg(feature = "flowy_unit_test")]
let (revision_ack_notifier, _) = tokio::sync::broadcast::channel(1); let (revision_ack_notifier, _) = tokio::sync::broadcast::channel(1);
@ -47,10 +43,10 @@ impl RevisionManager {
object_id: object_id.to_string(), object_id: object_id.to_string(),
user_id: user_id.to_owned(), user_id: user_id.to_owned(),
rev_id_counter, rev_id_counter,
cache, rev_compressor,
#[cfg(feature = "flowy_unit_test")] #[cfg(feature = "flowy_unit_test")]
revision_ack_notifier, rev_ack_notifier: revision_ack_notifier,
} }
} }
@ -63,7 +59,7 @@ impl RevisionManager {
object_id: self.object_id.clone(), object_id: self.object_id.clone(),
user_id: self.user_id.clone(), user_id: self.user_id.clone(),
cloud, cloud,
cache: self.cache.clone(), rev_compressor: self.rev_compressor.clone(),
} }
.load::<C>() .load::<C>()
.await?; .await?;
@ -75,7 +71,7 @@ impl RevisionManager {
pub async fn reset_object(&self, revisions: RepeatedRevision) -> FlowyResult<()> { pub async fn reset_object(&self, revisions: RepeatedRevision) -> FlowyResult<()> {
let rev_id = pair_rev_id_from_revisions(&revisions).1; let rev_id = pair_rev_id_from_revisions(&revisions).1;
let write_guard = self.cache.write().await; let write_guard = self.rev_compressor.write().await;
let _ = write_guard.reset(revisions.into_inner()).await?; let _ = write_guard.reset(revisions.into_inner()).await?;
self.rev_id_counter.set(rev_id); self.rev_id_counter.set(rev_id);
Ok(()) Ok(())
@ -87,7 +83,7 @@ impl RevisionManager {
return Err(FlowyError::internal().context("Delta data should be empty")); return Err(FlowyError::internal().context("Delta data should be empty"));
} }
let write_guard = self.cache.write().await; let write_guard = self.rev_compressor.write().await;
let _ = write_guard.add_ack_revision(revision).await?; let _ = write_guard.add_ack_revision(revision).await?;
self.rev_id_counter.set(revision.rev_id); self.rev_id_counter.set(revision.rev_id);
Ok(()) Ok(())
@ -101,7 +97,7 @@ impl RevisionManager {
if revision.delta_data.is_empty() { if revision.delta_data.is_empty() {
return Err(FlowyError::internal().context("Delta data should be empty")); return Err(FlowyError::internal().context("Delta data should be empty"));
} }
let mut write_guard = self.cache.write().await; let mut write_guard = self.rev_compressor.write().await;
let rev_id = write_guard.write_sync_revision::<C>(revision).await?; let rev_id = write_guard.write_sync_revision::<C>(revision).await?;
self.rev_id_counter.set(rev_id); self.rev_id_counter.set(rev_id);
@ -110,9 +106,9 @@ impl RevisionManager {
#[tracing::instrument(level = "debug", skip(self), err)] #[tracing::instrument(level = "debug", skip(self), err)]
pub async fn ack_revision(&self, rev_id: i64) -> Result<(), FlowyError> { pub async fn ack_revision(&self, rev_id: i64) -> Result<(), FlowyError> {
if self.cache.write().await.ack_revision(rev_id).await.is_ok() { if self.rev_compressor.write().await.ack_revision(rev_id).await.is_ok() {
#[cfg(feature = "flowy_unit_test")] #[cfg(feature = "flowy_unit_test")]
let _ = self.revision_ack_notifier.send(rev_id); let _ = self.rev_ack_notifier.send(rev_id);
} }
Ok(()) Ok(())
} }
@ -128,37 +124,42 @@ impl RevisionManager {
} }
pub async fn get_revisions_in_range(&self, range: RevisionRange) -> Result<Vec<Revision>, FlowyError> { pub async fn get_revisions_in_range(&self, range: RevisionRange) -> Result<Vec<Revision>, FlowyError> {
let revisions = self.cache.read().await.revisions_in_range(&range).await?; let revisions = self.rev_compressor.read().await.revisions_in_range(&range).await?;
Ok(revisions) Ok(revisions)
} }
pub async fn next_sync_revision(&self) -> FlowyResult<Option<Revision>> { pub async fn next_sync_revision(&self) -> FlowyResult<Option<Revision>> {
Ok(self.cache.read().await.next_sync_revision().await?) Ok(self.rev_compressor.read().await.next_sync_revision().await?)
} }
pub async fn get_revision(&self, rev_id: i64) -> Option<Revision> { pub async fn get_revision(&self, rev_id: i64) -> Option<Revision> {
self.cache.read().await.get(rev_id).await.map(|record| record.revision) self.rev_compressor
.read()
.await
.get(rev_id)
.await
.map(|record| record.revision)
} }
} }
#[cfg(feature = "flowy_unit_test")] #[cfg(feature = "flowy_unit_test")]
impl RevisionManager { impl RevisionManager {
pub async fn revision_cache(&self) -> Arc<RevisionCache> { pub async fn revision_cache(&self) -> Arc<RevisionCache> {
self.cache.read().await.inner.clone() self.rev_compressor.read().await.inner.clone()
} }
pub fn revision_ack_receiver(&self) -> tokio::sync::broadcast::Receiver<i64> { pub fn ack_notify(&self) -> tokio::sync::broadcast::Receiver<i64> {
self.revision_ack_notifier.subscribe() self.rev_ack_notifier.subscribe()
} }
} }
struct RevisionCacheCompact { struct RevisionCompressor {
object_id: String, object_id: String,
user_id: String, user_id: String,
inner: Arc<RevisionCache>, inner: Arc<RevisionCache>,
sync_seq: RevisionSyncSequence, sync_seq: RevisionSyncSequence,
} }
impl RevisionCacheCompact { impl RevisionCompressor {
fn new(object_id: &str, user_id: &str, inner: Arc<RevisionCache>) -> Self { fn new(object_id: &str, user_id: &str, inner: Arc<RevisionCache>) -> Self {
let sync_seq = RevisionSyncSequence::new(); let sync_seq = RevisionSyncSequence::new();
let object_id = object_id.to_owned(); let object_id = object_id.to_owned();
@ -251,7 +252,7 @@ impl RevisionCacheCompact {
} }
} }
impl std::ops::Deref for RevisionCacheCompact { impl std::ops::Deref for RevisionCompressor {
type Target = Arc<RevisionCache>; type Target = Arc<RevisionCache>;
fn deref(&self) -> &Self::Target { fn deref(&self) -> &Self::Target {
@ -323,7 +324,7 @@ struct RevisionLoader {
object_id: String, object_id: String,
user_id: String, user_id: String,
cloud: Arc<dyn RevisionCloudService>, cloud: Arc<dyn RevisionCloudService>,
cache: Arc<RwLock<RevisionCacheCompact>>, rev_compressor: Arc<RwLock<RevisionCompressor>>,
} }
impl RevisionLoader { impl RevisionLoader {
@ -331,14 +332,14 @@ impl RevisionLoader {
where where
C: RevisionCompact, C: RevisionCompact,
{ {
let records = self.cache.read().await.batch_get(&self.object_id)?; let records = self.rev_compressor.read().await.batch_get(&self.object_id)?;
let revisions: Vec<Revision>; let revisions: Vec<Revision>;
let mut rev_id = 0; let mut rev_id = 0;
if records.is_empty() { if records.is_empty() {
let remote_revisions = self.cloud.fetch_object(&self.user_id, &self.object_id).await?; let remote_revisions = self.cloud.fetch_object(&self.user_id, &self.object_id).await?;
for revision in &remote_revisions { for revision in &remote_revisions {
rev_id = revision.rev_id; rev_id = revision.rev_id;
let _ = self.cache.read().await.add_ack_revision(revision).await?; let _ = self.rev_compressor.read().await.add_ack_revision(revision).await?;
} }
revisions = remote_revisions; revisions = remote_revisions;
} else { } else {
@ -346,7 +347,12 @@ impl RevisionLoader {
rev_id = record.revision.rev_id; rev_id = record.revision.rev_id;
if record.state == RevisionState::Sync { if record.state == RevisionState::Sync {
// Sync the records if their state is RevisionState::Sync. // Sync the records if their state is RevisionState::Sync.
let _ = self.cache.write().await.add_sync_revision(&record.revision).await?; let _ = self
.rev_compressor
.write()
.await
.add_sync_revision(&record.revision)
.await?;
} }
} }
revisions = records.into_iter().map(|record| record.revision).collect::<_>(); revisions = records.into_iter().map(|record| record.revision).collect::<_>();

View file

@ -44,7 +44,7 @@ pub struct RevisionWebSocketManager {
pub object_id: String, pub object_id: String,
sink_provider: Arc<dyn RevisionWSSinkDataProvider>, sink_provider: Arc<dyn RevisionWSSinkDataProvider>,
stream_consumer: Arc<dyn RevisionWSSteamConsumer>, stream_consumer: Arc<dyn RevisionWSSteamConsumer>,
web_socket: Arc<dyn RevisionWebSocket>, rev_web_socket: Arc<dyn RevisionWebSocket>,
pub ws_passthrough_tx: Sender<ServerRevisionWSData>, pub ws_passthrough_tx: Sender<ServerRevisionWSData>,
ws_passthrough_rx: Option<Receiver<ServerRevisionWSData>>, ws_passthrough_rx: Option<Receiver<ServerRevisionWSData>>,
pub state_passthrough_tx: broadcast::Sender<WSConnectState>, pub state_passthrough_tx: broadcast::Sender<WSConnectState>,
@ -60,7 +60,7 @@ impl RevisionWebSocketManager {
pub fn new( pub fn new(
object_name: &str, object_name: &str,
object_id: &str, object_id: &str,
web_socket: Arc<dyn RevisionWebSocket>, rev_web_socket: Arc<dyn RevisionWebSocket>,
sink_provider: Arc<dyn RevisionWSSinkDataProvider>, sink_provider: Arc<dyn RevisionWSSinkDataProvider>,
stream_consumer: Arc<dyn RevisionWSSteamConsumer>, stream_consumer: Arc<dyn RevisionWSSteamConsumer>,
ping_duration: Duration, ping_duration: Duration,
@ -75,7 +75,7 @@ impl RevisionWebSocketManager {
object_name, object_name,
sink_provider, sink_provider,
stream_consumer, stream_consumer,
web_socket, rev_web_socket,
ws_passthrough_tx, ws_passthrough_tx,
ws_passthrough_rx: Some(ws_passthrough_rx), ws_passthrough_rx: Some(ws_passthrough_rx),
state_passthrough_tx, state_passthrough_tx,
@ -91,7 +91,7 @@ impl RevisionWebSocketManager {
&self.object_id, &self.object_id,
&self.object_name, &self.object_name,
self.sink_provider.clone(), self.sink_provider.clone(),
self.web_socket.clone(), self.rev_web_socket.clone(),
self.stop_sync_tx.subscribe(), self.stop_sync_tx.subscribe(),
ping_duration, ping_duration,
); );

View file

@ -28,7 +28,7 @@ impl CrateConfig {
.flowy_config .flowy_config
.proto_crates .proto_crates
.iter() .iter()
.map(|name| path_buf_with_component(&self.crate_path, vec![&name])) .map(|name| path_buf_with_component(&self.crate_path, vec![name]))
.collect::<Vec<PathBuf>>(); .collect::<Vec<PathBuf>>();
proto_paths proto_paths
} }

View file

@ -31,7 +31,7 @@ pub fn parse_crate_protobuf(crate_paths: Vec<String>) -> Vec<ProtobufCrateContex
.collect::<Vec<ProtobufCrateContext>>() .collect::<Vec<ProtobufCrateContext>>()
} }
fn parse_files_protobuf(proto_crate_path: &PathBuf, proto_output_dir: &PathBuf) -> Vec<ProtoFile> { fn parse_files_protobuf(proto_crate_path: &Path, proto_output_dir: &Path) -> Vec<ProtoFile> {
let mut gen_proto_vec: Vec<ProtoFile> = vec![]; let mut gen_proto_vec: Vec<ProtoFile> = vec![];
// file_stem https://doc.rust-lang.org/std/path/struct.Path.html#method.file_stem // file_stem https://doc.rust-lang.org/std/path/struct.Path.html#method.file_stem
for (path, file_name) in WalkDir::new(proto_crate_path) for (path, file_name) in WalkDir::new(proto_crate_path)
@ -54,7 +54,7 @@ fn parse_files_protobuf(proto_crate_path: &PathBuf, proto_output_dir: &PathBuf)
.unwrap_or_else(|_| panic!("Unable to parse file at {}", path)); .unwrap_or_else(|_| panic!("Unable to parse file at {}", path));
let structs = get_ast_structs(&ast); let structs = get_ast_structs(&ast);
let proto_file = format!("{}.proto", &file_name); let proto_file = format!("{}.proto", &file_name);
let proto_file_path = path_string_with_component(&proto_output_dir, vec![&proto_file]); let proto_file_path = path_string_with_component(proto_output_dir, vec![&proto_file]);
let mut proto_file_content = parse_or_init_proto_file(proto_file_path.as_ref()); let mut proto_file_content = parse_or_init_proto_file(proto_file_path.as_ref());
structs.iter().for_each(|s| { structs.iter().for_each(|s| {

View file

@ -12,7 +12,7 @@ pub use proto_gen::*;
pub use proto_info::*; pub use proto_info::*;
use std::fs::File; use std::fs::File;
use std::io::Write; use std::io::Write;
use std::path::PathBuf; use std::path::{Path, PathBuf};
use std::process::Command; use std::process::Command;
use walkdir::WalkDir; use walkdir::WalkDir;
@ -56,7 +56,7 @@ pub fn gen(crate_name: &str, proto_file_dir: &str) {
generate_rust_protobuf_files(&protoc_bin_path, &proto_file_paths, proto_file_dir); generate_rust_protobuf_files(&protoc_bin_path, &proto_file_paths, proto_file_dir);
} }
fn generate_rust_protobuf_files(protoc_bin_path: &PathBuf, proto_file_paths: &Vec<String>, proto_file_dir: &str) { fn generate_rust_protobuf_files(protoc_bin_path: &Path, proto_file_paths: &[String], proto_file_dir: &str) {
protoc_rust::Codegen::new() protoc_rust::Codegen::new()
.out_dir("./src/protobuf/model") .out_dir("./src/protobuf/model")
.protoc_path(protoc_bin_path) .protoc_path(protoc_bin_path)

View file

@ -94,18 +94,18 @@ pub fn is_hidden(entry: &walkdir::DirEntry) -> bool {
entry.file_name().to_str().map(|s| s.starts_with('.')).unwrap_or(false) entry.file_name().to_str().map(|s| s.starts_with('.')).unwrap_or(false)
} }
pub fn create_dir_if_not_exist(dir: &PathBuf) { pub fn create_dir_if_not_exist(dir: &Path) {
if !dir.as_path().exists() { if !dir.exists() {
std::fs::create_dir_all(dir).unwrap(); std::fs::create_dir_all(dir).unwrap();
} }
} }
pub fn path_string_with_component(path: &PathBuf, components: Vec<&str>) -> String { pub fn path_string_with_component(path: &Path, components: Vec<&str>) -> String {
path_buf_with_component(path, components).to_str().unwrap().to_string() path_buf_with_component(path, components).to_str().unwrap().to_string()
} }
pub fn path_buf_with_component(path: &PathBuf, components: Vec<&str>) -> PathBuf { pub fn path_buf_with_component(path: &Path, components: Vec<&str>) -> PathBuf {
let mut path_buf = path.clone(); let mut path_buf = path.to_path_buf();
for component in components { for component in components {
path_buf.push(component); path_buf.push(component);
} }