|
@@ -1,4 +1,4 @@
|
|
|
-use crate::disk::RevisionRecord;
|
|
|
+use crate::disk::SyncRecord;
|
|
|
use crate::REVISION_WRITE_INTERVAL_IN_MILLIS;
|
|
|
use dashmap::DashMap;
|
|
|
use flowy_error::{FlowyError, FlowyResult};
|
|
@@ -7,15 +7,15 @@ use std::{borrow::Cow, sync::Arc, time::Duration};
|
|
|
use tokio::{sync::RwLock, task::JoinHandle};
|
|
|
|
|
|
pub(crate) trait RevisionMemoryCacheDelegate: Send + Sync {
|
|
|
- fn checkpoint_tick(&self, records: Vec<RevisionRecord>) -> FlowyResult<()>;
|
|
|
+ fn send_sync(&self, records: Vec<SyncRecord>) -> FlowyResult<()>;
|
|
|
fn receive_ack(&self, object_id: &str, rev_id: i64);
|
|
|
}
|
|
|
|
|
|
pub(crate) struct RevisionMemoryCache {
|
|
|
object_id: String,
|
|
|
- revs_map: Arc<DashMap<i64, RevisionRecord>>,
|
|
|
+ revs_map: Arc<DashMap<i64, SyncRecord>>,
|
|
|
delegate: Arc<dyn RevisionMemoryCacheDelegate>,
|
|
|
- pending_write_revs: Arc<RwLock<Vec<i64>>>,
|
|
|
+ defer_write_revs: Arc<RwLock<Vec<i64>>>,
|
|
|
defer_save: RwLock<Option<JoinHandle<()>>>,
|
|
|
}
|
|
|
|
|
@@ -25,7 +25,7 @@ impl RevisionMemoryCache {
|
|
|
object_id: object_id.to_owned(),
|
|
|
revs_map: Arc::new(DashMap::new()),
|
|
|
delegate,
|
|
|
- pending_write_revs: Arc::new(RwLock::new(vec![])),
|
|
|
+ defer_write_revs: Arc::new(RwLock::new(vec![])),
|
|
|
defer_save: RwLock::new(None),
|
|
|
}
|
|
|
}
|
|
@@ -34,7 +34,7 @@ impl RevisionMemoryCache {
|
|
|
self.revs_map.contains_key(rev_id)
|
|
|
}
|
|
|
|
|
|
- pub(crate) async fn add<'a>(&'a self, record: Cow<'a, RevisionRecord>) {
|
|
|
+ pub(crate) async fn add<'a>(&'a self, record: Cow<'a, SyncRecord>) {
|
|
|
let record = match record {
|
|
|
Cow::Borrowed(record) => record.clone(),
|
|
|
Cow::Owned(record) => record,
|
|
@@ -43,11 +43,11 @@ impl RevisionMemoryCache {
|
|
|
let rev_id = record.revision.rev_id;
|
|
|
self.revs_map.insert(rev_id, record);
|
|
|
|
|
|
- let mut write_guard = self.pending_write_revs.write().await;
|
|
|
+ let mut write_guard = self.defer_write_revs.write().await;
|
|
|
if !write_guard.contains(&rev_id) {
|
|
|
write_guard.push(rev_id);
|
|
|
drop(write_guard);
|
|
|
- self.make_checkpoint().await;
|
|
|
+ self.tick_checkpoint().await;
|
|
|
}
|
|
|
}
|
|
|
|
|
@@ -57,8 +57,8 @@ impl RevisionMemoryCache {
|
|
|
Some(mut record) => record.ack(),
|
|
|
}
|
|
|
|
|
|
- if self.pending_write_revs.read().await.contains(rev_id) {
|
|
|
- self.make_checkpoint().await;
|
|
|
+ if self.defer_write_revs.read().await.contains(rev_id) {
|
|
|
+ self.tick_checkpoint().await;
|
|
|
} else {
|
|
|
// The revision must be saved on disk if the pending_write_revs
|
|
|
// doesn't contains the rev_id.
|
|
@@ -66,7 +66,7 @@ impl RevisionMemoryCache {
|
|
|
}
|
|
|
}
|
|
|
|
|
|
- pub(crate) async fn get(&self, rev_id: &i64) -> Option<RevisionRecord> {
|
|
|
+ pub(crate) async fn get(&self, rev_id: &i64) -> Option<SyncRecord> {
|
|
|
self.revs_map.get(rev_id).map(|r| r.value().clone())
|
|
|
}
|
|
|
|
|
@@ -80,21 +80,21 @@ impl RevisionMemoryCache {
|
|
|
}
|
|
|
}
|
|
|
|
|
|
- pub(crate) async fn get_with_range(&self, range: &RevisionRange) -> Result<Vec<RevisionRecord>, FlowyError> {
|
|
|
+ pub(crate) async fn get_with_range(&self, range: &RevisionRange) -> Result<Vec<SyncRecord>, FlowyError> {
|
|
|
let revs = range
|
|
|
.iter()
|
|
|
.flat_map(|rev_id| self.revs_map.get(&rev_id).map(|record| record.clone()))
|
|
|
- .collect::<Vec<RevisionRecord>>();
|
|
|
+ .collect::<Vec<SyncRecord>>();
|
|
|
Ok(revs)
|
|
|
}
|
|
|
|
|
|
- pub(crate) async fn reset_with_revisions(&self, revision_records: Vec<RevisionRecord>) {
|
|
|
+ pub(crate) async fn reset_with_revisions(&self, revision_records: Vec<SyncRecord>) {
|
|
|
self.revs_map.clear();
|
|
|
if let Some(handler) = self.defer_save.write().await.take() {
|
|
|
handler.abort();
|
|
|
}
|
|
|
|
|
|
- let mut write_guard = self.pending_write_revs.write().await;
|
|
|
+ let mut write_guard = self.defer_write_revs.write().await;
|
|
|
write_guard.clear();
|
|
|
for record in revision_records {
|
|
|
write_guard.push(record.revision.rev_id);
|
|
@@ -102,21 +102,21 @@ impl RevisionMemoryCache {
|
|
|
}
|
|
|
drop(write_guard);
|
|
|
|
|
|
- self.make_checkpoint().await;
|
|
|
+ self.tick_checkpoint().await;
|
|
|
}
|
|
|
|
|
|
- async fn make_checkpoint(&self) {
|
|
|
+ async fn tick_checkpoint(&self) {
|
|
|
// https://github.com/async-graphql/async-graphql/blob/ed8449beec3d9c54b94da39bab33cec809903953/src/dataloader/mod.rs#L362
|
|
|
if let Some(handler) = self.defer_save.write().await.take() {
|
|
|
handler.abort();
|
|
|
}
|
|
|
|
|
|
- if self.pending_write_revs.read().await.is_empty() {
|
|
|
+ if self.defer_write_revs.read().await.is_empty() {
|
|
|
return;
|
|
|
}
|
|
|
|
|
|
let rev_map = self.revs_map.clone();
|
|
|
- let pending_write_revs = self.pending_write_revs.clone();
|
|
|
+ let pending_write_revs = self.defer_write_revs.clone();
|
|
|
let delegate = self.delegate.clone();
|
|
|
|
|
|
*self.defer_save.write().await = Some(tokio::spawn(async move {
|
|
@@ -128,7 +128,7 @@ impl RevisionMemoryCache {
|
|
|
//
|
|
|
// Use saturating_sub and split_off ?
|
|
|
// https://stackoverflow.com/questions/28952411/what-is-the-idiomatic-way-to-pop-the-last-n-elements-in-a-mutable-vec
|
|
|
- let mut save_records: Vec<RevisionRecord> = vec![];
|
|
|
+ let mut save_records: Vec<SyncRecord> = vec![];
|
|
|
revs_write_guard.iter().for_each(|rev_id| match rev_map.get(rev_id) {
|
|
|
None => {}
|
|
|
Some(value) => {
|
|
@@ -136,7 +136,7 @@ impl RevisionMemoryCache {
|
|
|
}
|
|
|
});
|
|
|
|
|
|
- if delegate.checkpoint_tick(save_records).is_ok() {
|
|
|
+ if delegate.send_sync(save_records).is_ok() {
|
|
|
revs_write_guard.clear();
|
|
|
drop(revs_write_guard);
|
|
|
}
|