| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441 |
- use std::collections::HashMap;
- use std::sync::{Arc, Weak};
- use collab::core::collab::{CollabRawData, MutexCollab};
- use collab_database::blocks::BlockEvent;
- use collab_database::database::{DatabaseData, YrsDocAction};
- use collab_database::error::DatabaseError;
- use collab_database::user::{
- CollabFuture, CollabObjectUpdate, CollabObjectUpdateByOid, DatabaseCollabService,
- WorkspaceDatabase,
- };
- use collab_database::views::{CreateDatabaseParams, CreateViewParams, DatabaseLayout};
- use collab_define::CollabType;
- use tokio::sync::RwLock;
- use collab_integrate::collab_builder::AppFlowyCollabBuilder;
- use collab_integrate::{CollabPersistenceConfig, RocksCollabDB};
- use flowy_database_deps::cloud::DatabaseCloudService;
- use flowy_error::{internal_error, FlowyError, FlowyResult};
- use flowy_task::TaskDispatcher;
- use crate::entities::{
- DatabaseDescriptionPB, DatabaseLayoutPB, DatabaseSnapshotPB, DidFetchRowPB,
- RepeatedDatabaseDescriptionPB,
- };
- use crate::notification::{send_notification, DatabaseNotification};
- use crate::services::database::DatabaseEditor;
- use crate::services::database_view::DatabaseLayoutDepsResolver;
- use crate::services::field_settings::default_field_settings_by_layout_map;
- use crate::services::share::csv::{CSVFormat, CSVImporter, ImportResult};
- pub trait DatabaseUser: Send + Sync {
- fn user_id(&self) -> Result<i64, FlowyError>;
- fn token(&self) -> Result<Option<String>, FlowyError>;
- fn collab_db(&self, uid: i64) -> Result<Weak<RocksCollabDB>, FlowyError>;
- }
- pub struct DatabaseManager {
- user: Arc<dyn DatabaseUser>,
- workspace_database: Arc<RwLock<Option<Arc<WorkspaceDatabase>>>>,
- task_scheduler: Arc<RwLock<TaskDispatcher>>,
- editors: RwLock<HashMap<String, Arc<DatabaseEditor>>>,
- collab_builder: Arc<AppFlowyCollabBuilder>,
- cloud_service: Arc<dyn DatabaseCloudService>,
- }
- impl DatabaseManager {
- pub fn new(
- database_user: Arc<dyn DatabaseUser>,
- task_scheduler: Arc<RwLock<TaskDispatcher>>,
- collab_builder: Arc<AppFlowyCollabBuilder>,
- cloud_service: Arc<dyn DatabaseCloudService>,
- ) -> Self {
- Self {
- user: database_user,
- workspace_database: Default::default(),
- task_scheduler,
- editors: Default::default(),
- collab_builder,
- cloud_service,
- }
- }
- fn is_collab_exist(&self, uid: i64, collab_db: &Weak<RocksCollabDB>, object_id: &str) -> bool {
- match collab_db.upgrade() {
- None => false,
- Some(collab_db) => {
- let read_txn = collab_db.read_txn();
- read_txn.is_exist(uid, object_id)
- },
- }
- }
- pub async fn initialize(
- &self,
- uid: i64,
- _workspace_id: String,
- database_views_aggregate_id: String,
- ) -> FlowyResult<()> {
- let collab_db = self.user.collab_db(uid)?;
- let collab_builder = UserDatabaseCollabServiceImpl {
- collab_builder: self.collab_builder.clone(),
- cloud_service: self.cloud_service.clone(),
- };
- let config = CollabPersistenceConfig::new().snapshot_per_update(10);
- let mut collab_raw_data = CollabRawData::default();
- // If the workspace database not exist in disk, try to fetch from remote.
- if !self.is_collab_exist(uid, &collab_db, &database_views_aggregate_id) {
- tracing::trace!("workspace database not exist, try to fetch from remote");
- match self
- .cloud_service
- .get_collab_update(&database_views_aggregate_id, CollabType::WorkspaceDatabase)
- .await
- {
- Ok(updates) => {
- collab_raw_data = updates;
- },
- Err(err) => {
- return Err(FlowyError::record_not_found().with_context(format!(
- "get workspace database :{} failed: {}",
- database_views_aggregate_id, err,
- )));
- },
- }
- }
- // Construct the workspace database.
- tracing::trace!("open workspace database: {}", &database_views_aggregate_id);
- let collab = collab_builder.build_collab_with_config(
- uid,
- &database_views_aggregate_id,
- CollabType::WorkspaceDatabase,
- collab_db.clone(),
- collab_raw_data,
- &config,
- );
- let workspace_database =
- WorkspaceDatabase::open(uid, collab, collab_db, config, collab_builder);
- subscribe_block_event(&workspace_database);
- *self.workspace_database.write().await = Some(Arc::new(workspace_database));
- // Remove all existing editors
- self.editors.write().await.clear();
- Ok(())
- }
- pub async fn initialize_with_new_user(
- &self,
- user_id: i64,
- workspace_id: String,
- database_views_aggregate_id: String,
- ) -> FlowyResult<()> {
- self
- .initialize(user_id, workspace_id, database_views_aggregate_id)
- .await?;
- Ok(())
- }
- pub async fn get_all_databases_description(&self) -> RepeatedDatabaseDescriptionPB {
- let mut items = vec![];
- if let Ok(wdb) = self.get_workspace_database().await {
- items = wdb
- .get_all_databases()
- .into_iter()
- .map(DatabaseDescriptionPB::from)
- .collect();
- }
- RepeatedDatabaseDescriptionPB { items }
- }
- pub async fn get_database_with_view_id(&self, view_id: &str) -> FlowyResult<Arc<DatabaseEditor>> {
- let database_id = self.get_database_id_with_view_id(view_id).await?;
- self.get_database(&database_id).await
- }
- pub async fn get_database_id_with_view_id(&self, view_id: &str) -> FlowyResult<String> {
- let wdb = self.get_workspace_database().await?;
- wdb.get_database_id_with_view_id(view_id).ok_or_else(|| {
- FlowyError::record_not_found()
- .with_context(format!("The database for view id: {} not found", view_id))
- })
- }
- pub async fn get_database(&self, database_id: &str) -> FlowyResult<Arc<DatabaseEditor>> {
- if let Some(editor) = self.editors.read().await.get(database_id) {
- return Ok(editor.clone());
- }
- self.open_database(database_id).await
- }
- pub async fn open_database(&self, database_id: &str) -> FlowyResult<Arc<DatabaseEditor>> {
- tracing::trace!("create new editor for database {}", database_id);
- let mut editors = self.editors.write().await;
- let wdb = self.get_workspace_database().await?;
- let database = wdb
- .get_database(database_id)
- .await
- .ok_or_else(FlowyError::collab_not_sync)?;
- let editor = Arc::new(DatabaseEditor::new(database, self.task_scheduler.clone()).await?);
- editors.insert(database_id.to_string(), editor.clone());
- Ok(editor)
- }
- #[tracing::instrument(level = "debug", skip_all)]
- pub async fn close_database_view<T: AsRef<str>>(&self, view_id: T) -> FlowyResult<()> {
- // TODO(natan): defer closing the database if the sync is not finished
- let view_id = view_id.as_ref();
- let wdb = self.get_workspace_database().await?;
- let database_id = wdb.get_database_id_with_view_id(view_id);
- if database_id.is_some() {
- wdb.close_database(database_id.as_ref().unwrap());
- }
- if let Some(database_id) = database_id {
- let mut editors = self.editors.write().await;
- if let Some(editor) = editors.get(&database_id) {
- if editor.close_view_editor(view_id).await {
- editor.close().await;
- editors.remove(&database_id);
- }
- }
- }
- Ok(())
- }
- pub async fn delete_database_view(&self, view_id: &str) -> FlowyResult<()> {
- let database = self.get_database_with_view_id(view_id).await?;
- let _ = database.delete_database_view(view_id).await?;
- Ok(())
- }
- pub async fn duplicate_database(&self, view_id: &str) -> FlowyResult<Vec<u8>> {
- let wdb = self.get_workspace_database().await?;
- let data = wdb.get_database_duplicated_data(view_id).await?;
- let json_bytes = data.to_json_bytes()?;
- Ok(json_bytes)
- }
- /// Create a new database with the given data that can be deserialized to [DatabaseData].
- #[tracing::instrument(level = "trace", skip_all, err)]
- pub async fn create_database_with_database_data(
- &self,
- view_id: &str,
- data: Vec<u8>,
- ) -> FlowyResult<()> {
- let mut database_data = DatabaseData::from_json_bytes(data)?;
- database_data.view.id = view_id.to_string();
- let wdb = self.get_workspace_database().await?;
- let _ = wdb.create_database_with_data(database_data)?;
- Ok(())
- }
- pub async fn create_database_with_params(&self, params: CreateDatabaseParams) -> FlowyResult<()> {
- let wdb = self.get_workspace_database().await?;
- let _ = wdb.create_database(params)?;
- Ok(())
- }
- /// A linked view is a view that is linked to existing database.
- #[tracing::instrument(level = "trace", skip(self), err)]
- pub async fn create_linked_view(
- &self,
- name: String,
- layout: DatabaseLayout,
- database_id: String,
- database_view_id: String,
- ) -> FlowyResult<()> {
- let wdb = self.get_workspace_database().await?;
- let mut params = CreateViewParams::new(database_id.clone(), database_view_id, name, layout);
- if let Some(database) = wdb.get_database(&database_id).await {
- let (field, layout_setting) = DatabaseLayoutDepsResolver::new(database, layout)
- .resolve_deps_when_create_database_linked_view();
- if let Some(field) = field {
- params = params.with_deps_fields(vec![field], vec![default_field_settings_by_layout_map()]);
- }
- if let Some(layout_setting) = layout_setting {
- params = params.with_layout_setting(layout_setting);
- }
- };
- wdb.create_database_linked_view(params).await?;
- Ok(())
- }
- pub async fn import_csv(
- &self,
- view_id: String,
- content: String,
- format: CSVFormat,
- ) -> FlowyResult<ImportResult> {
- let params = tokio::task::spawn_blocking(move || {
- CSVImporter.import_csv_from_string(view_id, content, format)
- })
- .await
- .map_err(internal_error)??;
- let result = ImportResult {
- database_id: params.database_id.clone(),
- view_id: params.view_id.clone(),
- };
- self.create_database_with_params(params).await?;
- Ok(result)
- }
- // will implement soon
- pub async fn import_csv_from_file(
- &self,
- _file_path: String,
- _format: CSVFormat,
- ) -> FlowyResult<()> {
- Ok(())
- }
- pub async fn export_csv(&self, view_id: &str, style: CSVFormat) -> FlowyResult<String> {
- let database = self.get_database_with_view_id(view_id).await?;
- database.export_csv(style).await
- }
- pub async fn update_database_layout(
- &self,
- view_id: &str,
- layout: DatabaseLayoutPB,
- ) -> FlowyResult<()> {
- let database = self.get_database_with_view_id(view_id).await?;
- database.update_view_layout(view_id, layout.into()).await
- }
- pub async fn get_database_snapshots(
- &self,
- view_id: &str,
- limit: usize,
- ) -> FlowyResult<Vec<DatabaseSnapshotPB>> {
- let database_id = self.get_database_id_with_view_id(view_id).await?;
- let snapshots = self
- .cloud_service
- .get_collab_snapshots(&database_id, limit)
- .await?
- .into_iter()
- .map(|snapshot| DatabaseSnapshotPB {
- snapshot_id: snapshot.snapshot_id,
- snapshot_desc: "".to_string(),
- created_at: snapshot.created_at,
- data: snapshot.data,
- })
- .collect::<Vec<_>>();
- Ok(snapshots)
- }
- async fn get_workspace_database(&self) -> FlowyResult<Arc<WorkspaceDatabase>> {
- let database = self.workspace_database.read().await;
- match &*database {
- None => Err(FlowyError::internal().with_context("Workspace database not initialized")),
- Some(user_database) => Ok(user_database.clone()),
- }
- }
- /// Only expose this method for testing
- #[cfg(debug_assertions)]
- pub fn get_cloud_service(&self) -> &Arc<dyn DatabaseCloudService> {
- &self.cloud_service
- }
- }
- /// Send notification to all clients that are listening to the given object.
- fn subscribe_block_event(workspace_database: &WorkspaceDatabase) {
- let mut block_event_rx = workspace_database.subscribe_block_event();
- tokio::spawn(async move {
- while let Ok(event) = block_event_rx.recv().await {
- match event {
- BlockEvent::DidFetchRow(row_details) => {
- for row_detail in row_details {
- tracing::trace!("Did fetch row: {:?}", row_detail.row.id);
- let row_id = row_detail.row.id.clone();
- let pb = DidFetchRowPB::from(row_detail);
- send_notification(&row_id, DatabaseNotification::DidFetchRow)
- .payload(pb)
- .send();
- }
- },
- }
- }
- });
- }
- struct UserDatabaseCollabServiceImpl {
- collab_builder: Arc<AppFlowyCollabBuilder>,
- cloud_service: Arc<dyn DatabaseCloudService>,
- }
- impl DatabaseCollabService for UserDatabaseCollabServiceImpl {
- fn get_collab_update(
- &self,
- object_id: &str,
- object_ty: CollabType,
- ) -> CollabFuture<Result<CollabObjectUpdate, DatabaseError>> {
- let object_id = object_id.to_string();
- let weak_cloud_service = Arc::downgrade(&self.cloud_service);
- Box::pin(async move {
- match weak_cloud_service.upgrade() {
- None => {
- tracing::warn!("Cloud service is dropped");
- Ok(vec![])
- },
- Some(cloud_service) => {
- let updates = cloud_service
- .get_collab_update(&object_id, object_ty)
- .await?;
- Ok(updates)
- },
- }
- })
- }
- fn batch_get_collab_update(
- &self,
- object_ids: Vec<String>,
- object_ty: CollabType,
- ) -> CollabFuture<Result<CollabObjectUpdateByOid, DatabaseError>> {
- let weak_cloud_service = Arc::downgrade(&self.cloud_service);
- Box::pin(async move {
- match weak_cloud_service.upgrade() {
- None => {
- tracing::warn!("Cloud service is dropped");
- Ok(CollabObjectUpdateByOid::default())
- },
- Some(cloud_service) => {
- let updates = cloud_service
- .batch_get_collab_updates(object_ids, object_ty)
- .await?;
- Ok(updates)
- },
- }
- })
- }
- fn build_collab_with_config(
- &self,
- uid: i64,
- object_id: &str,
- object_type: CollabType,
- collab_db: Weak<RocksCollabDB>,
- collab_raw_data: CollabRawData,
- config: &CollabPersistenceConfig,
- ) -> Arc<MutexCollab> {
- self
- .collab_builder
- .build_with_config(
- uid,
- object_id,
- object_type,
- collab_db,
- collab_raw_data,
- config,
- )
- .unwrap()
- }
- }
|