From 1b8d6c6e85025ef0b6f102d3e7e857ab137707d9 Mon Sep 17 00:00:00 2001 From: Damien Elmes Date: Thu, 11 Mar 2021 15:47:31 +1000 Subject: [PATCH] split out sync, notetypes and config code --- rslib/backend.proto | 142 ++--- rslib/build/protobuf.rs | 2 +- rslib/src/backend/config.rs | 76 +++ rslib/src/backend/decks.rs | 127 ++++ rslib/src/backend/mod.rs | 560 ++---------------- rslib/src/backend/notes.rs | 172 ++++++ rslib/src/backend/notetypes.rs | 89 +++ rslib/src/backend/scheduler/mod.rs | 2 +- rslib/src/backend/{sync.rs => sync/mod.rs} | 89 ++- .../{http_sync_server.rs => sync/server.rs} | 21 +- 10 files changed, 664 insertions(+), 616 deletions(-) create mode 100644 rslib/src/backend/decks.rs create mode 100644 rslib/src/backend/notes.rs create mode 100644 rslib/src/backend/notetypes.rs rename rslib/src/backend/{sync.rs => sync/mod.rs} (78%) rename rslib/src/backend/{http_sync_server.rs => sync/server.rs} (94%) diff --git a/rslib/backend.proto b/rslib/backend.proto index 11a390895..a7b02f105 100644 --- a/rslib/backend.proto +++ b/rslib/backend.proto @@ -80,7 +80,12 @@ message DeckConfigID { /// sure all clients agree on the same service indices enum ServiceIndex { SERVICE_INDEX_SCHEDULING = 0; - SERVICE_INDEX_BACKEND = 1; + SERVICE_INDEX_DECKS = 1; + SERVICE_INDEX_NOTES = 2; + SERVICE_INDEX_SYNC = 3; + SERVICE_INDEX_NOTETYPES = 4; + SERVICE_INDEX_CONFIG = 5; + SERVICE_INDEX_BACKEND = 99; } service SchedulingService { @@ -108,6 +113,72 @@ service SchedulingService { rpc GetQueuedCards(GetQueuedCardsIn) returns (GetQueuedCardsOut); } +service DecksService { + rpc AddOrUpdateDeckLegacy(AddOrUpdateDeckLegacyIn) returns (DeckID); + rpc DeckTree(DeckTreeIn) returns (DeckTreeNode); + rpc DeckTreeLegacy(Empty) returns (Json); + rpc GetAllDecksLegacy(Empty) returns (Json); + rpc GetDeckIDByName(String) returns (DeckID); + rpc GetDeckLegacy(DeckID) returns (Json); + rpc GetDeckNames(GetDeckNamesIn) returns (DeckNames); + rpc NewDeckLegacy(Bool) returns (Json); + rpc RemoveDeck(DeckID) returns (Empty); + rpc DragDropDecks(DragDropDecksIn) returns (Empty); +} + +service NotesService { + rpc NewNote(NoteTypeID) returns (Note); + rpc AddNote(AddNoteIn) returns (NoteID); + rpc DefaultsForAdding(DefaultsForAddingIn) returns (DeckAndNotetype); + rpc DefaultDeckForNotetype(NoteTypeID) returns (DeckID); + rpc UpdateNote(UpdateNoteIn) returns (Empty); + rpc GetNote(NoteID) returns (Note); + rpc RemoveNotes(RemoveNotesIn) returns (Empty); + rpc AddNoteTags(AddNoteTagsIn) returns (UInt32); + rpc UpdateNoteTags(UpdateNoteTagsIn) returns (UInt32); + rpc ClozeNumbersInNote(Note) returns (ClozeNumbersInNoteOut); + rpc AfterNoteUpdates(AfterNoteUpdatesIn) returns (Empty); + rpc FieldNamesForNotes(FieldNamesForNotesIn) returns (FieldNamesForNotesOut); + rpc NoteIsDuplicateOrEmpty(Note) returns (NoteIsDuplicateOrEmptyOut); + rpc CardsOfNote(NoteID) returns (CardIDs); +} + +service SyncService { + rpc SyncMedia(SyncAuth) returns (Empty); + rpc AbortSync(Empty) returns (Empty); + rpc AbortMediaSync(Empty) returns (Empty); + rpc BeforeUpload(Empty) returns (Empty); + rpc SyncLogin(SyncLoginIn) returns (SyncAuth); + rpc SyncStatus(SyncAuth) returns (SyncStatusOut); + rpc SyncCollection(SyncAuth) returns (SyncCollectionOut); + rpc FullUpload(SyncAuth) returns (Empty); + rpc FullDownload(SyncAuth) returns (Empty); + rpc SyncServerMethod(SyncServerMethodIn) returns (Json); +} + +service ConfigService { + rpc GetConfigJson(String) returns (Json); + rpc SetConfigJson(SetConfigJsonIn) returns (Empty); + rpc RemoveConfig(String) returns (Empty); + rpc GetAllConfig(Empty) returns (Json); + rpc GetConfigBool(Config.Bool) returns (Bool); + rpc SetConfigBool(SetConfigBoolIn) returns (Empty); + rpc GetConfigString(Config.String) returns (String); + rpc SetConfigString(SetConfigStringIn) returns (Empty); + rpc GetPreferences(Empty) returns (Preferences); + rpc SetPreferences(Preferences) returns (Empty); +} + +service NoteTypesService { + rpc AddOrUpdateNotetype(AddOrUpdateNotetypeIn) returns (NoteTypeID); + rpc GetStockNotetypeLegacy(StockNoteType) returns (Json); + rpc GetNotetypeLegacy(NoteTypeID) returns (Json); + rpc GetNotetypeNames(Empty) returns (NoteTypeNames); + rpc GetNotetypeNamesAndCounts(Empty) returns (NoteTypeUseCounts); + rpc GetNotetypeIDByName(String) returns (NoteTypeID); + rpc RemoveNotetype(NoteTypeID) returns (Empty); +} + service BackendService { rpc LatestProgress(Empty) returns (Progress); rpc SetWantsAbort(Empty) returns (Empty); @@ -145,19 +216,6 @@ service BackendService { rpc EmptyTrash(Empty) returns (Empty); rpc RestoreTrash(Empty) returns (Empty); - // decks - - rpc AddOrUpdateDeckLegacy(AddOrUpdateDeckLegacyIn) returns (DeckID); - rpc DeckTree(DeckTreeIn) returns (DeckTreeNode); - rpc DeckTreeLegacy(Empty) returns (Json); - rpc GetAllDecksLegacy(Empty) returns (Json); - rpc GetDeckIDByName(String) returns (DeckID); - rpc GetDeckLegacy(DeckID) returns (Json); - rpc GetDeckNames(GetDeckNamesIn) returns (DeckNames); - rpc NewDeckLegacy(Bool) returns (Json); - rpc RemoveDeck(DeckID) returns (Empty); - rpc DragDropDecks(DragDropDecksIn) returns (Empty); - // deck config rpc AddOrUpdateDeckConfigLegacy(AddOrUpdateDeckConfigLegacyIn) @@ -174,33 +232,6 @@ service BackendService { rpc RemoveCards(RemoveCardsIn) returns (Empty); rpc SetDeck(SetDeckIn) returns (Empty); - // notes - - rpc NewNote(NoteTypeID) returns (Note); - rpc AddNote(AddNoteIn) returns (NoteID); - rpc DefaultsForAdding(DefaultsForAddingIn) returns (DeckAndNotetype); - rpc DefaultDeckForNotetype(NoteTypeID) returns (DeckID); - rpc UpdateNote(UpdateNoteIn) returns (Empty); - rpc GetNote(NoteID) returns (Note); - rpc RemoveNotes(RemoveNotesIn) returns (Empty); - rpc AddNoteTags(AddNoteTagsIn) returns (UInt32); - rpc UpdateNoteTags(UpdateNoteTagsIn) returns (UInt32); - rpc ClozeNumbersInNote(Note) returns (ClozeNumbersInNoteOut); - rpc AfterNoteUpdates(AfterNoteUpdatesIn) returns (Empty); - rpc FieldNamesForNotes(FieldNamesForNotesIn) returns (FieldNamesForNotesOut); - rpc NoteIsDuplicateOrEmpty(Note) returns (NoteIsDuplicateOrEmptyOut); - rpc CardsOfNote(NoteID) returns (CardIDs); - - // note types - - rpc AddOrUpdateNotetype(AddOrUpdateNotetypeIn) returns (NoteTypeID); - rpc GetStockNotetypeLegacy(StockNoteType) returns (Json); - rpc GetNotetypeLegacy(NoteTypeID) returns (Json); - rpc GetNotetypeNames(Empty) returns (NoteTypeNames); - rpc GetNotetypeNamesAndCounts(Empty) returns (NoteTypeUseCounts); - rpc GetNotetypeIDByName(String) returns (NoteTypeID); - rpc RemoveNotetype(NoteTypeID) returns (Empty); - // collection rpc OpenCollection(OpenCollectionIn) returns (Empty); @@ -210,19 +241,6 @@ service BackendService { rpc Undo(Empty) returns (UndoStatus); rpc Redo(Empty) returns (UndoStatus); - // sync - - rpc SyncMedia(SyncAuth) returns (Empty); - rpc AbortSync(Empty) returns (Empty); - rpc AbortMediaSync(Empty) returns (Empty); - rpc BeforeUpload(Empty) returns (Empty); - rpc SyncLogin(SyncLoginIn) returns (SyncAuth); - rpc SyncStatus(SyncAuth) returns (SyncStatusOut); - rpc SyncCollection(SyncAuth) returns (SyncCollectionOut); - rpc FullUpload(SyncAuth) returns (Empty); - rpc FullDownload(SyncAuth) returns (Empty); - rpc SyncServerMethod(SyncServerMethodIn) returns (Json); - // translation/messages/text manipulation rpc TranslateString(TranslateStringIn) returns (String); @@ -238,22 +256,6 @@ service BackendService { rpc ClearTag(String) returns (Empty); rpc TagTree(Empty) returns (TagTreeNode); rpc DragDropTags(DragDropTagsIn) returns (Empty); - - // config - - rpc GetConfigJson(String) returns (Json); - rpc SetConfigJson(SetConfigJsonIn) returns (Empty); - rpc RemoveConfig(String) returns (Empty); - rpc GetAllConfig(Empty) returns (Json); - rpc GetConfigBool(Config.Bool) returns (Bool); - rpc SetConfigBool(SetConfigBoolIn) returns (Empty); - rpc GetConfigString(Config.String) returns (String); - rpc SetConfigString(SetConfigStringIn) returns (Empty); - - // preferences - - rpc GetPreferences(Empty) returns (Preferences); - rpc SetPreferences(Preferences) returns (Empty); } // Protobuf stored in .anki2 files diff --git a/rslib/build/protobuf.rs b/rslib/build/protobuf.rs index aad63055b..65c1c75ff 100644 --- a/rslib/build/protobuf.rs +++ b/rslib/build/protobuf.rs @@ -55,7 +55,7 @@ impl prost_build::ServiceGenerator for CustomGenerator { fn generate(&mut self, service: prost_build::Service, buf: &mut String) { write!( buf, - "pub mod {name} {{ + "pub mod {name}_service {{ use super::*; use prost::Message; use crate::err::Result; diff --git a/rslib/src/backend/config.rs b/rslib/src/backend/config.rs index 0deaa12e7..965be7e6d 100644 --- a/rslib/src/backend/config.rs +++ b/rslib/src/backend/config.rs @@ -1,12 +1,16 @@ // Copyright: Ankitects Pty Ltd and contributors // License: GNU AGPL, version 3 or later; http://www.gnu.org/licenses/agpl.html +use super::Backend; use crate::{ backend_proto as pb, config::{BoolKey, StringKey}, + prelude::*, }; use pb::config::bool::Key as BoolKeyProto; use pb::config::string::Key as StringKeyProto; +pub(super) use pb::config_service::Service as ConfigService; +use serde_json::Value; impl From for BoolKey { fn from(k: BoolKeyProto) -> Self { @@ -38,3 +42,75 @@ impl From for StringKey { } } } + +impl ConfigService for Backend { + fn get_config_json(&self, input: pb::String) -> Result { + self.with_col(|col| { + let val: Option = col.get_config_optional(input.val.as_str()); + val.ok_or(AnkiError::NotFound) + .and_then(|v| serde_json::to_vec(&v).map_err(Into::into)) + .map(Into::into) + }) + } + + fn set_config_json(&self, input: pb::SetConfigJsonIn) -> Result { + self.with_col(|col| { + col.transact(None, |col| { + // ensure it's a well-formed object + let val: Value = serde_json::from_slice(&input.value_json)?; + col.set_config(input.key.as_str(), &val) + }) + }) + .map(Into::into) + } + + fn remove_config(&self, input: pb::String) -> Result { + self.with_col(|col| col.transact(None, |col| col.remove_config(input.val.as_str()))) + .map(Into::into) + } + + fn get_all_config(&self, _input: pb::Empty) -> Result { + self.with_col(|col| { + let conf = col.storage.get_all_config()?; + serde_json::to_vec(&conf).map_err(Into::into) + }) + .map(Into::into) + } + + fn get_config_bool(&self, input: pb::config::Bool) -> Result { + self.with_col(|col| { + Ok(pb::Bool { + val: col.get_bool(input.key().into()), + }) + }) + } + + fn set_config_bool(&self, input: pb::SetConfigBoolIn) -> Result { + self.with_col(|col| col.transact(None, |col| col.set_bool(input.key().into(), input.value))) + .map(Into::into) + } + + fn get_config_string(&self, input: pb::config::String) -> Result { + self.with_col(|col| { + Ok(pb::String { + val: col.get_string(input.key().into()), + }) + }) + } + + fn set_config_string(&self, input: pb::SetConfigStringIn) -> Result { + self.with_col(|col| { + col.transact(None, |col| col.set_string(input.key().into(), &input.value)) + }) + .map(Into::into) + } + + fn get_preferences(&self, _input: pb::Empty) -> Result { + self.with_col(|col| col.get_preferences()) + } + + fn set_preferences(&self, input: pb::Preferences) -> Result { + self.with_col(|col| col.set_preferences(input)) + .map(Into::into) + } +} diff --git a/rslib/src/backend/decks.rs b/rslib/src/backend/decks.rs new file mode 100644 index 000000000..e072f777f --- /dev/null +++ b/rslib/src/backend/decks.rs @@ -0,0 +1,127 @@ +// Copyright: Ankitects Pty Ltd and contributors +// License: GNU AGPL, version 3 or later; http://www.gnu.org/licenses/agpl.html + +use super::Backend; +use crate::{ + backend_proto::{self as pb}, + decks::{Deck, DeckID, DeckSchema11}, + prelude::*, +}; +pub(super) use pb::decks_service::Service as DecksService; + +impl DecksService for Backend { + fn add_or_update_deck_legacy(&self, input: pb::AddOrUpdateDeckLegacyIn) -> Result { + self.with_col(|col| { + let schema11: DeckSchema11 = serde_json::from_slice(&input.deck)?; + let mut deck: Deck = schema11.into(); + if input.preserve_usn_and_mtime { + col.transact(None, |col| { + let usn = col.usn()?; + col.add_or_update_single_deck_with_existing_id(&mut deck, usn) + })?; + } else { + col.add_or_update_deck(&mut deck)?; + } + Ok(pb::DeckId { did: deck.id.0 }) + }) + } + + fn deck_tree(&self, input: pb::DeckTreeIn) -> Result { + let lim = if input.top_deck_id > 0 { + Some(DeckID(input.top_deck_id)) + } else { + None + }; + self.with_col(|col| { + let now = if input.now == 0 { + None + } else { + Some(TimestampSecs(input.now)) + }; + col.deck_tree(now, lim) + }) + } + + fn deck_tree_legacy(&self, _input: pb::Empty) -> Result { + self.with_col(|col| { + let tree = col.legacy_deck_tree()?; + serde_json::to_vec(&tree) + .map_err(Into::into) + .map(Into::into) + }) + } + + fn get_all_decks_legacy(&self, _input: pb::Empty) -> Result { + self.with_col(|col| { + let decks = col.storage.get_all_decks_as_schema11()?; + serde_json::to_vec(&decks).map_err(Into::into) + }) + .map(Into::into) + } + + fn get_deck_id_by_name(&self, input: pb::String) -> Result { + self.with_col(|col| { + col.get_deck_id(&input.val).and_then(|d| { + d.ok_or(AnkiError::NotFound) + .map(|d| pb::DeckId { did: d.0 }) + }) + }) + } + + fn get_deck_legacy(&self, input: pb::DeckId) -> Result { + self.with_col(|col| { + let deck: DeckSchema11 = col + .storage + .get_deck(input.into())? + .ok_or(AnkiError::NotFound)? + .into(); + serde_json::to_vec(&deck) + .map_err(Into::into) + .map(Into::into) + }) + } + + fn get_deck_names(&self, input: pb::GetDeckNamesIn) -> Result { + self.with_col(|col| { + let names = if input.include_filtered { + col.get_all_deck_names(input.skip_empty_default)? + } else { + col.get_all_normal_deck_names()? + }; + Ok(pb::DeckNames { + entries: names + .into_iter() + .map(|(id, name)| pb::DeckNameId { id: id.0, name }) + .collect(), + }) + }) + } + + fn new_deck_legacy(&self, input: pb::Bool) -> Result { + let deck = if input.val { + Deck::new_filtered() + } else { + Deck::new_normal() + }; + let schema11: DeckSchema11 = deck.into(); + serde_json::to_vec(&schema11) + .map_err(Into::into) + .map(Into::into) + } + + fn remove_deck(&self, input: pb::DeckId) -> Result { + self.with_col(|col| col.remove_deck_and_child_decks(input.into())) + .map(Into::into) + } + + fn drag_drop_decks(&self, input: pb::DragDropDecksIn) -> Result { + let source_dids: Vec<_> = input.source_deck_ids.into_iter().map(Into::into).collect(); + let target_did = if input.target_deck_id == 0 { + None + } else { + Some(input.target_deck_id.into()) + }; + self.with_col(|col| col.drag_drop_decks(&source_dids, target_did)) + .map(Into::into) + } +} diff --git a/rslib/src/backend/mod.rs b/rslib/src/backend/mod.rs index da51c86a5..ecc65e4f7 100644 --- a/rslib/src/backend/mod.rs +++ b/rslib/src/backend/mod.rs @@ -5,26 +5,33 @@ mod adding; mod card; mod config; mod dbproxy; +mod decks; mod err; mod generic; -mod http_sync_server; +mod notes; +mod notetypes; mod progress; mod scheduler; mod search; mod sync; -use self::scheduler::SchedulingService; -use crate::backend_proto::backend::Service as BackendService; +use self::{ + config::ConfigService, + decks::DecksService, + notes::NotesService, + notetypes::NoteTypesService, + scheduler::SchedulingService, + sync::{SyncService, SyncState}, +}; +use crate::backend_proto::backend_service::Service as BackendService; use crate::{ backend::dbproxy::db_command_bytes, backend_proto as pb, - backend_proto::{AddOrUpdateDeckConfigLegacyIn, Empty, RenderedTemplateReplacement}, + backend_proto::{AddOrUpdateDeckConfigLegacyIn, RenderedTemplateReplacement}, card::{Card, CardID}, - cloze::add_cloze_numbers_in_string, collection::{open_collection, Collection}, deckconf::{DeckConf, DeckConfSchema11}, - decks::{Deck, DeckID, DeckSchema11}, err::{AnkiError, Result}, i18n::I18n, latex::{extract_latex, extract_latex_expanding_clozes, ExtractedLatex}, @@ -33,26 +40,20 @@ use crate::{ markdown::render_markdown, media::check::MediaChecker, media::MediaManager, - notes::{Note, NoteID}, - notetype::{ - all_stock_notetypes, CardTemplateSchema11, NoteType, NoteTypeSchema11, RenderCardOutput, - }, + notes::NoteID, + notetype::{CardTemplateSchema11, RenderCardOutput}, scheduler::timespan::{answer_button_time, time_span}, search::{concatenate_searches, replace_search_node, write_nodes, Node}, - sync::{http::SyncRequest, LocalServer}, template::RenderedNode, text::{extract_av_tags, sanitize_html_no_images, strip_av_tags, AVTag}, - timestamp::TimestampSecs, undo::UndoableOpKind, }; use fluent::FluentValue; -use futures::future::AbortHandle; use log::error; use once_cell::sync::OnceCell; use progress::{AbortHandleSlot, Progress}; use prost::Message; -use serde_json::Value as JsonValue; -use std::{collections::HashSet, convert::TryInto}; +use std::convert::TryInto; use std::{ result, sync::{Arc, Mutex}, @@ -62,7 +63,6 @@ use tokio::runtime::{self, Runtime}; use self::{ err::anki_error_to_proto_error, progress::{progress_to_proto, ProgressState}, - sync::RemoteSyncStatus, }; pub struct Backend { @@ -79,9 +79,7 @@ pub struct Backend { #[derive(Default)] struct BackendState { - remote_sync_status: RemoteSyncStatus, - media_sync_abort: Option, - http_sync_server: Option, + sync: SyncState, } pub fn init_backend(init_msg: &[u8]) -> std::result::Result { @@ -100,12 +98,12 @@ pub fn init_backend(init_msg: &[u8]) -> std::result::Result { } impl BackendService for Backend { - fn latest_progress(&self, _input: Empty) -> Result { + fn latest_progress(&self, _input: pb::Empty) -> Result { let progress = self.progress_state.lock().unwrap().last_progress; Ok(progress_to_proto(progress, &self.i18n)) } - fn set_wants_abort(&self, _input: Empty) -> Result { + fn set_wants_abort(&self, _input: pb::Empty) -> Result { self.progress_state.lock().unwrap().want_abort = true; Ok(().into()) } @@ -304,7 +302,7 @@ impl BackendService for Backend { self.with_col(|col| col.get_graph_preferences()) } - fn set_graph_preferences(&self, input: pb::GraphPreferences) -> Result { + fn set_graph_preferences(&self, input: pb::GraphPreferences) -> Result { self.with_col(|col| col.set_graph_preferences(input)) .map(Into::into) } @@ -334,7 +332,7 @@ impl BackendService for Backend { }) } - fn trash_media_files(&self, input: pb::TrashMediaFilesIn) -> Result { + fn trash_media_files(&self, input: pb::TrashMediaFilesIn) -> Result { self.with_col(|col| { let mgr = MediaManager::new(&col.media_folder, &col.media_db)?; let mut ctx = mgr.dbctx(); @@ -354,7 +352,7 @@ impl BackendService for Backend { }) } - fn empty_trash(&self, _input: Empty) -> Result { + fn empty_trash(&self, _input: pb::Empty) -> Result { let mut handler = self.new_progress_handler(); let progress_fn = move |progress| handler.update(Progress::MediaCheck(progress as u32), true); @@ -370,7 +368,7 @@ impl BackendService for Backend { .map(Into::into) } - fn restore_trash(&self, _input: Empty) -> Result { + fn restore_trash(&self, _input: pb::Empty) -> Result { let mut handler = self.new_progress_handler(); let progress_fn = move |progress| handler.update(Progress::MediaCheck(progress as u32), true); @@ -386,124 +384,6 @@ impl BackendService for Backend { .map(Into::into) } - // decks - //---------------------------------------------------- - - fn add_or_update_deck_legacy(&self, input: pb::AddOrUpdateDeckLegacyIn) -> Result { - self.with_col(|col| { - let schema11: DeckSchema11 = serde_json::from_slice(&input.deck)?; - let mut deck: Deck = schema11.into(); - if input.preserve_usn_and_mtime { - col.transact(None, |col| { - let usn = col.usn()?; - col.add_or_update_single_deck_with_existing_id(&mut deck, usn) - })?; - } else { - col.add_or_update_deck(&mut deck)?; - } - Ok(pb::DeckId { did: deck.id.0 }) - }) - } - - fn deck_tree(&self, input: pb::DeckTreeIn) -> Result { - let lim = if input.top_deck_id > 0 { - Some(DeckID(input.top_deck_id)) - } else { - None - }; - self.with_col(|col| { - let now = if input.now == 0 { - None - } else { - Some(TimestampSecs(input.now)) - }; - col.deck_tree(now, lim) - }) - } - - fn deck_tree_legacy(&self, _input: pb::Empty) -> Result { - self.with_col(|col| { - let tree = col.legacy_deck_tree()?; - serde_json::to_vec(&tree) - .map_err(Into::into) - .map(Into::into) - }) - } - - fn get_all_decks_legacy(&self, _input: Empty) -> Result { - self.with_col(|col| { - let decks = col.storage.get_all_decks_as_schema11()?; - serde_json::to_vec(&decks).map_err(Into::into) - }) - .map(Into::into) - } - - fn get_deck_id_by_name(&self, input: pb::String) -> Result { - self.with_col(|col| { - col.get_deck_id(&input.val).and_then(|d| { - d.ok_or(AnkiError::NotFound) - .map(|d| pb::DeckId { did: d.0 }) - }) - }) - } - - fn get_deck_legacy(&self, input: pb::DeckId) -> Result { - self.with_col(|col| { - let deck: DeckSchema11 = col - .storage - .get_deck(input.into())? - .ok_or(AnkiError::NotFound)? - .into(); - serde_json::to_vec(&deck) - .map_err(Into::into) - .map(Into::into) - }) - } - - fn get_deck_names(&self, input: pb::GetDeckNamesIn) -> Result { - self.with_col(|col| { - let names = if input.include_filtered { - col.get_all_deck_names(input.skip_empty_default)? - } else { - col.get_all_normal_deck_names()? - }; - Ok(pb::DeckNames { - entries: names - .into_iter() - .map(|(id, name)| pb::DeckNameId { id: id.0, name }) - .collect(), - }) - }) - } - - fn new_deck_legacy(&self, input: pb::Bool) -> Result { - let deck = if input.val { - Deck::new_filtered() - } else { - Deck::new_normal() - }; - let schema11: DeckSchema11 = deck.into(); - serde_json::to_vec(&schema11) - .map_err(Into::into) - .map(Into::into) - } - - fn remove_deck(&self, input: pb::DeckId) -> Result { - self.with_col(|col| col.remove_deck_and_child_decks(input.into())) - .map(Into::into) - } - - fn drag_drop_decks(&self, input: pb::DragDropDecksIn) -> Result { - let source_dids: Vec<_> = input.source_deck_ids.into_iter().map(Into::into).collect(); - let target_did = if input.target_deck_id == 0 { - None - } else { - Some(input.target_deck_id.into()) - }; - self.with_col(|col| col.drag_drop_decks(&source_dids, target_did)) - .map(Into::into) - } - // deck config //---------------------------------------------------- @@ -522,7 +402,7 @@ impl BackendService for Backend { .map(Into::into) } - fn all_deck_config_legacy(&self, _input: Empty) -> Result { + fn all_deck_config_legacy(&self, _input: pb::Empty) -> Result { self.with_col(|col| { let conf: Vec = col .storage @@ -544,13 +424,13 @@ impl BackendService for Backend { .map(Into::into) } - fn new_deck_config_legacy(&self, _input: Empty) -> Result { + fn new_deck_config_legacy(&self, _input: pb::Empty) -> Result { serde_json::to_vec(&DeckConfSchema11::default()) .map_err(Into::into) .map(Into::into) } - fn remove_deck_config(&self, input: pb::DeckConfigId) -> Result { + fn remove_deck_config(&self, input: pb::DeckConfigId) -> Result { self.with_col(|col| col.transact(None, |col| col.remove_deck_config(input.into()))) .map(Into::into) } @@ -567,7 +447,7 @@ impl BackendService for Backend { }) } - fn update_card(&self, input: pb::UpdateCardIn) -> Result { + fn update_card(&self, input: pb::UpdateCardIn) -> Result { self.with_col(|col| { let op = if input.skip_undo_entry { None @@ -580,7 +460,7 @@ impl BackendService for Backend { .map(Into::into) } - fn remove_cards(&self, input: pb::RemoveCardsIn) -> Result { + fn remove_cards(&self, input: pb::RemoveCardsIn) -> Result { self.with_col(|col| { col.transact(None, |col| { col.remove_cards_and_orphaned_notes( @@ -595,250 +475,16 @@ impl BackendService for Backend { }) } - fn set_deck(&self, input: pb::SetDeckIn) -> Result { + fn set_deck(&self, input: pb::SetDeckIn) -> Result { let cids: Vec<_> = input.card_ids.into_iter().map(CardID).collect(); let deck_id = input.deck_id.into(); self.with_col(|col| col.set_deck(&cids, deck_id).map(Into::into)) } - // notes - //------------------------------------------------------------------- - - fn new_note(&self, input: pb::NoteTypeId) -> Result { - self.with_col(|col| { - let nt = col.get_notetype(input.into())?.ok_or(AnkiError::NotFound)?; - Ok(nt.new_note().into()) - }) - } - - fn add_note(&self, input: pb::AddNoteIn) -> Result { - self.with_col(|col| { - let mut note: Note = input.note.ok_or(AnkiError::NotFound)?.into(); - col.add_note(&mut note, DeckID(input.deck_id)) - .map(|_| pb::NoteId { nid: note.id.0 }) - }) - } - - fn defaults_for_adding(&self, input: pb::DefaultsForAddingIn) -> Result { - self.with_col(|col| { - let home_deck: DeckID = input.home_deck_of_current_review_card.into(); - col.defaults_for_adding(home_deck).map(Into::into) - }) - } - - fn default_deck_for_notetype(&self, input: pb::NoteTypeId) -> Result { - self.with_col(|col| { - Ok(col - .default_deck_for_notetype(input.into())? - .unwrap_or(DeckID(0)) - .into()) - }) - } - - fn update_note(&self, input: pb::UpdateNoteIn) -> Result { - self.with_col(|col| { - let op = if input.skip_undo_entry { - None - } else { - Some(UndoableOpKind::UpdateNote) - }; - let mut note: Note = input.note.ok_or(AnkiError::NotFound)?.into(); - col.update_note_with_op(&mut note, op) - }) - .map(Into::into) - } - - fn get_note(&self, input: pb::NoteId) -> Result { - self.with_col(|col| { - col.storage - .get_note(input.into())? - .ok_or(AnkiError::NotFound) - .map(Into::into) - }) - } - - fn remove_notes(&self, input: pb::RemoveNotesIn) -> Result { - self.with_col(|col| { - if !input.note_ids.is_empty() { - col.remove_notes( - &input - .note_ids - .into_iter() - .map(Into::into) - .collect::>(), - )?; - } - if !input.card_ids.is_empty() { - let nids = col.storage.note_ids_of_cards( - &input - .card_ids - .into_iter() - .map(Into::into) - .collect::>(), - )?; - col.remove_notes(&nids.into_iter().collect::>())? - } - Ok(().into()) - }) - } - - fn add_note_tags(&self, input: pb::AddNoteTagsIn) -> Result { - self.with_col(|col| { - col.add_tags_to_notes(&to_nids(input.nids), &input.tags) - .map(|n| n as u32) - }) - .map(Into::into) - } - - fn update_note_tags(&self, input: pb::UpdateNoteTagsIn) -> Result { - self.with_col(|col| { - col.replace_tags_for_notes( - &to_nids(input.nids), - &input.tags, - &input.replacement, - input.regex, - ) - .map(|n| (n as u32).into()) - }) - } - - fn cloze_numbers_in_note(&self, note: pb::Note) -> Result { - let mut set = HashSet::with_capacity(4); - for field in ¬e.fields { - add_cloze_numbers_in_string(field, &mut set); - } - Ok(pb::ClozeNumbersInNoteOut { - numbers: set.into_iter().map(|n| n as u32).collect(), - }) - } - - fn after_note_updates(&self, input: pb::AfterNoteUpdatesIn) -> Result { - self.with_col(|col| { - col.transact(None, |col| { - col.after_note_updates( - &to_nids(input.nids), - input.generate_cards, - input.mark_notes_modified, - )?; - Ok(pb::Empty {}) - }) - }) - } - - fn field_names_for_notes( - &self, - input: pb::FieldNamesForNotesIn, - ) -> Result { - self.with_col(|col| { - let nids: Vec<_> = input.nids.into_iter().map(NoteID).collect(); - col.storage - .field_names_for_notes(&nids) - .map(|fields| pb::FieldNamesForNotesOut { fields }) - }) - } - - fn note_is_duplicate_or_empty(&self, input: pb::Note) -> Result { - let note: Note = input.into(); - self.with_col(|col| { - col.note_is_duplicate_or_empty(¬e) - .map(|r| pb::NoteIsDuplicateOrEmptyOut { state: r as i32 }) - }) - } - - fn cards_of_note(&self, input: pb::NoteId) -> Result { - self.with_col(|col| { - col.storage - .all_card_ids_of_note(NoteID(input.nid)) - .map(|v| pb::CardIDs { - cids: v.into_iter().map(Into::into).collect(), - }) - }) - } - - // notetypes - //------------------------------------------------------------------- - - fn add_or_update_notetype(&self, input: pb::AddOrUpdateNotetypeIn) -> Result { - self.with_col(|col| { - let legacy: NoteTypeSchema11 = serde_json::from_slice(&input.json)?; - let mut nt: NoteType = legacy.into(); - if nt.id.0 == 0 { - col.add_notetype(&mut nt)?; - } else { - col.update_notetype(&mut nt, input.preserve_usn_and_mtime)?; - } - Ok(pb::NoteTypeId { ntid: nt.id.0 }) - }) - } - - fn get_stock_notetype_legacy(&self, input: pb::StockNoteType) -> Result { - // fixme: use individual functions instead of full vec - let mut all = all_stock_notetypes(&self.i18n); - let idx = (input.kind as usize).min(all.len() - 1); - let nt = all.swap_remove(idx); - let schema11: NoteTypeSchema11 = nt.into(); - serde_json::to_vec(&schema11) - .map_err(Into::into) - .map(Into::into) - } - - fn get_notetype_legacy(&self, input: pb::NoteTypeId) -> Result { - self.with_col(|col| { - let schema11: NoteTypeSchema11 = col - .storage - .get_notetype(input.into())? - .ok_or(AnkiError::NotFound)? - .into(); - Ok(serde_json::to_vec(&schema11)?).map(Into::into) - }) - } - - fn get_notetype_names(&self, _input: Empty) -> Result { - self.with_col(|col| { - let entries: Vec<_> = col - .storage - .get_all_notetype_names()? - .into_iter() - .map(|(id, name)| pb::NoteTypeNameId { id: id.0, name }) - .collect(); - Ok(pb::NoteTypeNames { entries }) - }) - } - - fn get_notetype_names_and_counts(&self, _input: Empty) -> Result { - self.with_col(|col| { - let entries: Vec<_> = col - .storage - .get_notetype_use_counts()? - .into_iter() - .map(|(id, name, use_count)| pb::NoteTypeNameIdUseCount { - id: id.0, - name, - use_count, - }) - .collect(); - Ok(pb::NoteTypeUseCounts { entries }) - }) - } - - fn get_notetype_id_by_name(&self, input: pb::String) -> Result { - self.with_col(|col| { - col.storage - .get_notetype_id(&input.val) - .and_then(|nt| nt.ok_or(AnkiError::NotFound)) - .map(|ntid| pb::NoteTypeId { ntid: ntid.0 }) - }) - } - - fn remove_notetype(&self, input: pb::NoteTypeId) -> Result { - self.with_col(|col| col.remove_notetype(input.into())) - .map(Into::into) - } - // collection //------------------------------------------------------------------- - fn open_collection(&self, input: pb::OpenCollectionIn) -> Result { + fn open_collection(&self, input: pb::OpenCollectionIn) -> Result { let mut col = self.col.lock().unwrap(); if col.is_some() { return Err(AnkiError::CollectionAlreadyOpen); @@ -867,7 +513,7 @@ impl BackendService for Backend { Ok(().into()) } - fn close_collection(&self, input: pb::CloseCollectionIn) -> Result { + fn close_collection(&self, input: pb::CloseCollectionIn) -> Result { self.abort_media_sync_and_wait(); let mut col = self.col.lock().unwrap(); @@ -917,60 +563,6 @@ impl BackendService for Backend { }) } - // sync - //------------------------------------------------------------------- - - fn sync_media(&self, input: pb::SyncAuth) -> Result { - self.sync_media_inner(input).map(Into::into) - } - - fn abort_sync(&self, _input: Empty) -> Result { - if let Some(handle) = self.sync_abort.lock().unwrap().take() { - handle.abort(); - } - Ok(().into()) - } - - /// Abort the media sync. Does not wait for completion. - fn abort_media_sync(&self, _input: Empty) -> Result { - let guard = self.state.lock().unwrap(); - if let Some(handle) = &guard.media_sync_abort { - handle.abort(); - } - Ok(().into()) - } - - fn before_upload(&self, _input: Empty) -> Result { - self.with_col(|col| col.before_upload().map(Into::into)) - } - - fn sync_login(&self, input: pb::SyncLoginIn) -> Result { - self.sync_login_inner(input) - } - - fn sync_status(&self, input: pb::SyncAuth) -> Result { - self.sync_status_inner(input) - } - - fn sync_collection(&self, input: pb::SyncAuth) -> Result { - self.sync_collection_inner(input) - } - - fn full_upload(&self, input: pb::SyncAuth) -> Result { - self.full_sync_inner(input, true)?; - Ok(().into()) - } - - fn full_download(&self, input: pb::SyncAuth) -> Result { - self.full_sync_inner(input, false)?; - Ok(().into()) - } - - fn sync_server_method(&self, input: pb::SyncServerMethodIn) -> Result { - let req = SyncRequest::from_method_and_data(input.method(), input.data)?; - self.sync_server_method_inner(req).map(Into::into) - } - // i18n/messages //------------------------------------------------------------------- @@ -999,7 +591,7 @@ impl BackendService for Backend { .into()) } - fn i18n_resources(&self, _input: Empty) -> Result { + fn i18n_resources(&self, _input: pb::Empty) -> Result { serde_json::to_vec(&self.i18n.resources_for_js()) .map(Into::into) .map_err(Into::into) @@ -1021,7 +613,7 @@ impl BackendService for Backend { self.with_col(|col| col.transact(None, |col| col.clear_unused_tags().map(Into::into))) } - fn all_tags(&self, _input: Empty) -> Result { + fn all_tags(&self, _input: pb::Empty) -> Result { Ok(pb::StringList { vals: self.with_col(|col| { Ok(col @@ -1052,11 +644,11 @@ impl BackendService for Backend { }) } - fn tag_tree(&self, _input: Empty) -> Result { + fn tag_tree(&self, _input: pb::Empty) -> Result { self.with_col(|col| col.tag_tree()) } - fn drag_drop_tags(&self, input: pb::DragDropTagsIn) -> Result { + fn drag_drop_tags(&self, input: pb::DragDropTagsIn) -> Result { let source_tags = input.source_tags; let target_tag = if input.target_tag.is_empty() { None @@ -1066,79 +658,6 @@ impl BackendService for Backend { self.with_col(|col| col.drag_drop_tags(&source_tags, target_tag)) .map(Into::into) } - - // config/preferences - //------------------------------------------------------------------- - - fn get_config_json(&self, input: pb::String) -> Result { - self.with_col(|col| { - let val: Option = col.get_config_optional(input.val.as_str()); - val.ok_or(AnkiError::NotFound) - .and_then(|v| serde_json::to_vec(&v).map_err(Into::into)) - .map(Into::into) - }) - } - - fn set_config_json(&self, input: pb::SetConfigJsonIn) -> Result { - self.with_col(|col| { - col.transact(None, |col| { - // ensure it's a well-formed object - let val: JsonValue = serde_json::from_slice(&input.value_json)?; - col.set_config(input.key.as_str(), &val) - }) - }) - .map(Into::into) - } - - fn remove_config(&self, input: pb::String) -> Result { - self.with_col(|col| col.transact(None, |col| col.remove_config(input.val.as_str()))) - .map(Into::into) - } - - fn get_all_config(&self, _input: Empty) -> Result { - self.with_col(|col| { - let conf = col.storage.get_all_config()?; - serde_json::to_vec(&conf).map_err(Into::into) - }) - .map(Into::into) - } - - fn get_config_bool(&self, input: pb::config::Bool) -> Result { - self.with_col(|col| { - Ok(pb::Bool { - val: col.get_bool(input.key().into()), - }) - }) - } - - fn set_config_bool(&self, input: pb::SetConfigBoolIn) -> Result { - self.with_col(|col| col.transact(None, |col| col.set_bool(input.key().into(), input.value))) - .map(Into::into) - } - - fn get_config_string(&self, input: pb::config::String) -> Result { - self.with_col(|col| { - Ok(pb::String { - val: col.get_string(input.key().into()), - }) - }) - } - - fn set_config_string(&self, input: pb::SetConfigStringIn) -> Result { - self.with_col(|col| { - col.transact(None, |col| col.set_string(input.key().into(), &input.value)) - }) - .map(Into::into) - } - - fn get_preferences(&self, _input: Empty) -> Result { - self.with_col(|col| col.get_preferences()) - } - - fn set_preferences(&self, input: pb::Preferences) -> Result { - self.with_col(|col| col.set_preferences(input)) - .map(Into::into) - } } impl Backend { @@ -1172,6 +691,11 @@ impl Backend { .and_then(|service| match service { pb::ServiceIndex::Scheduling => SchedulingService::run_method(self, method, input), pb::ServiceIndex::Backend => BackendService::run_method(self, method, input), + pb::ServiceIndex::Decks => DecksService::run_method(self, method, input), + pb::ServiceIndex::Notes => NotesService::run_method(self, method, input), + pb::ServiceIndex::Notetypes => NoteTypesService::run_method(self, method, input), + pb::ServiceIndex::Config => ConfigService::run_method(self, method, input), + pb::ServiceIndex::Sync => SyncService::run_method(self, method, input), }) .map_err(|err| { let backend_err = anki_error_to_proto_error(err, &self.i18n); @@ -1225,10 +749,6 @@ impl Backend { } } -fn to_nids(ids: Vec) -> Vec { - ids.into_iter().map(NoteID).collect() -} - fn translate_arg_to_fluent_val(arg: &pb::TranslateArgValue) -> FluentValue { use pb::translate_arg_value::Value as V; match &arg.value { diff --git a/rslib/src/backend/notes.rs b/rslib/src/backend/notes.rs new file mode 100644 index 000000000..5c82ad5a4 --- /dev/null +++ b/rslib/src/backend/notes.rs @@ -0,0 +1,172 @@ +// Copyright: Ankitects Pty Ltd and contributors +// License: GNU AGPL, version 3 or later; http://www.gnu.org/licenses/agpl.html + +use std::collections::HashSet; + +use super::Backend; +use crate::{ + backend_proto::{self as pb}, + cloze::add_cloze_numbers_in_string, + prelude::*, +}; +pub(super) use pb::notes_service::Service as NotesService; + +impl NotesService for Backend { + // notes + //------------------------------------------------------------------- + + fn new_note(&self, input: pb::NoteTypeId) -> Result { + self.with_col(|col| { + let nt = col.get_notetype(input.into())?.ok_or(AnkiError::NotFound)?; + Ok(nt.new_note().into()) + }) + } + + fn add_note(&self, input: pb::AddNoteIn) -> Result { + self.with_col(|col| { + let mut note: Note = input.note.ok_or(AnkiError::NotFound)?.into(); + col.add_note(&mut note, DeckID(input.deck_id)) + .map(|_| pb::NoteId { nid: note.id.0 }) + }) + } + + fn defaults_for_adding(&self, input: pb::DefaultsForAddingIn) -> Result { + self.with_col(|col| { + let home_deck: DeckID = input.home_deck_of_current_review_card.into(); + col.defaults_for_adding(home_deck).map(Into::into) + }) + } + + fn default_deck_for_notetype(&self, input: pb::NoteTypeId) -> Result { + self.with_col(|col| { + Ok(col + .default_deck_for_notetype(input.into())? + .unwrap_or(DeckID(0)) + .into()) + }) + } + + fn update_note(&self, input: pb::UpdateNoteIn) -> Result { + self.with_col(|col| { + let op = if input.skip_undo_entry { + None + } else { + Some(UndoableOpKind::UpdateNote) + }; + let mut note: Note = input.note.ok_or(AnkiError::NotFound)?.into(); + col.update_note_with_op(&mut note, op) + }) + .map(Into::into) + } + + fn get_note(&self, input: pb::NoteId) -> Result { + self.with_col(|col| { + col.storage + .get_note(input.into())? + .ok_or(AnkiError::NotFound) + .map(Into::into) + }) + } + + fn remove_notes(&self, input: pb::RemoveNotesIn) -> Result { + self.with_col(|col| { + if !input.note_ids.is_empty() { + col.remove_notes( + &input + .note_ids + .into_iter() + .map(Into::into) + .collect::>(), + )?; + } + if !input.card_ids.is_empty() { + let nids = col.storage.note_ids_of_cards( + &input + .card_ids + .into_iter() + .map(Into::into) + .collect::>(), + )?; + col.remove_notes(&nids.into_iter().collect::>())? + } + Ok(().into()) + }) + } + + fn add_note_tags(&self, input: pb::AddNoteTagsIn) -> Result { + self.with_col(|col| { + col.add_tags_to_notes(&to_nids(input.nids), &input.tags) + .map(|n| n as u32) + }) + .map(Into::into) + } + + fn update_note_tags(&self, input: pb::UpdateNoteTagsIn) -> Result { + self.with_col(|col| { + col.replace_tags_for_notes( + &to_nids(input.nids), + &input.tags, + &input.replacement, + input.regex, + ) + .map(|n| (n as u32).into()) + }) + } + + fn cloze_numbers_in_note(&self, note: pb::Note) -> Result { + let mut set = HashSet::with_capacity(4); + for field in ¬e.fields { + add_cloze_numbers_in_string(field, &mut set); + } + Ok(pb::ClozeNumbersInNoteOut { + numbers: set.into_iter().map(|n| n as u32).collect(), + }) + } + + fn after_note_updates(&self, input: pb::AfterNoteUpdatesIn) -> Result { + self.with_col(|col| { + col.transact(None, |col| { + col.after_note_updates( + &to_nids(input.nids), + input.generate_cards, + input.mark_notes_modified, + )?; + Ok(pb::Empty {}) + }) + }) + } + + fn field_names_for_notes( + &self, + input: pb::FieldNamesForNotesIn, + ) -> Result { + self.with_col(|col| { + let nids: Vec<_> = input.nids.into_iter().map(NoteID).collect(); + col.storage + .field_names_for_notes(&nids) + .map(|fields| pb::FieldNamesForNotesOut { fields }) + }) + } + + fn note_is_duplicate_or_empty(&self, input: pb::Note) -> Result { + let note: Note = input.into(); + self.with_col(|col| { + col.note_is_duplicate_or_empty(¬e) + .map(|r| pb::NoteIsDuplicateOrEmptyOut { state: r as i32 }) + }) + } + + fn cards_of_note(&self, input: pb::NoteId) -> Result { + self.with_col(|col| { + col.storage + .all_card_ids_of_note(NoteID(input.nid)) + .map(|v| pb::CardIDs { + cids: v.into_iter().map(Into::into).collect(), + }) + }) + } +} + +fn to_nids(ids: Vec) -> Vec { + ids.into_iter().map(NoteID).collect() +} diff --git a/rslib/src/backend/notetypes.rs b/rslib/src/backend/notetypes.rs new file mode 100644 index 000000000..cab29ba03 --- /dev/null +++ b/rslib/src/backend/notetypes.rs @@ -0,0 +1,89 @@ +// Copyright: Ankitects Pty Ltd and contributors +// License: GNU AGPL, version 3 or later; http://www.gnu.org/licenses/agpl.html + +use super::Backend; +use crate::{ + backend_proto as pb, + notetype::{all_stock_notetypes, NoteType, NoteTypeSchema11}, + prelude::*, +}; +pub(super) use pb::notetypes_service::Service as NoteTypesService; + +impl NoteTypesService for Backend { + fn add_or_update_notetype(&self, input: pb::AddOrUpdateNotetypeIn) -> Result { + self.with_col(|col| { + let legacy: NoteTypeSchema11 = serde_json::from_slice(&input.json)?; + let mut nt: NoteType = legacy.into(); + if nt.id.0 == 0 { + col.add_notetype(&mut nt)?; + } else { + col.update_notetype(&mut nt, input.preserve_usn_and_mtime)?; + } + Ok(pb::NoteTypeId { ntid: nt.id.0 }) + }) + } + + fn get_stock_notetype_legacy(&self, input: pb::StockNoteType) -> Result { + // fixme: use individual functions instead of full vec + let mut all = all_stock_notetypes(&self.i18n); + let idx = (input.kind as usize).min(all.len() - 1); + let nt = all.swap_remove(idx); + let schema11: NoteTypeSchema11 = nt.into(); + serde_json::to_vec(&schema11) + .map_err(Into::into) + .map(Into::into) + } + + fn get_notetype_legacy(&self, input: pb::NoteTypeId) -> Result { + self.with_col(|col| { + let schema11: NoteTypeSchema11 = col + .storage + .get_notetype(input.into())? + .ok_or(AnkiError::NotFound)? + .into(); + Ok(serde_json::to_vec(&schema11)?).map(Into::into) + }) + } + + fn get_notetype_names(&self, _input: pb::Empty) -> Result { + self.with_col(|col| { + let entries: Vec<_> = col + .storage + .get_all_notetype_names()? + .into_iter() + .map(|(id, name)| pb::NoteTypeNameId { id: id.0, name }) + .collect(); + Ok(pb::NoteTypeNames { entries }) + }) + } + + fn get_notetype_names_and_counts(&self, _input: pb::Empty) -> Result { + self.with_col(|col| { + let entries: Vec<_> = col + .storage + .get_notetype_use_counts()? + .into_iter() + .map(|(id, name, use_count)| pb::NoteTypeNameIdUseCount { + id: id.0, + name, + use_count, + }) + .collect(); + Ok(pb::NoteTypeUseCounts { entries }) + }) + } + + fn get_notetype_id_by_name(&self, input: pb::String) -> Result { + self.with_col(|col| { + col.storage + .get_notetype_id(&input.val) + .and_then(|nt| nt.ok_or(AnkiError::NotFound)) + .map(|ntid| pb::NoteTypeId { ntid: ntid.0 }) + }) + } + + fn remove_notetype(&self, input: pb::NoteTypeId) -> Result { + self.with_col(|col| col.remove_notetype(input.into())) + .map(Into::into) + } +} diff --git a/rslib/src/backend/scheduler/mod.rs b/rslib/src/backend/scheduler/mod.rs index 73b0b1536..1e8a0a19c 100644 --- a/rslib/src/backend/scheduler/mod.rs +++ b/rslib/src/backend/scheduler/mod.rs @@ -15,7 +15,7 @@ use crate::{ }, stats::studied_today, }; -pub(super) use pb::scheduling::Service as SchedulingService; +pub(super) use pb::scheduling_service::Service as SchedulingService; impl SchedulingService for Backend { /// This behaves like _updateCutoff() in older code - it also unburies at the start of diff --git a/rslib/src/backend/sync.rs b/rslib/src/backend/sync/mod.rs similarity index 78% rename from rslib/src/backend/sync.rs rename to rslib/src/backend/sync/mod.rs index 17c1fd5b0..7d894ee0a 100644 --- a/rslib/src/backend/sync.rs +++ b/rslib/src/backend/sync/mod.rs @@ -1,6 +1,8 @@ // Copyright: Ankitects Pty Ltd and contributors // License: GNU AGPL, version 3 or later; http://www.gnu.org/licenses/agpl.html +mod server; + use std::sync::Arc; use futures::future::{AbortHandle, AbortRegistration, Abortable}; @@ -12,12 +14,20 @@ use crate::{ media::MediaManager, prelude::*, sync::{ - get_remote_sync_meta, sync_abort, sync_login, FullSyncProgress, NormalSyncProgress, - SyncActionRequired, SyncAuth, SyncMeta, SyncOutput, + get_remote_sync_meta, http::SyncRequest, sync_abort, sync_login, FullSyncProgress, + LocalServer, NormalSyncProgress, SyncActionRequired, SyncAuth, SyncMeta, SyncOutput, }, }; use super::{progress::AbortHandleSlot, Backend}; +pub(super) use pb::sync_service::Service as SyncService; + +#[derive(Default)] +pub(super) struct SyncState { + remote_sync_status: RemoteSyncStatus, + media_sync_abort: Option, + http_sync_server: Option, +} #[derive(Default, Debug)] pub(super) struct RemoteSyncStatus { @@ -70,6 +80,59 @@ impl From for SyncAuth { } } +impl SyncService for Backend { + fn sync_media(&self, input: pb::SyncAuth) -> Result { + self.sync_media_inner(input).map(Into::into) + } + + fn abort_sync(&self, _input: pb::Empty) -> Result { + if let Some(handle) = self.sync_abort.lock().unwrap().take() { + handle.abort(); + } + Ok(().into()) + } + + /// Abort the media sync. Does not wait for completion. + fn abort_media_sync(&self, _input: pb::Empty) -> Result { + let guard = self.state.lock().unwrap(); + if let Some(handle) = &guard.sync.media_sync_abort { + handle.abort(); + } + Ok(().into()) + } + + fn before_upload(&self, _input: pb::Empty) -> Result { + self.with_col(|col| col.before_upload().map(Into::into)) + } + + fn sync_login(&self, input: pb::SyncLoginIn) -> Result { + self.sync_login_inner(input) + } + + fn sync_status(&self, input: pb::SyncAuth) -> Result { + self.sync_status_inner(input) + } + + fn sync_collection(&self, input: pb::SyncAuth) -> Result { + self.sync_collection_inner(input) + } + + fn full_upload(&self, input: pb::SyncAuth) -> Result { + self.full_sync_inner(input, true)?; + Ok(().into()) + } + + fn full_download(&self, input: pb::SyncAuth) -> Result { + self.full_sync_inner(input, false)?; + Ok(().into()) + } + + fn sync_server_method(&self, input: pb::SyncServerMethodIn) -> Result { + let req = SyncRequest::from_method_and_data(input.method(), input.data)?; + self.sync_server_method_inner(req).map(Into::into) + } +} + impl Backend { fn sync_abort_handle( &self, @@ -104,11 +167,11 @@ impl Backend { let (abort_handle, abort_reg) = AbortHandle::new_pair(); { let mut guard = self.state.lock().unwrap(); - if guard.media_sync_abort.is_some() { + if guard.sync.media_sync_abort.is_some() { // media sync is already active return Ok(()); } else { - guard.media_sync_abort = Some(abort_handle); + guard.sync.media_sync_abort = Some(abort_handle); } } @@ -131,7 +194,7 @@ impl Backend { let result = rt.block_on(abortable_sync); // mark inactive - self.state.lock().unwrap().media_sync_abort.take(); + self.state.lock().unwrap().sync.media_sync_abort.take(); // return result match result { @@ -146,14 +209,14 @@ impl Backend { /// Abort the media sync. Won't return until aborted. pub(super) fn abort_media_sync_and_wait(&self) { let guard = self.state.lock().unwrap(); - if let Some(handle) = &guard.media_sync_abort { + if let Some(handle) = &guard.sync.media_sync_abort { handle.abort(); self.progress_state.lock().unwrap().want_abort = true; } drop(guard); // block until it aborts - while self.state.lock().unwrap().media_sync_abort.is_some() { + while self.state.lock().unwrap().sync.media_sync_abort.is_some() { std::thread::sleep(std::time::Duration::from_millis(100)); self.progress_state.lock().unwrap().want_abort = true; } @@ -185,8 +248,8 @@ impl Backend { // return cached server response if only a short time has elapsed { let guard = self.state.lock().unwrap(); - if guard.remote_sync_status.last_check.elapsed_secs() < 300 { - return Ok(guard.remote_sync_status.last_response.into()); + if guard.sync.remote_sync_status.last_check.elapsed_secs() < 300 { + return Ok(guard.sync.remote_sync_status.last_response.into()); } } @@ -201,9 +264,9 @@ impl Backend { // On startup, the sync status check will block on network access, and then automatic syncing begins, // taking hold of the mutex. By the time we reach here, our network status may be out of date, // so we discard it if stale. - if guard.remote_sync_status.last_check < time_at_check_begin { - guard.remote_sync_status.last_check = time_at_check_begin; - guard.remote_sync_status.last_response = response; + if guard.sync.remote_sync_status.last_check < time_at_check_begin { + guard.sync.remote_sync_status.last_check = time_at_check_begin; + guard.sync.remote_sync_status.last_response = response; } } @@ -247,6 +310,7 @@ impl Backend { self.state .lock() .unwrap() + .sync .remote_sync_status .update(output.required.into()); Ok(output.into()) @@ -302,6 +366,7 @@ impl Backend { self.state .lock() .unwrap() + .sync .remote_sync_status .update(pb::sync_status_out::Required::NoChanges); } diff --git a/rslib/src/backend/http_sync_server.rs b/rslib/src/backend/sync/server.rs similarity index 94% rename from rslib/src/backend/http_sync_server.rs rename to rslib/src/backend/sync/server.rs index efdc9f3ab..8b4816e9f 100644 --- a/rslib/src/backend/http_sync_server.rs +++ b/rslib/src/backend/sync/server.rs @@ -4,7 +4,7 @@ use std::{path::PathBuf, sync::MutexGuard}; use tokio::runtime::Runtime; -use super::{Backend, BackendState}; +use crate::backend::{Backend, BackendState}; use crate::{ err::SyncErrorKind, prelude::*, @@ -24,16 +24,12 @@ impl Backend { F: FnOnce(&mut LocalServer) -> Result, { let mut state_guard = self.state.lock().unwrap(); - let out = - func( - state_guard - .http_sync_server - .as_mut() - .ok_or_else(|| AnkiError::SyncError { - kind: SyncErrorKind::SyncNotStarted, - info: Default::default(), - })?, - ); + let out = func(state_guard.sync.http_sync_server.as_mut().ok_or_else(|| { + AnkiError::SyncError { + kind: SyncErrorKind::SyncNotStarted, + info: Default::default(), + } + })?); if out.is_err() { self.abort_and_restore_collection(Some(state_guard)) } @@ -82,6 +78,7 @@ impl Backend { fn take_server(&self, state_guard: Option>) -> Result { let mut state_guard = state_guard.unwrap_or_else(|| self.state.lock().unwrap()); state_guard + .sync .http_sync_server .take() .ok_or_else(|| AnkiError::SyncError { @@ -94,7 +91,7 @@ impl Backend { // place col into new server let server = self.col_into_server()?; let mut state_guard = self.state.lock().unwrap(); - assert!(state_guard.http_sync_server.replace(server).is_none()); + assert!(state_guard.sync.http_sync_server.replace(server).is_none()); drop(state_guard); self.with_sync_server(|server| {