// Copyright 2020 The Jujutsu Authors // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // https://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #![allow(missing_docs)] use std::any::Any; use std::collections::{BTreeMap, HashMap, HashSet}; use std::fmt::Debug; use std::io::{ErrorKind, Write}; use std::path::{Path, PathBuf}; use std::time::SystemTime; use std::{fs, io}; use itertools::Itertools as _; use prost::Message; use tempfile::NamedTempFile; use thiserror::Error; use crate::backend::{CommitId, MillisSinceEpoch, Timestamp}; use crate::content_hash::blake2b_hash; use crate::file_util::{persist_content_addressed_temp_file, IoResultExt as _, PathError}; use crate::merge::Merge; use crate::object_id::{HexPrefix, ObjectId, PrefixResolution}; use crate::op_store::{ OpStore, OpStoreError, OpStoreResult, Operation, OperationId, OperationMetadata, RefTarget, RemoteRef, RemoteRefState, RemoteView, View, ViewId, WorkspaceId, }; use crate::{dag_walk, git, op_store}; // BLAKE2b-512 hash length in bytes const OPERATION_ID_LENGTH: usize = 64; const VIEW_ID_LENGTH: usize = 64; #[derive(Debug, Error)] #[error("Failed to read {kind} with ID {id}")] struct DecodeError { kind: &'static str, id: String, #[source] err: prost::DecodeError, } impl From for OpStoreError { fn from(err: DecodeError) -> Self { OpStoreError::Other(err.into()) } } #[derive(Debug)] pub struct SimpleOpStore { path: PathBuf, empty_view_id: ViewId, root_operation_id: OperationId, } impl SimpleOpStore { pub fn name() -> &'static str { "simple_op_store" } /// Creates an empty OpStore, panics if it already exists pub fn init(store_path: &Path) -> Self { fs::create_dir(store_path.join("views")).unwrap(); fs::create_dir(store_path.join("operations")).unwrap(); Self::load(store_path) } /// Load an existing OpStore pub fn load(store_path: &Path) -> Self { SimpleOpStore { path: store_path.to_path_buf(), empty_view_id: ViewId::from_bytes(&[0; VIEW_ID_LENGTH]), root_operation_id: OperationId::from_bytes(&[0; OPERATION_ID_LENGTH]), } } fn view_path(&self, id: &ViewId) -> PathBuf { self.path.join("views").join(id.hex()) } fn operation_path(&self, id: &OperationId) -> PathBuf { self.path.join("operations").join(id.hex()) } } impl OpStore for SimpleOpStore { fn as_any(&self) -> &dyn Any { self } fn name(&self) -> &str { Self::name() } fn root_operation_id(&self) -> &OperationId { &self.root_operation_id } fn read_view(&self, id: &ViewId) -> OpStoreResult { if *id == self.empty_view_id { return Ok(View::default()); } let path = self.view_path(id); let buf = fs::read(path).map_err(|err| io_to_read_error(err, id))?; let proto = crate::protos::op_store::View::decode(&*buf).map_err(|err| DecodeError { kind: "view", id: id.hex(), err, })?; Ok(view_from_proto(proto)) } fn write_view(&self, view: &View) -> OpStoreResult { let temp_file = NamedTempFile::new_in(&self.path).map_err(|err| io_to_write_error(err, "view"))?; let proto = view_to_proto(view); temp_file .as_file() .write_all(&proto.encode_to_vec()) .map_err(|err| io_to_write_error(err, "view"))?; let id = ViewId::new(blake2b_hash(view).to_vec()); persist_content_addressed_temp_file(temp_file, self.view_path(&id)) .map_err(|err| io_to_write_error(err, "view"))?; Ok(id) } fn read_operation(&self, id: &OperationId) -> OpStoreResult { if *id == self.root_operation_id { return Ok(Operation::make_root(self.empty_view_id.clone())); } let path = self.operation_path(id); let buf = fs::read(path).map_err(|err| io_to_read_error(err, id))?; let proto = crate::protos::op_store::Operation::decode(&*buf).map_err(|err| DecodeError { kind: "operation", id: id.hex(), err, })?; let mut operation = operation_from_proto(proto); if operation.parents.is_empty() { // Repos created before we had the root operation will have an operation without // parents. operation.parents.push(self.root_operation_id.clone()); } Ok(operation) } fn write_operation(&self, operation: &Operation) -> OpStoreResult { assert!(!operation.parents.is_empty()); let temp_file = NamedTempFile::new_in(&self.path).map_err(|err| io_to_write_error(err, "operation"))?; let proto = operation_to_proto(operation); temp_file .as_file() .write_all(&proto.encode_to_vec()) .map_err(|err| io_to_write_error(err, "operation"))?; let id = OperationId::new(blake2b_hash(operation).to_vec()); persist_content_addressed_temp_file(temp_file, self.operation_path(&id)) .map_err(|err| io_to_write_error(err, "operation"))?; Ok(id) } fn resolve_operation_id_prefix( &self, prefix: &HexPrefix, ) -> OpStoreResult> { let op_dir = self.path.join("operations"); let find = || -> io::Result<_> { let matches_root = prefix.matches(&self.root_operation_id); let hex_prefix = prefix.hex(); if hex_prefix.len() == OPERATION_ID_LENGTH * 2 { // Fast path for full-length ID if matches_root || op_dir.join(hex_prefix).try_exists()? { let id = OperationId::from_bytes(prefix.as_full_bytes().unwrap()); return Ok(PrefixResolution::SingleMatch(id)); } else { return Ok(PrefixResolution::NoMatch); } } let mut matched = matches_root.then(|| self.root_operation_id.clone()); for entry in op_dir.read_dir()? { let Ok(name) = entry?.file_name().into_string() else { continue; // Skip invalid UTF-8 }; if !name.starts_with(&hex_prefix) { continue; } let Ok(id) = OperationId::try_from_hex(&name) else { continue; // Skip invalid hex }; if matched.is_some() { return Ok(PrefixResolution::AmbiguousMatch); } matched = Some(id); } if let Some(id) = matched { Ok(PrefixResolution::SingleMatch(id)) } else { Ok(PrefixResolution::NoMatch) } }; find() .context(&op_dir) .map_err(|err| OpStoreError::Other(err.into())) } #[tracing::instrument(skip(self))] fn gc(&self, head_ids: &[OperationId], keep_newer: SystemTime) -> OpStoreResult<()> { let to_op_id = |entry: &fs::DirEntry| -> Option { let name = entry.file_name().into_string().ok()?; OperationId::try_from_hex(&name).ok() }; let to_view_id = |entry: &fs::DirEntry| -> Option { let name = entry.file_name().into_string().ok()?; ViewId::try_from_hex(&name).ok() }; let remove_file_if_not_new = |entry: &fs::DirEntry| -> Result<(), PathError> { let path = entry.path(); // Check timestamp, but there's still TOCTOU problem if an existing // file is renewed. let metadata = entry.metadata().context(&path)?; let mtime = metadata.modified().expect("unsupported platform?"); if mtime > keep_newer { tracing::trace!(?path, "not removing"); Ok(()) } else { tracing::trace!(?path, "removing"); fs::remove_file(&path).context(&path) } }; // Reachable objects are resolved without considering the keep_newer // parameter. We could collect ancestors of the "new" operations here, // but more files can be added anyway after that. let read_op = |id: &OperationId| self.read_operation(id).map(|data| (id.clone(), data)); let reachable_ops: HashMap = dag_walk::dfs_ok( head_ids.iter().map(read_op), |(id, _)| id.clone(), |(_, data)| data.parents.iter().map(read_op).collect_vec(), ) .try_collect()?; let reachable_views: HashSet<&ViewId> = reachable_ops.values().map(|data| &data.view_id).collect(); tracing::info!( reachable_op_count = reachable_ops.len(), reachable_view_count = reachable_views.len(), "collected reachable objects" ); let prune_ops = || -> Result<(), PathError> { let op_dir = self.path.join("operations"); for entry in op_dir.read_dir().context(&op_dir)? { let entry = entry.context(&op_dir)?; let Some(id) = to_op_id(&entry) else { tracing::trace!(?entry, "skipping invalid file name"); continue; }; if reachable_ops.contains_key(&id) { continue; } // If the operation was added after collecting reachable_views, // its view mtime would also be renewed. So there's no need to // update the reachable_views set to preserve the view. remove_file_if_not_new(&entry)?; } Ok(()) }; prune_ops().map_err(|err| OpStoreError::Other(err.into()))?; let prune_views = || -> Result<(), PathError> { let view_dir = self.path.join("views"); for entry in view_dir.read_dir().context(&view_dir)? { let entry = entry.context(&view_dir)?; let Some(id) = to_view_id(&entry) else { tracing::trace!(?entry, "skipping invalid file name"); continue; }; if reachable_views.contains(&id) { continue; } remove_file_if_not_new(&entry)?; } Ok(()) }; prune_views().map_err(|err| OpStoreError::Other(err.into()))?; Ok(()) } } fn io_to_read_error(err: std::io::Error, id: &impl ObjectId) -> OpStoreError { if err.kind() == ErrorKind::NotFound { OpStoreError::ObjectNotFound { object_type: id.object_type(), hash: id.hex(), source: Box::new(err), } } else { OpStoreError::ReadObject { object_type: id.object_type(), hash: id.hex(), source: Box::new(err), } } } fn io_to_write_error(err: std::io::Error, object_type: &'static str) -> OpStoreError { OpStoreError::WriteObject { object_type, source: Box::new(err), } } fn timestamp_to_proto(timestamp: &Timestamp) -> crate::protos::op_store::Timestamp { crate::protos::op_store::Timestamp { millis_since_epoch: timestamp.timestamp.0, tz_offset: timestamp.tz_offset, } } fn timestamp_from_proto(proto: crate::protos::op_store::Timestamp) -> Timestamp { Timestamp { timestamp: MillisSinceEpoch(proto.millis_since_epoch), tz_offset: proto.tz_offset, } } fn operation_metadata_to_proto( metadata: &OperationMetadata, ) -> crate::protos::op_store::OperationMetadata { crate::protos::op_store::OperationMetadata { start_time: Some(timestamp_to_proto(&metadata.start_time)), end_time: Some(timestamp_to_proto(&metadata.end_time)), description: metadata.description.clone(), hostname: metadata.hostname.clone(), username: metadata.username.clone(), is_snapshot: metadata.is_snapshot, tags: metadata.tags.clone(), } } fn operation_metadata_from_proto( proto: crate::protos::op_store::OperationMetadata, ) -> OperationMetadata { let start_time = timestamp_from_proto(proto.start_time.unwrap_or_default()); let end_time = timestamp_from_proto(proto.end_time.unwrap_or_default()); OperationMetadata { start_time, end_time, description: proto.description, hostname: proto.hostname, username: proto.username, is_snapshot: proto.is_snapshot, tags: proto.tags, } } fn operation_to_proto(operation: &Operation) -> crate::protos::op_store::Operation { let mut proto = crate::protos::op_store::Operation { view_id: operation.view_id.as_bytes().to_vec(), metadata: Some(operation_metadata_to_proto(&operation.metadata)), ..Default::default() }; for parent in &operation.parents { proto.parents.push(parent.to_bytes()); } proto } fn operation_from_proto(proto: crate::protos::op_store::Operation) -> Operation { let parents = proto.parents.into_iter().map(OperationId::new).collect(); let view_id = ViewId::new(proto.view_id); let metadata = operation_metadata_from_proto(proto.metadata.unwrap_or_default()); Operation { view_id, parents, metadata, } } fn view_to_proto(view: &View) -> crate::protos::op_store::View { let mut proto = crate::protos::op_store::View { // New/loaded view should have been migrated to the latest format has_git_refs_migrated_to_remote: true, ..Default::default() }; for (workspace_id, commit_id) in &view.wc_commit_ids { proto .wc_commit_ids .insert(workspace_id.as_str().to_string(), commit_id.to_bytes()); } for head_id in &view.head_ids { proto.head_ids.push(head_id.to_bytes()); } proto.branches = branch_views_to_proto_legacy(&view.local_branches, &view.remote_views); for (name, target) in &view.tags { proto.tags.push(crate::protos::op_store::Tag { name: name.clone(), target: ref_target_to_proto(target), }); } for (git_ref_name, target) in &view.git_refs { proto.git_refs.push(crate::protos::op_store::GitRef { name: git_ref_name.clone(), target: ref_target_to_proto(target), ..Default::default() }); } proto.git_head = ref_target_to_proto(&view.git_head); proto } fn view_from_proto(proto: crate::protos::op_store::View) -> View { let mut view = View::default(); // For compatibility with old repos before we had support for multiple working // copies #[allow(deprecated)] if !proto.wc_commit_id.is_empty() { view.wc_commit_ids .insert(WorkspaceId::default(), CommitId::new(proto.wc_commit_id)); } for (workspace_id, commit_id) in proto.wc_commit_ids { view.wc_commit_ids .insert(WorkspaceId::new(workspace_id), CommitId::new(commit_id)); } for head_id_bytes in proto.head_ids { view.head_ids.insert(CommitId::new(head_id_bytes)); } let (local_branches, remote_views) = branch_views_from_proto_legacy(proto.branches); view.local_branches = local_branches; view.remote_views = remote_views; for tag_proto in proto.tags { view.tags .insert(tag_proto.name, ref_target_from_proto(tag_proto.target)); } for git_ref in proto.git_refs { let target = if git_ref.target.is_some() { ref_target_from_proto(git_ref.target) } else { // Legacy format RefTarget::normal(CommitId::new(git_ref.commit_id)) }; view.git_refs.insert(git_ref.name, target); } #[allow(deprecated)] if proto.git_head.is_some() { view.git_head = ref_target_from_proto(proto.git_head); } else if !proto.git_head_legacy.is_empty() { view.git_head = RefTarget::normal(CommitId::new(proto.git_head_legacy)); } if !proto.has_git_refs_migrated_to_remote { migrate_git_refs_to_remote(&mut view); } view } fn branch_views_to_proto_legacy( local_branches: &BTreeMap, remote_views: &BTreeMap, ) -> Vec { op_store::merge_join_branch_views(local_branches, remote_views) .map(|(name, branch_target)| { let local_target = ref_target_to_proto(branch_target.local_target); let remote_branches = branch_target .remote_refs .iter() .map( |&(remote_name, remote_ref)| crate::protos::op_store::RemoteBranch { remote_name: remote_name.to_owned(), target: ref_target_to_proto(&remote_ref.target), state: remote_ref_state_to_proto(remote_ref.state), }, ) .collect(); crate::protos::op_store::Branch { name: name.to_owned(), local_target, remote_branches, } }) .collect() } fn branch_views_from_proto_legacy( branches_legacy: Vec, ) -> (BTreeMap, BTreeMap) { let mut local_branches: BTreeMap = BTreeMap::new(); let mut remote_views: BTreeMap = BTreeMap::new(); for branch_proto in branches_legacy { let local_target = ref_target_from_proto(branch_proto.local_target); for remote_branch in branch_proto.remote_branches { let state = remote_ref_state_from_proto(remote_branch.state).unwrap_or_else(|| { // If local branch doesn't exist, we assume that the remote branch hasn't been // merged because git.auto-local-branch was off. That's probably more common // than deleted but yet-to-be-pushed local branch. Alternatively, we could read // git.auto-local-branch setting here, but that wouldn't always work since the // setting could be toggled after the branch got merged. let is_git_tracking = remote_branch.remote_name == git::REMOTE_NAME_FOR_LOCAL_GIT_REPO; let default_state = if is_git_tracking || local_target.is_present() { RemoteRefState::Tracking } else { RemoteRefState::New }; tracing::trace!( ?branch_proto.name, ?remote_branch.remote_name, ?default_state, "generated tracking state", ); default_state }); let remote_view = remote_views.entry(remote_branch.remote_name).or_default(); let remote_ref = RemoteRef { target: ref_target_from_proto(remote_branch.target), state, }; remote_view .branches .insert(branch_proto.name.clone(), remote_ref); } if local_target.is_present() { local_branches.insert(branch_proto.name, local_target); } } (local_branches, remote_views) } fn migrate_git_refs_to_remote(view: &mut View) { if view.git_refs.is_empty() { // Not a repo backed by Git? return; } tracing::info!("migrating Git-tracking branches"); let mut git_view = RemoteView::default(); for (full_name, target) in &view.git_refs { if let Some(name) = full_name.strip_prefix("refs/heads/") { assert!(!name.is_empty()); let remote_ref = RemoteRef { target: target.clone(), // Git-tracking branches should never be untracked. state: RemoteRefState::Tracking, }; git_view.branches.insert(name.to_owned(), remote_ref); } } view.remote_views .insert(git::REMOTE_NAME_FOR_LOCAL_GIT_REPO.to_owned(), git_view); // jj < 0.9 might have imported refs from remote named "git" let reserved_git_ref_prefix = format!("refs/remotes/{}/", git::REMOTE_NAME_FOR_LOCAL_GIT_REPO); view.git_refs .retain(|name, _| !name.starts_with(&reserved_git_ref_prefix)); } fn ref_target_to_proto(value: &RefTarget) -> Option { let term_to_proto = |term: &Option| crate::protos::op_store::ref_conflict::Term { value: term.as_ref().map(|id| id.to_bytes()), }; let merge = value.as_merge(); let conflict_proto = crate::protos::op_store::RefConflict { removes: merge.removes().map(term_to_proto).collect(), adds: merge.adds().map(term_to_proto).collect(), }; let proto = crate::protos::op_store::RefTarget { value: Some(crate::protos::op_store::ref_target::Value::Conflict( conflict_proto, )), }; Some(proto) } #[allow(deprecated)] #[cfg(test)] fn ref_target_to_proto_legacy(value: &RefTarget) -> Option { if let Some(id) = value.as_normal() { let proto = crate::protos::op_store::RefTarget { value: Some(crate::protos::op_store::ref_target::Value::CommitId( id.to_bytes(), )), }; Some(proto) } else if value.has_conflict() { let ref_conflict_proto = crate::protos::op_store::RefConflictLegacy { removes: value.removed_ids().map(|id| id.to_bytes()).collect(), adds: value.added_ids().map(|id| id.to_bytes()).collect(), }; let proto = crate::protos::op_store::RefTarget { value: Some(crate::protos::op_store::ref_target::Value::ConflictLegacy( ref_conflict_proto, )), }; Some(proto) } else { assert!(value.is_absent()); None } } fn ref_target_from_proto(maybe_proto: Option) -> RefTarget { // TODO: Delete legacy format handling when we decide to drop support for views // saved by jj <= 0.8. let Some(proto) = maybe_proto else { // Legacy absent id return RefTarget::absent(); }; match proto.value.unwrap() { #[allow(deprecated)] crate::protos::op_store::ref_target::Value::CommitId(id) => { // Legacy non-conflicting id RefTarget::normal(CommitId::new(id)) } #[allow(deprecated)] crate::protos::op_store::ref_target::Value::ConflictLegacy(conflict) => { // Legacy conflicting ids let removes = conflict.removes.into_iter().map(CommitId::new); let adds = conflict.adds.into_iter().map(CommitId::new); RefTarget::from_legacy_form(removes, adds) } crate::protos::op_store::ref_target::Value::Conflict(conflict) => { let term_from_proto = |term: crate::protos::op_store::ref_conflict::Term| term.value.map(CommitId::new); let removes = conflict.removes.into_iter().map(term_from_proto); let adds = conflict.adds.into_iter().map(term_from_proto); RefTarget::from_merge(Merge::from_removes_adds(removes, adds)) } } } fn remote_ref_state_to_proto(state: RemoteRefState) -> Option { let proto_state = match state { RemoteRefState::New => crate::protos::op_store::RemoteRefState::New, RemoteRefState::Tracking => crate::protos::op_store::RemoteRefState::Tracking, }; Some(proto_state as i32) } fn remote_ref_state_from_proto(proto_value: Option) -> Option { let proto_state = proto_value?.try_into().ok()?; let state = match proto_state { crate::protos::op_store::RemoteRefState::New => RemoteRefState::New, crate::protos::op_store::RemoteRefState::Tracking => RemoteRefState::Tracking, }; Some(state) } #[cfg(test)] mod tests { use insta::assert_snapshot; use itertools::Itertools as _; use maplit::{btreemap, hashmap, hashset}; use super::*; fn create_view() -> View { let new_remote_ref = |target: &RefTarget| RemoteRef { target: target.clone(), state: RemoteRefState::New, }; let tracking_remote_ref = |target: &RefTarget| RemoteRef { target: target.clone(), state: RemoteRefState::Tracking, }; let head_id1 = CommitId::from_hex("aaa111"); let head_id2 = CommitId::from_hex("aaa222"); let branch_main_local_target = RefTarget::normal(CommitId::from_hex("ccc111")); let branch_main_origin_target = RefTarget::normal(CommitId::from_hex("ccc222")); let branch_deleted_origin_target = RefTarget::normal(CommitId::from_hex("ccc333")); let tag_v1_target = RefTarget::normal(CommitId::from_hex("ddd111")); let git_refs_main_target = RefTarget::normal(CommitId::from_hex("fff111")); let git_refs_feature_target = RefTarget::from_legacy_form( [CommitId::from_hex("fff111")], [CommitId::from_hex("fff222"), CommitId::from_hex("fff333")], ); let default_wc_commit_id = CommitId::from_hex("abc111"); let test_wc_commit_id = CommitId::from_hex("abc222"); View { head_ids: hashset! {head_id1, head_id2}, local_branches: btreemap! { "main".to_string() => branch_main_local_target, }, tags: btreemap! { "v1.0".to_string() => tag_v1_target, }, remote_views: btreemap! { "origin".to_string() => RemoteView { branches: btreemap! { "main".to_string() => tracking_remote_ref(&branch_main_origin_target), "deleted".to_string() => new_remote_ref(&branch_deleted_origin_target), }, }, }, git_refs: btreemap! { "refs/heads/main".to_string() => git_refs_main_target, "refs/heads/feature".to_string() => git_refs_feature_target, }, git_head: RefTarget::normal(CommitId::from_hex("fff111")), wc_commit_ids: hashmap! { WorkspaceId::default() => default_wc_commit_id, WorkspaceId::new("test".to_string()) => test_wc_commit_id, }, } } fn create_operation() -> Operation { Operation { view_id: ViewId::from_hex("aaa111"), parents: vec![ OperationId::from_hex("bbb111"), OperationId::from_hex("bbb222"), ], metadata: OperationMetadata { start_time: Timestamp { timestamp: MillisSinceEpoch(123456789), tz_offset: 3600, }, end_time: Timestamp { timestamp: MillisSinceEpoch(123456800), tz_offset: 3600, }, description: "check out foo".to_string(), hostname: "some.host.example.com".to_string(), username: "someone".to_string(), is_snapshot: false, tags: hashmap! { "key1".to_string() => "value1".to_string(), "key2".to_string() => "value2".to_string(), }, }, } } #[test] fn test_hash_view() { // Test exact output so we detect regressions in compatibility assert_snapshot!( ViewId::new(blake2b_hash(&create_view()).to_vec()).hex(), @"f426676b3a2f7c6b9ec8677cb05ed249d0d244ab7e86a7c51117e2d8a4829db65e55970c761231e2107d303bf3d33a1f2afdd4ed2181f223e99753674b20a35e" ); } #[test] fn test_hash_operation() { // Test exact output so we detect regressions in compatibility assert_snapshot!( OperationId::new(blake2b_hash(&create_operation()).to_vec()).hex(), @"20b495d54aa3be3a672a2ed6dbbf7a711dabce4cc0161d657e5177070491c1e780eec3fd35c2aa9dcc22371462aeb412a502a847f29419e65718f56a0ad1b2d0" ); } #[test] fn test_read_write_view() { let temp_dir = testutils::new_temp_dir(); let store = SimpleOpStore::init(temp_dir.path()); let view = create_view(); let view_id = store.write_view(&view).unwrap(); let read_view = store.read_view(&view_id).unwrap(); assert_eq!(read_view, view); } #[test] fn test_read_write_operation() { let temp_dir = testutils::new_temp_dir(); let store = SimpleOpStore::init(temp_dir.path()); let operation = create_operation(); let op_id = store.write_operation(&operation).unwrap(); let read_operation = store.read_operation(&op_id).unwrap(); assert_eq!(read_operation, operation); } #[test] fn test_branch_views_legacy_roundtrip() { let new_remote_ref = |target: &RefTarget| RemoteRef { target: target.clone(), state: RemoteRefState::New, }; let tracking_remote_ref = |target: &RefTarget| RemoteRef { target: target.clone(), state: RemoteRefState::Tracking, }; let local_branch1_target = RefTarget::normal(CommitId::from_hex("111111")); let local_branch3_target = RefTarget::normal(CommitId::from_hex("222222")); let git_branch1_target = RefTarget::normal(CommitId::from_hex("333333")); let remote1_branch1_target = RefTarget::normal(CommitId::from_hex("444444")); let remote2_branch2_target = RefTarget::normal(CommitId::from_hex("555555")); let remote2_branch4_target = RefTarget::normal(CommitId::from_hex("666666")); let local_branches = btreemap! { "branch1".to_owned() => local_branch1_target.clone(), "branch3".to_owned() => local_branch3_target.clone(), }; let remote_views = btreemap! { "git".to_owned() => RemoteView { branches: btreemap! { "branch1".to_owned() => tracking_remote_ref(&git_branch1_target), }, }, "remote1".to_owned() => RemoteView { branches: btreemap! { "branch1".to_owned() => tracking_remote_ref(&remote1_branch1_target), }, }, "remote2".to_owned() => RemoteView { branches: btreemap! { // "branch2" is non-tracking. "branch4" is tracking, but locally deleted. "branch2".to_owned() => new_remote_ref(&remote2_branch2_target), "branch4".to_owned() => tracking_remote_ref(&remote2_branch4_target), }, }, }; let branches_legacy = branch_views_to_proto_legacy(&local_branches, &remote_views); assert_eq!( branches_legacy .iter() .map(|proto| &proto.name) .sorted() .collect_vec(), vec!["branch1", "branch2", "branch3", "branch4"], ); let (local_branches_reconstructed, remote_views_reconstructed) = branch_views_from_proto_legacy(branches_legacy); assert_eq!(local_branches_reconstructed, local_branches); assert_eq!(remote_views_reconstructed, remote_views); } #[test] fn test_migrate_git_refs_remote_named_git() { let normal_ref_target = |id_hex| RefTarget::normal(CommitId::from_hex(id_hex)); let normal_new_remote_ref = |id_hex| RemoteRef { target: normal_ref_target(id_hex), state: RemoteRefState::New, }; let normal_tracking_remote_ref = |id_hex| RemoteRef { target: normal_ref_target(id_hex), state: RemoteRefState::Tracking, }; let branch_to_proto = |name: &str, local_ref_target, remote_branches| crate::protos::op_store::Branch { name: name.to_owned(), local_target: ref_target_to_proto(local_ref_target), remote_branches, }; let remote_branch_to_proto = |remote_name: &str, ref_target| crate::protos::op_store::RemoteBranch { remote_name: remote_name.to_owned(), target: ref_target_to_proto(ref_target), state: None, // to be generated based on local branch existence }; let git_ref_to_proto = |name: &str, ref_target| crate::protos::op_store::GitRef { name: name.to_owned(), target: ref_target_to_proto(ref_target), ..Default::default() }; let proto = crate::protos::op_store::View { branches: vec![ branch_to_proto( "main", &normal_ref_target("111111"), vec![ remote_branch_to_proto("git", &normal_ref_target("222222")), remote_branch_to_proto("gita", &normal_ref_target("333333")), ], ), branch_to_proto( "untracked", RefTarget::absent_ref(), vec![remote_branch_to_proto("gita", &normal_ref_target("777777"))], ), ], git_refs: vec![ git_ref_to_proto("refs/heads/main", &normal_ref_target("444444")), git_ref_to_proto("refs/remotes/git/main", &normal_ref_target("555555")), git_ref_to_proto("refs/remotes/gita/main", &normal_ref_target("666666")), git_ref_to_proto("refs/remotes/gita/untracked", &normal_ref_target("888888")), ], has_git_refs_migrated_to_remote: false, ..Default::default() }; let view = view_from_proto(proto); assert_eq!( view.local_branches, btreemap! { "main".to_owned() => normal_ref_target("111111"), }, ); assert_eq!( view.remote_views, btreemap! { "git".to_owned() => RemoteView { branches: btreemap! { "main".to_owned() => normal_tracking_remote_ref("444444"), // refs/heads/main }, }, "gita".to_owned() => RemoteView { branches: btreemap! { "main".to_owned() => normal_tracking_remote_ref("333333"), "untracked".to_owned() => normal_new_remote_ref("777777"), }, }, }, ); assert_eq!( view.git_refs, btreemap! { "refs/heads/main".to_owned() => normal_ref_target("444444"), "refs/remotes/gita/main".to_owned() => normal_ref_target("666666"), "refs/remotes/gita/untracked".to_owned() => normal_ref_target("888888"), }, ); // Once migrated, "git" remote branches shouldn't be populated again. let mut proto = view_to_proto(&view); assert!(proto.has_git_refs_migrated_to_remote); proto.branches.clear(); let view = view_from_proto(proto); assert!(!view.remote_views.contains_key("git")); } #[test] fn test_ref_target_change_delete_order_roundtrip() { let target = RefTarget::from_merge(Merge::from_removes_adds( vec![Some(CommitId::from_hex("111111"))], vec![Some(CommitId::from_hex("222222")), None], )); let maybe_proto = ref_target_to_proto(&target); assert_eq!(ref_target_from_proto(maybe_proto), target); // If it were legacy format, order of None entry would be lost. let target = RefTarget::from_merge(Merge::from_removes_adds( vec![Some(CommitId::from_hex("111111"))], vec![None, Some(CommitId::from_hex("222222"))], )); let maybe_proto = ref_target_to_proto(&target); assert_eq!(ref_target_from_proto(maybe_proto), target); } #[test] fn test_ref_target_legacy_roundtrip() { let target = RefTarget::absent(); let maybe_proto = ref_target_to_proto_legacy(&target); assert_eq!(ref_target_from_proto(maybe_proto), target); let target = RefTarget::normal(CommitId::from_hex("111111")); let maybe_proto = ref_target_to_proto_legacy(&target); assert_eq!(ref_target_from_proto(maybe_proto), target); // N-way conflict let target = RefTarget::from_legacy_form( [CommitId::from_hex("111111"), CommitId::from_hex("222222")], [ CommitId::from_hex("333333"), CommitId::from_hex("444444"), CommitId::from_hex("555555"), ], ); let maybe_proto = ref_target_to_proto_legacy(&target); assert_eq!(ref_target_from_proto(maybe_proto), target); // Change-delete conflict let target = RefTarget::from_legacy_form( [CommitId::from_hex("111111")], [CommitId::from_hex("222222")], ); let maybe_proto = ref_target_to_proto_legacy(&target); assert_eq!(ref_target_from_proto(maybe_proto), target); } }