mirror of
https://github.com/zed-industries/zed.git
synced 2024-12-26 18:41:10 +00:00
Avoid persisting project's state before it has been initialized
This commit is contained in:
parent
e18bc24989
commit
24aafde1e8
4 changed files with 202 additions and 27 deletions
|
@ -42,7 +42,7 @@ pub struct Contact {
|
||||||
pub projects: Vec<ProjectMetadata>,
|
pub projects: Vec<ProjectMetadata>,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug)]
|
#[derive(Clone, Debug, PartialEq)]
|
||||||
pub struct ProjectMetadata {
|
pub struct ProjectMetadata {
|
||||||
pub id: u64,
|
pub id: u64,
|
||||||
pub worktree_root_names: Vec<String>,
|
pub worktree_root_names: Vec<String>,
|
||||||
|
|
|
@ -7,7 +7,7 @@ use ::rpc::Peer;
|
||||||
use anyhow::anyhow;
|
use anyhow::anyhow;
|
||||||
use client::{
|
use client::{
|
||||||
self, proto, test::FakeHttpClient, Channel, ChannelDetails, ChannelList, Client, Connection,
|
self, proto, test::FakeHttpClient, Channel, ChannelDetails, ChannelList, Client, Connection,
|
||||||
Credentials, EstablishConnectionError, UserStore, RECEIVE_TIMEOUT,
|
Credentials, EstablishConnectionError, ProjectMetadata, UserStore, RECEIVE_TIMEOUT,
|
||||||
};
|
};
|
||||||
use collections::{BTreeMap, HashMap, HashSet};
|
use collections::{BTreeMap, HashMap, HashSet};
|
||||||
use editor::{
|
use editor::{
|
||||||
|
@ -479,8 +479,8 @@ async fn test_cancel_join_request(
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[gpui::test(iterations = 3)]
|
#[gpui::test(iterations = 10)]
|
||||||
async fn test_private_projects(
|
async fn test_offline_projects(
|
||||||
deterministic: Arc<Deterministic>,
|
deterministic: Arc<Deterministic>,
|
||||||
cx_a: &mut TestAppContext,
|
cx_a: &mut TestAppContext,
|
||||||
cx_b: &mut TestAppContext,
|
cx_b: &mut TestAppContext,
|
||||||
|
@ -489,58 +489,229 @@ async fn test_private_projects(
|
||||||
let mut server = TestServer::start(cx_a.foreground(), cx_a.background()).await;
|
let mut server = TestServer::start(cx_a.foreground(), cx_a.background()).await;
|
||||||
let client_a = server.create_client(cx_a, "user_a").await;
|
let client_a = server.create_client(cx_a, "user_a").await;
|
||||||
let client_b = server.create_client(cx_b, "user_b").await;
|
let client_b = server.create_client(cx_b, "user_b").await;
|
||||||
|
let user_a = UserId::from_proto(client_a.user_id().unwrap());
|
||||||
server
|
server
|
||||||
.make_contacts(vec![(&client_a, cx_a), (&client_b, cx_b)])
|
.make_contacts(vec![(&client_a, cx_a), (&client_b, cx_b)])
|
||||||
.await;
|
.await;
|
||||||
|
|
||||||
let user_a = UserId::from_proto(client_a.user_id().unwrap());
|
// Set up observers of the project and user stores. Any time either of
|
||||||
|
// these models update, they should be in a consistent state with each
|
||||||
|
// other. There should not be an observable moment where the current
|
||||||
|
// user's contact entry contains a project that does not match one of
|
||||||
|
// the current open projects. That would cause a duplicate entry to be
|
||||||
|
// shown in the contacts panel.
|
||||||
|
let mut subscriptions = vec![];
|
||||||
|
let (window_id, view) = cx_a.add_window(|cx| {
|
||||||
|
subscriptions.push(cx.observe(&client_a.user_store, {
|
||||||
|
let project_store = client_a.project_store.clone();
|
||||||
|
let user_store = client_a.user_store.clone();
|
||||||
|
move |_, _, cx| check_project_list(project_store.clone(), user_store.clone(), cx)
|
||||||
|
}));
|
||||||
|
|
||||||
let fs = FakeFs::new(cx_a.background());
|
subscriptions.push(cx.observe(&client_a.project_store, {
|
||||||
fs.insert_tree("/a", json!({})).await;
|
let project_store = client_a.project_store.clone();
|
||||||
|
let user_store = client_a.user_store.clone();
|
||||||
|
move |_, _, cx| check_project_list(project_store.clone(), user_store.clone(), cx)
|
||||||
|
}));
|
||||||
|
|
||||||
// Create a private project
|
fn check_project_list(
|
||||||
let project_a = cx_a.update(|cx| {
|
project_store: ModelHandle<ProjectStore>,
|
||||||
|
user_store: ModelHandle<UserStore>,
|
||||||
|
cx: &mut gpui::MutableAppContext,
|
||||||
|
) {
|
||||||
|
let open_project_ids = project_store
|
||||||
|
.read(cx)
|
||||||
|
.projects(cx)
|
||||||
|
.filter_map(|project| project.read(cx).remote_id())
|
||||||
|
.collect::<Vec<_>>();
|
||||||
|
|
||||||
|
let user_store = user_store.read(cx);
|
||||||
|
for contact in user_store.contacts() {
|
||||||
|
if contact.user.id == user_store.current_user().unwrap().id {
|
||||||
|
for project in &contact.projects {
|
||||||
|
if !open_project_ids.contains(&project.id) {
|
||||||
|
panic!(
|
||||||
|
concat!(
|
||||||
|
"current user's contact data has a project",
|
||||||
|
"that doesn't match any open project {:?}",
|
||||||
|
),
|
||||||
|
project
|
||||||
|
);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
EmptyView
|
||||||
|
});
|
||||||
|
|
||||||
|
// Build an offline project with two worktrees.
|
||||||
|
client_a
|
||||||
|
.fs
|
||||||
|
.insert_tree(
|
||||||
|
"/code",
|
||||||
|
json!({
|
||||||
|
"crate1": { "a.rs": "" },
|
||||||
|
"crate2": { "b.rs": "" },
|
||||||
|
}),
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
let project = cx_a.update(|cx| {
|
||||||
Project::local(
|
Project::local(
|
||||||
false,
|
false,
|
||||||
client_a.client.clone(),
|
client_a.client.clone(),
|
||||||
client_a.user_store.clone(),
|
client_a.user_store.clone(),
|
||||||
client_a.project_store.clone(),
|
client_a.project_store.clone(),
|
||||||
client_a.language_registry.clone(),
|
client_a.language_registry.clone(),
|
||||||
fs.clone(),
|
client_a.fs.clone(),
|
||||||
cx,
|
cx,
|
||||||
)
|
)
|
||||||
});
|
});
|
||||||
|
project
|
||||||
|
.update(cx_a, |p, cx| {
|
||||||
|
p.find_or_create_local_worktree("/code/crate1", true, cx)
|
||||||
|
})
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
project
|
||||||
|
.update(cx_a, |p, cx| {
|
||||||
|
p.find_or_create_local_worktree("/code/crate2", true, cx)
|
||||||
|
})
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
project
|
||||||
|
.update(cx_a, |p, cx| p.restore_state(cx))
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
// Private projects are not registered with the server.
|
// When a project is offline, no information about it is sent to the server.
|
||||||
deterministic.run_until_parked();
|
deterministic.run_until_parked();
|
||||||
assert!(project_a.read_with(cx_a, |project, _| project.remote_id().is_none()));
|
assert!(server
|
||||||
|
.store
|
||||||
|
.read()
|
||||||
|
.await
|
||||||
|
.project_metadata_for_user(user_a)
|
||||||
|
.is_empty());
|
||||||
|
assert!(project.read_with(cx_a, |project, _| project.remote_id().is_none()));
|
||||||
assert!(client_b
|
assert!(client_b
|
||||||
.user_store
|
.user_store
|
||||||
.read_with(cx_b, |store, _| { store.contacts()[0].projects.is_empty() }));
|
.read_with(cx_b, |store, _| { store.contacts()[0].projects.is_empty() }));
|
||||||
|
|
||||||
// The project is registered when it is made public.
|
// When the project is taken online, its metadata is sent to the server
|
||||||
project_a.update(cx_a, |project, cx| project.set_online(true, cx));
|
// and broadcasted to other users.
|
||||||
|
project.update(cx_a, |p, cx| p.set_online(true, cx));
|
||||||
deterministic.run_until_parked();
|
deterministic.run_until_parked();
|
||||||
assert!(project_a.read_with(cx_a, |project, _| project.remote_id().is_some()));
|
let project_id = project.read_with(cx_a, |p, _| p.remote_id()).unwrap();
|
||||||
assert!(!client_b
|
client_b.user_store.read_with(cx_b, |store, _| {
|
||||||
.user_store
|
assert_eq!(
|
||||||
.read_with(cx_b, |store, _| { store.contacts()[0].projects.is_empty() }));
|
store.contacts()[0].projects,
|
||||||
|
&[ProjectMetadata {
|
||||||
|
id: project_id,
|
||||||
|
worktree_root_names: vec!["crate1".into(), "crate2".into()],
|
||||||
|
guests: Default::default(),
|
||||||
|
}]
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
// The project is registered again when it loses and regains connection.
|
// The project is registered again when the host loses and regains connection.
|
||||||
server.disconnect_client(user_a);
|
server.disconnect_client(user_a);
|
||||||
server.forbid_connections();
|
server.forbid_connections();
|
||||||
cx_a.foreground().advance_clock(rpc::RECEIVE_TIMEOUT);
|
cx_a.foreground().advance_clock(rpc::RECEIVE_TIMEOUT);
|
||||||
// deterministic.run_until_parked();
|
assert!(server
|
||||||
assert!(project_a.read_with(cx_a, |project, _| project.remote_id().is_none()));
|
.store
|
||||||
|
.read()
|
||||||
|
.await
|
||||||
|
.project_metadata_for_user(user_a)
|
||||||
|
.is_empty());
|
||||||
|
assert!(project.read_with(cx_a, |p, _| p.remote_id().is_none()));
|
||||||
assert!(client_b
|
assert!(client_b
|
||||||
.user_store
|
.user_store
|
||||||
.read_with(cx_b, |store, _| { store.contacts()[0].projects.is_empty() }));
|
.read_with(cx_b, |store, _| { store.contacts()[0].projects.is_empty() }));
|
||||||
|
|
||||||
server.allow_connections();
|
server.allow_connections();
|
||||||
cx_b.foreground().advance_clock(Duration::from_secs(10));
|
cx_b.foreground().advance_clock(Duration::from_secs(10));
|
||||||
assert!(project_a.read_with(cx_a, |project, _| project.remote_id().is_some()));
|
let project_id = project.read_with(cx_a, |p, _| p.remote_id()).unwrap();
|
||||||
assert!(!client_b
|
client_b.user_store.read_with(cx_b, |store, _| {
|
||||||
.user_store
|
assert_eq!(
|
||||||
.read_with(cx_b, |store, _| { store.contacts()[0].projects.is_empty() }));
|
store.contacts()[0].projects,
|
||||||
|
&[ProjectMetadata {
|
||||||
|
id: project_id,
|
||||||
|
worktree_root_names: vec!["crate1".into(), "crate2".into()],
|
||||||
|
guests: Default::default(),
|
||||||
|
}]
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
project
|
||||||
|
.update(cx_a, |p, cx| {
|
||||||
|
p.find_or_create_local_worktree("/code/crate3", true, cx)
|
||||||
|
})
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
deterministic.run_until_parked();
|
||||||
|
client_b.user_store.read_with(cx_b, |store, _| {
|
||||||
|
assert_eq!(
|
||||||
|
store.contacts()[0].projects,
|
||||||
|
&[ProjectMetadata {
|
||||||
|
id: project_id,
|
||||||
|
worktree_root_names: vec!["crate1".into(), "crate2".into(), "crate3".into()],
|
||||||
|
guests: Default::default(),
|
||||||
|
}]
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
// Build another project using a directory which was previously part of
|
||||||
|
// an online project. Restore the project's state from the host's database.
|
||||||
|
let project2 = cx_a.update(|cx| {
|
||||||
|
Project::local(
|
||||||
|
false,
|
||||||
|
client_a.client.clone(),
|
||||||
|
client_a.user_store.clone(),
|
||||||
|
client_a.project_store.clone(),
|
||||||
|
client_a.language_registry.clone(),
|
||||||
|
client_a.fs.clone(),
|
||||||
|
cx,
|
||||||
|
)
|
||||||
|
});
|
||||||
|
project2
|
||||||
|
.update(cx_a, |p, cx| {
|
||||||
|
p.find_or_create_local_worktree("/code/crate3", true, cx)
|
||||||
|
})
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
project2
|
||||||
|
.update(cx_a, |project, cx| project.restore_state(cx))
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
// This project is now online, because its directory was previously online.
|
||||||
|
project2.read_with(cx_a, |project, _| assert!(project.is_online()));
|
||||||
|
deterministic.run_until_parked();
|
||||||
|
let project2_id = project2.read_with(cx_a, |p, _| p.remote_id()).unwrap();
|
||||||
|
client_b.user_store.read_with(cx_b, |store, _| {
|
||||||
|
assert_eq!(
|
||||||
|
store.contacts()[0].projects,
|
||||||
|
&[
|
||||||
|
ProjectMetadata {
|
||||||
|
id: project_id,
|
||||||
|
worktree_root_names: vec!["crate1".into(), "crate2".into(), "crate3".into()],
|
||||||
|
guests: Default::default(),
|
||||||
|
},
|
||||||
|
ProjectMetadata {
|
||||||
|
id: project2_id,
|
||||||
|
worktree_root_names: vec!["crate3".into()],
|
||||||
|
guests: Default::default(),
|
||||||
|
}
|
||||||
|
]
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
cx_a.update(|cx| {
|
||||||
|
drop(subscriptions);
|
||||||
|
drop(view);
|
||||||
|
cx.remove_window(window_id);
|
||||||
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
#[gpui::test(iterations = 10)]
|
#[gpui::test(iterations = 10)]
|
||||||
|
|
|
@ -32,7 +32,7 @@ pub struct Project {
|
||||||
#[serde(skip)]
|
#[serde(skip)]
|
||||||
pub join_requests: HashMap<UserId, Vec<Receipt<proto::JoinProject>>>,
|
pub join_requests: HashMap<UserId, Vec<Receipt<proto::JoinProject>>>,
|
||||||
pub active_replica_ids: HashSet<ReplicaId>,
|
pub active_replica_ids: HashSet<ReplicaId>,
|
||||||
pub worktrees: HashMap<u64, Worktree>,
|
pub worktrees: BTreeMap<u64, Worktree>,
|
||||||
pub language_servers: Vec<proto::LanguageServer>,
|
pub language_servers: Vec<proto::LanguageServer>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -99,6 +99,7 @@ pub struct Project {
|
||||||
opened_buffers: HashMap<u64, OpenBuffer>,
|
opened_buffers: HashMap<u64, OpenBuffer>,
|
||||||
buffer_snapshots: HashMap<u64, Vec<(i32, TextBufferSnapshot)>>,
|
buffer_snapshots: HashMap<u64, Vec<(i32, TextBufferSnapshot)>>,
|
||||||
nonce: u128,
|
nonce: u128,
|
||||||
|
initialized_persistent_state: bool,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Error, Debug)]
|
#[derive(Error, Debug)]
|
||||||
|
@ -385,6 +386,7 @@ impl Project {
|
||||||
language_server_settings: Default::default(),
|
language_server_settings: Default::default(),
|
||||||
next_language_server_id: 0,
|
next_language_server_id: 0,
|
||||||
nonce: StdRng::from_entropy().gen(),
|
nonce: StdRng::from_entropy().gen(),
|
||||||
|
initialized_persistent_state: false,
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
@ -497,6 +499,7 @@ impl Project {
|
||||||
opened_buffers: Default::default(),
|
opened_buffers: Default::default(),
|
||||||
buffer_snapshots: Default::default(),
|
buffer_snapshots: Default::default(),
|
||||||
nonce: StdRng::from_entropy().gen(),
|
nonce: StdRng::from_entropy().gen(),
|
||||||
|
initialized_persistent_state: false,
|
||||||
};
|
};
|
||||||
for worktree in worktrees {
|
for worktree in worktrees {
|
||||||
this.add_worktree(&worktree, cx);
|
this.add_worktree(&worktree, cx);
|
||||||
|
@ -566,6 +569,7 @@ impl Project {
|
||||||
cx.spawn(|this, mut cx| async move {
|
cx.spawn(|this, mut cx| async move {
|
||||||
let online = read_online.await.log_err().unwrap_or(false);
|
let online = read_online.await.log_err().unwrap_or(false);
|
||||||
this.update(&mut cx, |this, cx| {
|
this.update(&mut cx, |this, cx| {
|
||||||
|
this.initialized_persistent_state = true;
|
||||||
if let ProjectClientState::Local { online_tx, .. } = &mut this.client_state {
|
if let ProjectClientState::Local { online_tx, .. } = &mut this.client_state {
|
||||||
let mut online_tx = online_tx.borrow_mut();
|
let mut online_tx = online_tx.borrow_mut();
|
||||||
if *online_tx != online {
|
if *online_tx != online {
|
||||||
|
@ -580,7 +584,7 @@ impl Project {
|
||||||
}
|
}
|
||||||
|
|
||||||
fn persist_state(&mut self, cx: &mut ModelContext<Self>) -> Task<Result<()>> {
|
fn persist_state(&mut self, cx: &mut ModelContext<Self>) -> Task<Result<()>> {
|
||||||
if self.is_remote() {
|
if self.is_remote() || !self.initialized_persistent_state {
|
||||||
return Task::ready(Ok(()));
|
return Task::ready(Ok(()));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue