mirror of
https://github.com/zed-industries/zed.git
synced 2024-12-26 10:40:54 +00:00
Merge branch 'main' into namespace-actions
This commit is contained in:
commit
ed2b690b9e
10 changed files with 523 additions and 405 deletions
3
Cargo.lock
generated
3
Cargo.lock
generated
|
@ -4007,11 +4007,11 @@ dependencies = [
|
||||||
"async-tungstenite",
|
"async-tungstenite",
|
||||||
"base64 0.13.0",
|
"base64 0.13.0",
|
||||||
"clock",
|
"clock",
|
||||||
|
"collections",
|
||||||
"futures",
|
"futures",
|
||||||
"gpui",
|
"gpui",
|
||||||
"log",
|
"log",
|
||||||
"parking_lot",
|
"parking_lot",
|
||||||
"postage",
|
|
||||||
"prost",
|
"prost",
|
||||||
"prost-build",
|
"prost-build",
|
||||||
"rand 0.8.3",
|
"rand 0.8.3",
|
||||||
|
@ -6122,7 +6122,6 @@ dependencies = [
|
||||||
"oauth2",
|
"oauth2",
|
||||||
"oauth2-surf",
|
"oauth2-surf",
|
||||||
"parking_lot",
|
"parking_lot",
|
||||||
"postage",
|
|
||||||
"project",
|
"project",
|
||||||
"rand 0.8.3",
|
"rand 0.8.3",
|
||||||
"rpc",
|
"rpc",
|
||||||
|
|
|
@ -6,7 +6,6 @@ use anyhow::{anyhow, Result};
|
||||||
use futures::{future::BoxFuture, stream::BoxStream, Future, StreamExt};
|
use futures::{future::BoxFuture, stream::BoxStream, Future, StreamExt};
|
||||||
use gpui::{executor, ModelHandle, TestAppContext};
|
use gpui::{executor, ModelHandle, TestAppContext};
|
||||||
use parking_lot::Mutex;
|
use parking_lot::Mutex;
|
||||||
use postage::barrier;
|
|
||||||
use rpc::{proto, ConnectionId, Peer, Receipt, TypedEnvelope};
|
use rpc::{proto, ConnectionId, Peer, Receipt, TypedEnvelope};
|
||||||
use std::{fmt, rc::Rc, sync::Arc};
|
use std::{fmt, rc::Rc, sync::Arc};
|
||||||
|
|
||||||
|
@ -23,7 +22,6 @@ struct FakeServerState {
|
||||||
connection_id: Option<ConnectionId>,
|
connection_id: Option<ConnectionId>,
|
||||||
forbid_connections: bool,
|
forbid_connections: bool,
|
||||||
auth_count: usize,
|
auth_count: usize,
|
||||||
connection_killer: Option<barrier::Sender>,
|
|
||||||
access_token: usize,
|
access_token: usize,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -76,15 +74,13 @@ impl FakeServer {
|
||||||
Err(EstablishConnectionError::Unauthorized)?
|
Err(EstablishConnectionError::Unauthorized)?
|
||||||
}
|
}
|
||||||
|
|
||||||
let (client_conn, server_conn, kill) =
|
let (client_conn, server_conn, _) = Connection::in_memory(cx.background());
|
||||||
Connection::in_memory(cx.background());
|
|
||||||
let (connection_id, io, incoming) =
|
let (connection_id, io, incoming) =
|
||||||
peer.add_test_connection(server_conn, cx.background()).await;
|
peer.add_test_connection(server_conn, cx.background()).await;
|
||||||
cx.background().spawn(io).detach();
|
cx.background().spawn(io).detach();
|
||||||
let mut state = state.lock();
|
let mut state = state.lock();
|
||||||
state.connection_id = Some(connection_id);
|
state.connection_id = Some(connection_id);
|
||||||
state.incoming = Some(incoming);
|
state.incoming = Some(incoming);
|
||||||
state.connection_killer = Some(kill);
|
|
||||||
Ok(client_conn)
|
Ok(client_conn)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
|
@ -1216,7 +1216,7 @@ impl Project {
|
||||||
let file = File::from_dyn(buffer.file())?;
|
let file = File::from_dyn(buffer.file())?;
|
||||||
let abs_path = file.as_local()?.abs_path(cx);
|
let abs_path = file.as_local()?.abs_path(cx);
|
||||||
let uri = lsp::Url::from_file_path(abs_path).unwrap();
|
let uri = lsp::Url::from_file_path(abs_path).unwrap();
|
||||||
let buffer_snapshots = self.buffer_snapshots.entry(buffer.remote_id()).or_default();
|
let buffer_snapshots = self.buffer_snapshots.get_mut(&buffer.remote_id())?;
|
||||||
let (version, prev_snapshot) = buffer_snapshots.last()?;
|
let (version, prev_snapshot) = buffer_snapshots.last()?;
|
||||||
let next_snapshot = buffer.text_snapshot();
|
let next_snapshot = buffer.text_snapshot();
|
||||||
let next_version = version + 1;
|
let next_version = version + 1;
|
||||||
|
@ -3850,7 +3850,7 @@ impl Project {
|
||||||
let buffer = this
|
let buffer = this
|
||||||
.opened_buffers
|
.opened_buffers
|
||||||
.get(&buffer_id)
|
.get(&buffer_id)
|
||||||
.map(|buffer| buffer.upgrade(cx).unwrap())
|
.and_then(|buffer| buffer.upgrade(cx))
|
||||||
.ok_or_else(|| anyhow!("unknown buffer id {}", buffer_id))?;
|
.ok_or_else(|| anyhow!("unknown buffer id {}", buffer_id))?;
|
||||||
Ok::<_, anyhow::Error>((project_id, buffer))
|
Ok::<_, anyhow::Error>((project_id, buffer))
|
||||||
})?;
|
})?;
|
||||||
|
@ -3882,7 +3882,7 @@ impl Project {
|
||||||
buffers.insert(
|
buffers.insert(
|
||||||
this.opened_buffers
|
this.opened_buffers
|
||||||
.get(buffer_id)
|
.get(buffer_id)
|
||||||
.map(|buffer| buffer.upgrade(cx).unwrap())
|
.and_then(|buffer| buffer.upgrade(cx))
|
||||||
.ok_or_else(|| anyhow!("unknown buffer id {}", buffer_id))?,
|
.ok_or_else(|| anyhow!("unknown buffer id {}", buffer_id))?,
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
@ -3911,7 +3911,7 @@ impl Project {
|
||||||
buffers.insert(
|
buffers.insert(
|
||||||
this.opened_buffers
|
this.opened_buffers
|
||||||
.get(buffer_id)
|
.get(buffer_id)
|
||||||
.map(|buffer| buffer.upgrade(cx).unwrap())
|
.and_then(|buffer| buffer.upgrade(cx))
|
||||||
.ok_or_else(|| anyhow!("unknown buffer id {}", buffer_id))?,
|
.ok_or_else(|| anyhow!("unknown buffer id {}", buffer_id))?,
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
@ -3942,7 +3942,7 @@ impl Project {
|
||||||
let buffer = this.read_with(&cx, |this, cx| {
|
let buffer = this.read_with(&cx, |this, cx| {
|
||||||
this.opened_buffers
|
this.opened_buffers
|
||||||
.get(&envelope.payload.buffer_id)
|
.get(&envelope.payload.buffer_id)
|
||||||
.map(|buffer| buffer.upgrade(cx).unwrap())
|
.and_then(|buffer| buffer.upgrade(cx))
|
||||||
.ok_or_else(|| anyhow!("unknown buffer id {}", envelope.payload.buffer_id))
|
.ok_or_else(|| anyhow!("unknown buffer id {}", envelope.payload.buffer_id))
|
||||||
})?;
|
})?;
|
||||||
buffer
|
buffer
|
||||||
|
@ -3972,7 +3972,7 @@ impl Project {
|
||||||
let buffer = this
|
let buffer = this
|
||||||
.opened_buffers
|
.opened_buffers
|
||||||
.get(&envelope.payload.buffer_id)
|
.get(&envelope.payload.buffer_id)
|
||||||
.map(|buffer| buffer.upgrade(cx).unwrap())
|
.and_then(|buffer| buffer.upgrade(cx))
|
||||||
.ok_or_else(|| anyhow!("unknown buffer id {}", envelope.payload.buffer_id))?;
|
.ok_or_else(|| anyhow!("unknown buffer id {}", envelope.payload.buffer_id))?;
|
||||||
let language = buffer.read(cx).language();
|
let language = buffer.read(cx).language();
|
||||||
let completion = language::proto::deserialize_completion(
|
let completion = language::proto::deserialize_completion(
|
||||||
|
@ -4014,7 +4014,7 @@ impl Project {
|
||||||
let buffer = this.update(&mut cx, |this, cx| {
|
let buffer = this.update(&mut cx, |this, cx| {
|
||||||
this.opened_buffers
|
this.opened_buffers
|
||||||
.get(&envelope.payload.buffer_id)
|
.get(&envelope.payload.buffer_id)
|
||||||
.map(|buffer| buffer.upgrade(cx).unwrap())
|
.and_then(|buffer| buffer.upgrade(cx))
|
||||||
.ok_or_else(|| anyhow!("unknown buffer id {}", envelope.payload.buffer_id))
|
.ok_or_else(|| anyhow!("unknown buffer id {}", envelope.payload.buffer_id))
|
||||||
})?;
|
})?;
|
||||||
buffer
|
buffer
|
||||||
|
@ -4055,7 +4055,7 @@ impl Project {
|
||||||
let buffer = this
|
let buffer = this
|
||||||
.opened_buffers
|
.opened_buffers
|
||||||
.get(&envelope.payload.buffer_id)
|
.get(&envelope.payload.buffer_id)
|
||||||
.map(|buffer| buffer.upgrade(cx).unwrap())
|
.and_then(|buffer| buffer.upgrade(cx))
|
||||||
.ok_or_else(|| anyhow!("unknown buffer id {}", envelope.payload.buffer_id))?;
|
.ok_or_else(|| anyhow!("unknown buffer id {}", envelope.payload.buffer_id))?;
|
||||||
Ok::<_, anyhow::Error>(this.apply_code_action(buffer, action, false, cx))
|
Ok::<_, anyhow::Error>(this.apply_code_action(buffer, action, false, cx))
|
||||||
})?;
|
})?;
|
||||||
|
|
|
@ -11,7 +11,10 @@ use client::{proto, Client, TypedEnvelope};
|
||||||
use clock::ReplicaId;
|
use clock::ReplicaId;
|
||||||
use collections::HashMap;
|
use collections::HashMap;
|
||||||
use futures::{
|
use futures::{
|
||||||
channel::mpsc::{self, UnboundedSender},
|
channel::{
|
||||||
|
mpsc::{self, UnboundedSender},
|
||||||
|
oneshot,
|
||||||
|
},
|
||||||
Stream, StreamExt,
|
Stream, StreamExt,
|
||||||
};
|
};
|
||||||
use fuzzy::CharBag;
|
use fuzzy::CharBag;
|
||||||
|
@ -26,7 +29,6 @@ use language::{
|
||||||
use lazy_static::lazy_static;
|
use lazy_static::lazy_static;
|
||||||
use parking_lot::Mutex;
|
use parking_lot::Mutex;
|
||||||
use postage::{
|
use postage::{
|
||||||
oneshot,
|
|
||||||
prelude::{Sink as _, Stream as _},
|
prelude::{Sink as _, Stream as _},
|
||||||
watch,
|
watch,
|
||||||
};
|
};
|
||||||
|
@ -727,11 +729,11 @@ impl LocalWorktree {
|
||||||
|
|
||||||
pub fn share(&mut self, project_id: u64, cx: &mut ModelContext<Worktree>) -> Task<Result<()>> {
|
pub fn share(&mut self, project_id: u64, cx: &mut ModelContext<Worktree>) -> Task<Result<()>> {
|
||||||
let register = self.register(project_id, cx);
|
let register = self.register(project_id, cx);
|
||||||
let (mut share_tx, mut share_rx) = oneshot::channel();
|
let (share_tx, share_rx) = oneshot::channel();
|
||||||
let (snapshots_to_send_tx, snapshots_to_send_rx) =
|
let (snapshots_to_send_tx, snapshots_to_send_rx) =
|
||||||
smol::channel::unbounded::<LocalSnapshot>();
|
smol::channel::unbounded::<LocalSnapshot>();
|
||||||
if self.share.is_some() {
|
if self.share.is_some() {
|
||||||
let _ = share_tx.try_send(Ok(()));
|
let _ = share_tx.send(Ok(()));
|
||||||
} else {
|
} else {
|
||||||
let rpc = self.client.clone();
|
let rpc = self.client.clone();
|
||||||
let worktree_id = cx.model_id() as u64;
|
let worktree_id = cx.model_id() as u64;
|
||||||
|
@ -756,15 +758,15 @@ impl LocalWorktree {
|
||||||
})
|
})
|
||||||
.await
|
.await
|
||||||
{
|
{
|
||||||
let _ = share_tx.try_send(Err(error));
|
let _ = share_tx.send(Err(error));
|
||||||
return Err(anyhow!("failed to send initial update worktree"));
|
return Err(anyhow!("failed to send initial update worktree"));
|
||||||
} else {
|
} else {
|
||||||
let _ = share_tx.try_send(Ok(()));
|
let _ = share_tx.send(Ok(()));
|
||||||
snapshot
|
snapshot
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
Err(error) => {
|
Err(error) => {
|
||||||
let _ = share_tx.try_send(Err(error.into()));
|
let _ = share_tx.send(Err(error.into()));
|
||||||
return Err(anyhow!("failed to send initial update worktree"));
|
return Err(anyhow!("failed to send initial update worktree"));
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
@ -804,9 +806,8 @@ impl LocalWorktree {
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
share_rx
|
share_rx
|
||||||
.next()
|
|
||||||
.await
|
.await
|
||||||
.unwrap_or_else(|| Err(anyhow!("share ended")))
|
.unwrap_or_else(|_| Err(anyhow!("share ended")))
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -9,9 +9,13 @@ path = "src/rpc.rs"
|
||||||
doctest = false
|
doctest = false
|
||||||
|
|
||||||
[features]
|
[features]
|
||||||
test-support = ["gpui/test-support"]
|
test-support = ["collections/test-support", "gpui/test-support"]
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
|
clock = { path = "../clock" }
|
||||||
|
collections = { path = "../collections" }
|
||||||
|
gpui = { path = "../gpui", optional = true }
|
||||||
|
util = { path = "../util" }
|
||||||
anyhow = "1.0"
|
anyhow = "1.0"
|
||||||
async-lock = "2.4"
|
async-lock = "2.4"
|
||||||
async-tungstenite = "0.16"
|
async-tungstenite = "0.16"
|
||||||
|
@ -19,21 +23,18 @@ base64 = "0.13"
|
||||||
futures = "0.3"
|
futures = "0.3"
|
||||||
log = "0.4"
|
log = "0.4"
|
||||||
parking_lot = "0.11.1"
|
parking_lot = "0.11.1"
|
||||||
postage = { version = "0.4.1", features = ["futures-traits"] }
|
|
||||||
prost = "0.8"
|
prost = "0.8"
|
||||||
rand = "0.8"
|
rand = "0.8"
|
||||||
rsa = "0.4"
|
rsa = "0.4"
|
||||||
serde = { version = "1", features = ["derive"] }
|
serde = { version = "1", features = ["derive"] }
|
||||||
smol-timeout = "0.6"
|
smol-timeout = "0.6"
|
||||||
zstd = "0.9"
|
zstd = "0.9"
|
||||||
clock = { path = "../clock" }
|
|
||||||
gpui = { path = "../gpui", optional = true }
|
|
||||||
util = { path = "../util" }
|
|
||||||
|
|
||||||
[build-dependencies]
|
[build-dependencies]
|
||||||
prost-build = "0.8"
|
prost-build = "0.8"
|
||||||
|
|
||||||
[dev-dependencies]
|
[dev-dependencies]
|
||||||
|
collections = { path = "../collections", features = ["test-support"] }
|
||||||
gpui = { path = "../gpui", features = ["test-support"] }
|
gpui = { path = "../gpui", features = ["test-support"] }
|
||||||
smol = "1.2.5"
|
smol = "1.2.5"
|
||||||
tempdir = "0.3.7"
|
tempdir = "0.3.7"
|
||||||
|
|
|
@ -35,21 +35,24 @@ impl Connection {
|
||||||
#[cfg(any(test, feature = "test-support"))]
|
#[cfg(any(test, feature = "test-support"))]
|
||||||
pub fn in_memory(
|
pub fn in_memory(
|
||||||
executor: std::sync::Arc<gpui::executor::Background>,
|
executor: std::sync::Arc<gpui::executor::Background>,
|
||||||
) -> (Self, Self, postage::barrier::Sender) {
|
) -> (Self, Self, std::sync::Arc<std::sync::atomic::AtomicBool>) {
|
||||||
use postage::prelude::Stream;
|
use std::sync::{
|
||||||
|
atomic::{AtomicBool, Ordering::SeqCst},
|
||||||
|
Arc,
|
||||||
|
};
|
||||||
|
|
||||||
let (kill_tx, kill_rx) = postage::barrier::channel();
|
let killed = Arc::new(AtomicBool::new(false));
|
||||||
let (a_tx, a_rx) = channel(kill_rx.clone(), executor.clone());
|
let (a_tx, a_rx) = channel(killed.clone(), executor.clone());
|
||||||
let (b_tx, b_rx) = channel(kill_rx, executor);
|
let (b_tx, b_rx) = channel(killed.clone(), executor);
|
||||||
return (
|
return (
|
||||||
Self { tx: a_tx, rx: b_rx },
|
Self { tx: a_tx, rx: b_rx },
|
||||||
Self { tx: b_tx, rx: a_rx },
|
Self { tx: b_tx, rx: a_rx },
|
||||||
kill_tx,
|
killed,
|
||||||
);
|
);
|
||||||
|
|
||||||
fn channel(
|
fn channel(
|
||||||
kill_rx: postage::barrier::Receiver,
|
killed: Arc<AtomicBool>,
|
||||||
executor: std::sync::Arc<gpui::executor::Background>,
|
executor: Arc<gpui::executor::Background>,
|
||||||
) -> (
|
) -> (
|
||||||
Box<dyn Send + Unpin + futures::Sink<WebSocketMessage, Error = WebSocketError>>,
|
Box<dyn Send + Unpin + futures::Sink<WebSocketMessage, Error = WebSocketError>>,
|
||||||
Box<
|
Box<
|
||||||
|
@ -57,20 +60,17 @@ impl Connection {
|
||||||
>,
|
>,
|
||||||
) {
|
) {
|
||||||
use futures::channel::mpsc;
|
use futures::channel::mpsc;
|
||||||
use std::{
|
use std::io::{Error, ErrorKind};
|
||||||
io::{Error, ErrorKind},
|
|
||||||
sync::Arc,
|
|
||||||
};
|
|
||||||
|
|
||||||
let (tx, rx) = mpsc::unbounded::<WebSocketMessage>();
|
let (tx, rx) = mpsc::unbounded::<WebSocketMessage>();
|
||||||
|
|
||||||
let tx = tx
|
let tx = tx
|
||||||
.sink_map_err(|e| WebSocketError::from(Error::new(ErrorKind::Other, e)))
|
.sink_map_err(|e| WebSocketError::from(Error::new(ErrorKind::Other, e)))
|
||||||
.with({
|
.with({
|
||||||
let kill_rx = kill_rx.clone();
|
let killed = killed.clone();
|
||||||
let executor = Arc::downgrade(&executor);
|
let executor = Arc::downgrade(&executor);
|
||||||
move |msg| {
|
move |msg| {
|
||||||
let mut kill_rx = kill_rx.clone();
|
let killed = killed.clone();
|
||||||
let executor = executor.clone();
|
let executor = executor.clone();
|
||||||
Box::pin(async move {
|
Box::pin(async move {
|
||||||
if let Some(executor) = executor.upgrade() {
|
if let Some(executor) = executor.upgrade() {
|
||||||
|
@ -78,7 +78,7 @@ impl Connection {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Writes to a half-open TCP connection will error.
|
// Writes to a half-open TCP connection will error.
|
||||||
if kill_rx.try_recv().is_ok() {
|
if killed.load(SeqCst) {
|
||||||
std::io::Result::Err(
|
std::io::Result::Err(
|
||||||
Error::new(ErrorKind::Other, "connection lost").into(),
|
Error::new(ErrorKind::Other, "connection lost").into(),
|
||||||
)?;
|
)?;
|
||||||
|
@ -90,10 +90,10 @@ impl Connection {
|
||||||
});
|
});
|
||||||
|
|
||||||
let rx = rx.then({
|
let rx = rx.then({
|
||||||
let kill_rx = kill_rx.clone();
|
let killed = killed.clone();
|
||||||
let executor = Arc::downgrade(&executor);
|
let executor = Arc::downgrade(&executor);
|
||||||
move |msg| {
|
move |msg| {
|
||||||
let mut kill_rx = kill_rx.clone();
|
let killed = killed.clone();
|
||||||
let executor = executor.clone();
|
let executor = executor.clone();
|
||||||
Box::pin(async move {
|
Box::pin(async move {
|
||||||
if let Some(executor) = executor.upgrade() {
|
if let Some(executor) = executor.upgrade() {
|
||||||
|
@ -101,7 +101,7 @@ impl Connection {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Reads from a half-open TCP connection will hang.
|
// Reads from a half-open TCP connection will hang.
|
||||||
if kill_rx.try_recv().is_ok() {
|
if killed.load(SeqCst) {
|
||||||
futures::future::pending::<()>().await;
|
futures::future::pending::<()>().await;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1,16 +1,18 @@
|
||||||
use super::proto::{self, AnyTypedEnvelope, EnvelopedMessage, MessageStream, RequestMessage};
|
use super::{
|
||||||
use super::Connection;
|
proto::{self, AnyTypedEnvelope, EnvelopedMessage, MessageStream, RequestMessage},
|
||||||
use anyhow::{anyhow, Context, Result};
|
Connection,
|
||||||
use futures::{channel::oneshot, stream::BoxStream, FutureExt as _, StreamExt};
|
|
||||||
use parking_lot::{Mutex, RwLock};
|
|
||||||
use postage::{
|
|
||||||
barrier, mpsc,
|
|
||||||
prelude::{Sink as _, Stream as _},
|
|
||||||
};
|
};
|
||||||
use smol_timeout::TimeoutExt as _;
|
use anyhow::{anyhow, Context, Result};
|
||||||
|
use collections::HashMap;
|
||||||
|
use futures::{
|
||||||
|
channel::{mpsc, oneshot},
|
||||||
|
stream::BoxStream,
|
||||||
|
FutureExt, SinkExt, StreamExt,
|
||||||
|
};
|
||||||
|
use parking_lot::{Mutex, RwLock};
|
||||||
|
use smol_timeout::TimeoutExt;
|
||||||
use std::sync::atomic::Ordering::SeqCst;
|
use std::sync::atomic::Ordering::SeqCst;
|
||||||
use std::{
|
use std::{
|
||||||
collections::HashMap,
|
|
||||||
fmt,
|
fmt,
|
||||||
future::Future,
|
future::Future,
|
||||||
marker::PhantomData,
|
marker::PhantomData,
|
||||||
|
@ -88,10 +90,10 @@ pub struct Peer {
|
||||||
|
|
||||||
#[derive(Clone)]
|
#[derive(Clone)]
|
||||||
pub struct ConnectionState {
|
pub struct ConnectionState {
|
||||||
outgoing_tx: futures::channel::mpsc::UnboundedSender<proto::Message>,
|
outgoing_tx: mpsc::UnboundedSender<proto::Message>,
|
||||||
next_message_id: Arc<AtomicU32>,
|
next_message_id: Arc<AtomicU32>,
|
||||||
response_channels:
|
response_channels:
|
||||||
Arc<Mutex<Option<HashMap<u32, oneshot::Sender<(proto::Envelope, barrier::Sender)>>>>>,
|
Arc<Mutex<Option<HashMap<u32, oneshot::Sender<(proto::Envelope, oneshot::Sender<()>)>>>>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
const KEEPALIVE_INTERVAL: Duration = Duration::from_secs(1);
|
const KEEPALIVE_INTERVAL: Duration = Duration::from_secs(1);
|
||||||
|
@ -125,7 +127,7 @@ impl Peer {
|
||||||
// bounded channel so that other peers will receive backpressure if they send
|
// bounded channel so that other peers will receive backpressure if they send
|
||||||
// messages faster than this peer can process them.
|
// messages faster than this peer can process them.
|
||||||
let (mut incoming_tx, incoming_rx) = mpsc::channel(64);
|
let (mut incoming_tx, incoming_rx) = mpsc::channel(64);
|
||||||
let (outgoing_tx, mut outgoing_rx) = futures::channel::mpsc::unbounded();
|
let (outgoing_tx, mut outgoing_rx) = mpsc::unbounded();
|
||||||
|
|
||||||
let connection_id = ConnectionId(self.next_connection_id.fetch_add(1, SeqCst));
|
let connection_id = ConnectionId(self.next_connection_id.fetch_add(1, SeqCst));
|
||||||
let connection_state = ConnectionState {
|
let connection_state = ConnectionState {
|
||||||
|
@ -173,8 +175,10 @@ impl Peer {
|
||||||
let incoming = incoming.context("received invalid RPC message")?;
|
let incoming = incoming.context("received invalid RPC message")?;
|
||||||
receive_timeout.set(create_timer(RECEIVE_TIMEOUT).fuse());
|
receive_timeout.set(create_timer(RECEIVE_TIMEOUT).fuse());
|
||||||
if let proto::Message::Envelope(incoming) = incoming {
|
if let proto::Message::Envelope(incoming) = incoming {
|
||||||
if incoming_tx.send(incoming).await.is_err() {
|
match incoming_tx.send(incoming).timeout(RECEIVE_TIMEOUT).await {
|
||||||
return Ok(());
|
Some(Ok(_)) => {},
|
||||||
|
Some(Err(_)) => return Ok(()),
|
||||||
|
None => Err(anyhow!("timed out processing incoming message"))?,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
break;
|
break;
|
||||||
|
@ -206,14 +210,14 @@ impl Peer {
|
||||||
if let Some(responding_to) = incoming.responding_to {
|
if let Some(responding_to) = incoming.responding_to {
|
||||||
let channel = response_channels.lock().as_mut()?.remove(&responding_to);
|
let channel = response_channels.lock().as_mut()?.remove(&responding_to);
|
||||||
if let Some(tx) = channel {
|
if let Some(tx) = channel {
|
||||||
let mut requester_resumed = barrier::channel();
|
let requester_resumed = oneshot::channel();
|
||||||
if let Err(error) = tx.send((incoming, requester_resumed.0)) {
|
if let Err(error) = tx.send((incoming, requester_resumed.0)) {
|
||||||
log::debug!(
|
log::debug!(
|
||||||
"received RPC but request future was dropped {:?}",
|
"received RPC but request future was dropped {:?}",
|
||||||
error.0
|
error.0
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
requester_resumed.1.recv().await;
|
let _ = requester_resumed.1.await;
|
||||||
} else {
|
} else {
|
||||||
log::warn!("received RPC response to unknown request {}", responding_to);
|
log::warn!("received RPC response to unknown request {}", responding_to);
|
||||||
}
|
}
|
||||||
|
@ -719,26 +723,26 @@ mod tests {
|
||||||
.add_test_connection(client_conn, cx.background())
|
.add_test_connection(client_conn, cx.background())
|
||||||
.await;
|
.await;
|
||||||
|
|
||||||
let (mut io_ended_tx, mut io_ended_rx) = postage::barrier::channel();
|
let (io_ended_tx, io_ended_rx) = oneshot::channel();
|
||||||
executor
|
executor
|
||||||
.spawn(async move {
|
.spawn(async move {
|
||||||
io_handler.await.ok();
|
io_handler.await.ok();
|
||||||
io_ended_tx.send(()).await.unwrap();
|
io_ended_tx.send(()).unwrap();
|
||||||
})
|
})
|
||||||
.detach();
|
.detach();
|
||||||
|
|
||||||
let (mut messages_ended_tx, mut messages_ended_rx) = postage::barrier::channel();
|
let (messages_ended_tx, messages_ended_rx) = oneshot::channel();
|
||||||
executor
|
executor
|
||||||
.spawn(async move {
|
.spawn(async move {
|
||||||
incoming.next().await;
|
incoming.next().await;
|
||||||
messages_ended_tx.send(()).await.unwrap();
|
messages_ended_tx.send(()).unwrap();
|
||||||
})
|
})
|
||||||
.detach();
|
.detach();
|
||||||
|
|
||||||
client.disconnect(connection_id);
|
client.disconnect(connection_id);
|
||||||
|
|
||||||
io_ended_rx.recv().await;
|
let _ = io_ended_rx.await;
|
||||||
messages_ended_rx.recv().await;
|
let _ = messages_ended_rx.await;
|
||||||
assert!(server_conn
|
assert!(server_conn
|
||||||
.send(WebSocketMessage::Binary(vec![]))
|
.send(WebSocketMessage::Binary(vec![]))
|
||||||
.await
|
.await
|
||||||
|
|
|
@ -14,7 +14,6 @@ required-features = ["seed-support"]
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
collections = { path = "../collections" }
|
collections = { path = "../collections" }
|
||||||
settings = { path = "../settings" }
|
|
||||||
rpc = { path = "../rpc" }
|
rpc = { path = "../rpc" }
|
||||||
anyhow = "1.0.40"
|
anyhow = "1.0.40"
|
||||||
async-io = "1.3"
|
async-io = "1.3"
|
||||||
|
@ -34,7 +33,6 @@ lipsum = { version = "0.8", optional = true }
|
||||||
oauth2 = { version = "4.0.0", default_features = false }
|
oauth2 = { version = "4.0.0", default_features = false }
|
||||||
oauth2-surf = "0.1.1"
|
oauth2-surf = "0.1.1"
|
||||||
parking_lot = "0.11.1"
|
parking_lot = "0.11.1"
|
||||||
postage = { version = "0.4.1", features = ["futures-traits"] }
|
|
||||||
rand = "0.8"
|
rand = "0.8"
|
||||||
rust-embed = { version = "6.3", features = ["include-exclude"] }
|
rust-embed = { version = "6.3", features = ["include-exclude"] }
|
||||||
scrypt = "0.7"
|
scrypt = "0.7"
|
||||||
|
@ -65,6 +63,7 @@ editor = { path = "../editor", features = ["test-support"] }
|
||||||
language = { path = "../language", features = ["test-support"] }
|
language = { path = "../language", features = ["test-support"] }
|
||||||
lsp = { path = "../lsp", features = ["test-support"] }
|
lsp = { path = "../lsp", features = ["test-support"] }
|
||||||
project = { path = "../project", features = ["test-support"] }
|
project = { path = "../project", features = ["test-support"] }
|
||||||
|
settings = { path = "../settings", features = ["test-support"] }
|
||||||
workspace = { path = "../workspace", features = ["test-support"] }
|
workspace = { path = "../workspace", features = ["test-support"] }
|
||||||
ctor = "0.1"
|
ctor = "0.1"
|
||||||
env_logger = "0.8"
|
env_logger = "0.8"
|
||||||
|
|
|
@ -1080,7 +1080,7 @@ mod tests {
|
||||||
use ::rpc::Peer;
|
use ::rpc::Peer;
|
||||||
use client::{
|
use client::{
|
||||||
self, test::FakeHttpClient, Channel, ChannelDetails, ChannelList, Client, Credentials,
|
self, test::FakeHttpClient, Channel, ChannelDetails, ChannelList, Client, Credentials,
|
||||||
EstablishConnectionError, UserStore,
|
EstablishConnectionError, UserStore, RECEIVE_TIMEOUT,
|
||||||
};
|
};
|
||||||
use collections::BTreeMap;
|
use collections::BTreeMap;
|
||||||
use editor::{
|
use editor::{
|
||||||
|
@ -1094,7 +1094,6 @@ mod tests {
|
||||||
};
|
};
|
||||||
use lsp::{self, FakeLanguageServer};
|
use lsp::{self, FakeLanguageServer};
|
||||||
use parking_lot::Mutex;
|
use parking_lot::Mutex;
|
||||||
use postage::barrier;
|
|
||||||
use project::{
|
use project::{
|
||||||
fs::{FakeFs, Fs as _},
|
fs::{FakeFs, Fs as _},
|
||||||
search::SearchQuery,
|
search::SearchQuery,
|
||||||
|
@ -1118,6 +1117,7 @@ mod tests {
|
||||||
},
|
},
|
||||||
time::Duration,
|
time::Duration,
|
||||||
};
|
};
|
||||||
|
use util::TryFutureExt;
|
||||||
use workspace::{Item, SplitDirection, ToggleFollow, Workspace, WorkspaceParams};
|
use workspace::{Item, SplitDirection, ToggleFollow, Workspace, WorkspaceParams};
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
|
@ -4995,6 +4995,7 @@ mod tests {
|
||||||
let operations = Rc::new(Cell::new(0));
|
let operations = Rc::new(Cell::new(0));
|
||||||
let mut server = TestServer::start(cx.foreground(), cx.background()).await;
|
let mut server = TestServer::start(cx.foreground(), cx.background()).await;
|
||||||
let mut clients = Vec::new();
|
let mut clients = Vec::new();
|
||||||
|
let mut user_ids = Vec::new();
|
||||||
let files = Arc::new(Mutex::new(Vec::new()));
|
let files = Arc::new(Mutex::new(Vec::new()));
|
||||||
|
|
||||||
let mut next_entity_id = 100000;
|
let mut next_entity_id = 100000;
|
||||||
|
@ -5162,6 +5163,8 @@ mod tests {
|
||||||
});
|
});
|
||||||
host_language_registry.add(Arc::new(language));
|
host_language_registry.add(Arc::new(language));
|
||||||
|
|
||||||
|
let host_disconnected = Rc::new(AtomicBool::new(false));
|
||||||
|
user_ids.push(host.current_user_id(&host_cx));
|
||||||
clients.push(cx.foreground().spawn(host.simulate_host(
|
clients.push(cx.foreground().spawn(host.simulate_host(
|
||||||
host_project,
|
host_project,
|
||||||
files,
|
files,
|
||||||
|
@ -5203,16 +5206,49 @@ mod tests {
|
||||||
)
|
)
|
||||||
.await
|
.await
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
user_ids.push(guest.current_user_id(&guest_cx));
|
||||||
clients.push(cx.foreground().spawn(guest.simulate_guest(
|
clients.push(cx.foreground().spawn(guest.simulate_guest(
|
||||||
guest_id,
|
guest_id,
|
||||||
guest_project,
|
guest_project,
|
||||||
operations.clone(),
|
operations.clone(),
|
||||||
max_operations,
|
max_operations,
|
||||||
rng.clone(),
|
rng.clone(),
|
||||||
|
host_disconnected.clone(),
|
||||||
guest_cx,
|
guest_cx,
|
||||||
)));
|
)));
|
||||||
|
|
||||||
log::info!("Guest {} added", guest_id);
|
log::info!("Guest {} added", guest_id);
|
||||||
|
} else if rng.lock().gen_bool(0.05) {
|
||||||
|
host_disconnected.store(true, SeqCst);
|
||||||
|
server.disconnect_client(user_ids[0]);
|
||||||
|
cx.foreground().advance_clock(RECEIVE_TIMEOUT);
|
||||||
|
let mut clients = futures::future::join_all(clients).await;
|
||||||
|
cx.foreground().run_until_parked();
|
||||||
|
|
||||||
|
let (host, mut host_cx) = clients.remove(0);
|
||||||
|
host.project
|
||||||
|
.as_ref()
|
||||||
|
.unwrap()
|
||||||
|
.read_with(&host_cx, |project, _| assert!(!project.is_shared()));
|
||||||
|
for (guest, mut guest_cx) in clients {
|
||||||
|
let contacts = server
|
||||||
|
.store
|
||||||
|
.read()
|
||||||
|
.contacts_for_user(guest.current_user_id(&guest_cx));
|
||||||
|
assert!(!contacts
|
||||||
|
.iter()
|
||||||
|
.flat_map(|contact| &contact.projects)
|
||||||
|
.any(|project| project.id == host_project_id));
|
||||||
|
guest
|
||||||
|
.project
|
||||||
|
.as_ref()
|
||||||
|
.unwrap()
|
||||||
|
.read_with(&guest_cx, |project, _| assert!(project.is_read_only()));
|
||||||
|
guest_cx.update(|_| drop(guest));
|
||||||
|
}
|
||||||
|
host_cx.update(|_| drop(host));
|
||||||
|
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -5325,7 +5361,7 @@ mod tests {
|
||||||
server: Arc<Server>,
|
server: Arc<Server>,
|
||||||
foreground: Rc<executor::Foreground>,
|
foreground: Rc<executor::Foreground>,
|
||||||
notifications: mpsc::UnboundedReceiver<()>,
|
notifications: mpsc::UnboundedReceiver<()>,
|
||||||
connection_killers: Arc<Mutex<HashMap<UserId, barrier::Sender>>>,
|
connection_killers: Arc<Mutex<HashMap<UserId, Arc<AtomicBool>>>>,
|
||||||
forbid_connections: Arc<AtomicBool>,
|
forbid_connections: Arc<AtomicBool>,
|
||||||
_test_db: TestDb,
|
_test_db: TestDb,
|
||||||
}
|
}
|
||||||
|
@ -5393,9 +5429,9 @@ mod tests {
|
||||||
"server is forbidding connections"
|
"server is forbidding connections"
|
||||||
)))
|
)))
|
||||||
} else {
|
} else {
|
||||||
let (client_conn, server_conn, kill_conn) =
|
let (client_conn, server_conn, killed) =
|
||||||
Connection::in_memory(cx.background());
|
Connection::in_memory(cx.background());
|
||||||
connection_killers.lock().insert(user_id, kill_conn);
|
connection_killers.lock().insert(user_id, killed);
|
||||||
cx.background()
|
cx.background()
|
||||||
.spawn(server.handle_connection(
|
.spawn(server.handle_connection(
|
||||||
server_conn,
|
server_conn,
|
||||||
|
@ -5437,7 +5473,11 @@ mod tests {
|
||||||
}
|
}
|
||||||
|
|
||||||
fn disconnect_client(&self, user_id: UserId) {
|
fn disconnect_client(&self, user_id: UserId) {
|
||||||
self.connection_killers.lock().remove(&user_id);
|
self.connection_killers
|
||||||
|
.lock()
|
||||||
|
.remove(&user_id)
|
||||||
|
.unwrap()
|
||||||
|
.store(true, SeqCst);
|
||||||
}
|
}
|
||||||
|
|
||||||
fn forbid_connections(&self) {
|
fn forbid_connections(&self) {
|
||||||
|
@ -5483,6 +5523,14 @@ mod tests {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl Deref for TestServer {
|
||||||
|
type Target = Server;
|
||||||
|
|
||||||
|
fn deref(&self) -> &Self::Target {
|
||||||
|
&self.server
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl Drop for TestServer {
|
impl Drop for TestServer {
|
||||||
fn drop(&mut self) {
|
fn drop(&mut self) {
|
||||||
self.peer.reset();
|
self.peer.reset();
|
||||||
|
@ -5604,117 +5652,138 @@ mod tests {
|
||||||
rng: Arc<Mutex<StdRng>>,
|
rng: Arc<Mutex<StdRng>>,
|
||||||
mut cx: TestAppContext,
|
mut cx: TestAppContext,
|
||||||
) -> (Self, TestAppContext) {
|
) -> (Self, TestAppContext) {
|
||||||
let fs = project.read_with(&cx, |project, _| project.fs().clone());
|
async fn simulate_host_internal(
|
||||||
while operations.get() < max_operations {
|
client: &mut TestClient,
|
||||||
operations.set(operations.get() + 1);
|
project: ModelHandle<Project>,
|
||||||
|
files: Arc<Mutex<Vec<PathBuf>>>,
|
||||||
|
operations: Rc<Cell<usize>>,
|
||||||
|
max_operations: usize,
|
||||||
|
rng: Arc<Mutex<StdRng>>,
|
||||||
|
cx: &mut TestAppContext,
|
||||||
|
) -> anyhow::Result<()> {
|
||||||
|
let fs = project.read_with(cx, |project, _| project.fs().clone());
|
||||||
|
while operations.get() < max_operations {
|
||||||
|
operations.set(operations.get() + 1);
|
||||||
|
|
||||||
let distribution = rng.lock().gen_range::<usize, _>(0..100);
|
let distribution = rng.lock().gen_range::<usize, _>(0..100);
|
||||||
match distribution {
|
match distribution {
|
||||||
0..=20 if !files.lock().is_empty() => {
|
0..=20 if !files.lock().is_empty() => {
|
||||||
let path = files.lock().choose(&mut *rng.lock()).unwrap().clone();
|
let path = files.lock().choose(&mut *rng.lock()).unwrap().clone();
|
||||||
let mut path = path.as_path();
|
let mut path = path.as_path();
|
||||||
while let Some(parent_path) = path.parent() {
|
while let Some(parent_path) = path.parent() {
|
||||||
path = parent_path;
|
path = parent_path;
|
||||||
|
if rng.lock().gen() {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
log::info!("Host: find/create local worktree {:?}", path);
|
||||||
|
let find_or_create_worktree = project.update(cx, |project, cx| {
|
||||||
|
project.find_or_create_local_worktree(path, true, cx)
|
||||||
|
});
|
||||||
if rng.lock().gen() {
|
if rng.lock().gen() {
|
||||||
break;
|
cx.background().spawn(find_or_create_worktree).detach();
|
||||||
|
} else {
|
||||||
|
find_or_create_worktree.await?;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
10..=80 if !files.lock().is_empty() => {
|
||||||
log::info!("Host: find/create local worktree {:?}", path);
|
let buffer = if client.buffers.is_empty() || rng.lock().gen() {
|
||||||
let find_or_create_worktree = project.update(&mut cx, |project, cx| {
|
let file = files.lock().choose(&mut *rng.lock()).unwrap().clone();
|
||||||
project.find_or_create_local_worktree(path, true, cx)
|
let (worktree, path) = project
|
||||||
});
|
.update(cx, |project, cx| {
|
||||||
let find_or_create_worktree = async move {
|
project.find_or_create_local_worktree(
|
||||||
find_or_create_worktree.await.unwrap();
|
file.clone(),
|
||||||
};
|
true,
|
||||||
if rng.lock().gen() {
|
cx,
|
||||||
cx.background().spawn(find_or_create_worktree).detach();
|
)
|
||||||
} else {
|
})
|
||||||
find_or_create_worktree.await;
|
.await?;
|
||||||
}
|
let project_path =
|
||||||
}
|
worktree.read_with(cx, |worktree, _| (worktree.id(), path));
|
||||||
10..=80 if !files.lock().is_empty() => {
|
|
||||||
let buffer = if self.buffers.is_empty() || rng.lock().gen() {
|
|
||||||
let file = files.lock().choose(&mut *rng.lock()).unwrap().clone();
|
|
||||||
let (worktree, path) = project
|
|
||||||
.update(&mut cx, |project, cx| {
|
|
||||||
project.find_or_create_local_worktree(file.clone(), true, cx)
|
|
||||||
})
|
|
||||||
.await
|
|
||||||
.unwrap();
|
|
||||||
let project_path =
|
|
||||||
worktree.read_with(&cx, |worktree, _| (worktree.id(), path));
|
|
||||||
log::info!(
|
|
||||||
"Host: opening path {:?}, worktree {}, relative_path {:?}",
|
|
||||||
file,
|
|
||||||
project_path.0,
|
|
||||||
project_path.1
|
|
||||||
);
|
|
||||||
let buffer = project
|
|
||||||
.update(&mut cx, |project, cx| {
|
|
||||||
project.open_buffer(project_path, cx)
|
|
||||||
})
|
|
||||||
.await
|
|
||||||
.unwrap();
|
|
||||||
self.buffers.insert(buffer.clone());
|
|
||||||
buffer
|
|
||||||
} else {
|
|
||||||
self.buffers
|
|
||||||
.iter()
|
|
||||||
.choose(&mut *rng.lock())
|
|
||||||
.unwrap()
|
|
||||||
.clone()
|
|
||||||
};
|
|
||||||
|
|
||||||
if rng.lock().gen_bool(0.1) {
|
|
||||||
cx.update(|cx| {
|
|
||||||
log::info!(
|
log::info!(
|
||||||
"Host: dropping buffer {:?}",
|
"Host: opening path {:?}, worktree {}, relative_path {:?}",
|
||||||
buffer.read(cx).file().unwrap().full_path(cx)
|
file,
|
||||||
|
project_path.0,
|
||||||
|
project_path.1
|
||||||
);
|
);
|
||||||
self.buffers.remove(&buffer);
|
let buffer = project
|
||||||
drop(buffer);
|
.update(cx, |project, cx| project.open_buffer(project_path, cx))
|
||||||
});
|
.await
|
||||||
} else {
|
.unwrap();
|
||||||
buffer.update(&mut cx, |buffer, cx| {
|
client.buffers.insert(buffer.clone());
|
||||||
log::info!(
|
buffer
|
||||||
"Host: updating buffer {:?} ({})",
|
} else {
|
||||||
buffer.file().unwrap().full_path(cx),
|
client
|
||||||
buffer.remote_id()
|
.buffers
|
||||||
);
|
.iter()
|
||||||
buffer.randomly_edit(&mut *rng.lock(), 5, cx)
|
.choose(&mut *rng.lock())
|
||||||
});
|
.unwrap()
|
||||||
|
.clone()
|
||||||
|
};
|
||||||
|
|
||||||
|
if rng.lock().gen_bool(0.1) {
|
||||||
|
cx.update(|cx| {
|
||||||
|
log::info!(
|
||||||
|
"Host: dropping buffer {:?}",
|
||||||
|
buffer.read(cx).file().unwrap().full_path(cx)
|
||||||
|
);
|
||||||
|
client.buffers.remove(&buffer);
|
||||||
|
drop(buffer);
|
||||||
|
});
|
||||||
|
} else {
|
||||||
|
buffer.update(cx, |buffer, cx| {
|
||||||
|
log::info!(
|
||||||
|
"Host: updating buffer {:?} ({})",
|
||||||
|
buffer.file().unwrap().full_path(cx),
|
||||||
|
buffer.remote_id()
|
||||||
|
);
|
||||||
|
buffer.randomly_edit(&mut *rng.lock(), 5, cx)
|
||||||
|
});
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
_ => loop {
|
||||||
|
let path_component_count = rng.lock().gen_range::<usize, _>(1..=5);
|
||||||
|
let mut path = PathBuf::new();
|
||||||
|
path.push("/");
|
||||||
|
for _ in 0..path_component_count {
|
||||||
|
let letter = rng.lock().gen_range(b'a'..=b'z');
|
||||||
|
path.push(std::str::from_utf8(&[letter]).unwrap());
|
||||||
|
}
|
||||||
|
path.set_extension("rs");
|
||||||
|
let parent_path = path.parent().unwrap();
|
||||||
|
|
||||||
|
log::info!("Host: creating file {:?}", path,);
|
||||||
|
|
||||||
|
if fs.create_dir(&parent_path).await.is_ok()
|
||||||
|
&& fs.create_file(&path, Default::default()).await.is_ok()
|
||||||
|
{
|
||||||
|
files.lock().push(path);
|
||||||
|
break;
|
||||||
|
} else {
|
||||||
|
log::info!("Host: cannot create file");
|
||||||
|
}
|
||||||
|
},
|
||||||
}
|
}
|
||||||
_ => loop {
|
|
||||||
let path_component_count = rng.lock().gen_range::<usize, _>(1..=5);
|
|
||||||
let mut path = PathBuf::new();
|
|
||||||
path.push("/");
|
|
||||||
for _ in 0..path_component_count {
|
|
||||||
let letter = rng.lock().gen_range(b'a'..=b'z');
|
|
||||||
path.push(std::str::from_utf8(&[letter]).unwrap());
|
|
||||||
}
|
|
||||||
path.set_extension("rs");
|
|
||||||
let parent_path = path.parent().unwrap();
|
|
||||||
|
|
||||||
log::info!("Host: creating file {:?}", path,);
|
cx.background().simulate_random_delay().await;
|
||||||
|
|
||||||
if fs.create_dir(&parent_path).await.is_ok()
|
|
||||||
&& fs.create_file(&path, Default::default()).await.is_ok()
|
|
||||||
{
|
|
||||||
files.lock().push(path);
|
|
||||||
break;
|
|
||||||
} else {
|
|
||||||
log::info!("Host: cannot create file");
|
|
||||||
}
|
|
||||||
},
|
|
||||||
}
|
}
|
||||||
|
|
||||||
cx.background().simulate_random_delay().await;
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
simulate_host_internal(
|
||||||
|
&mut self,
|
||||||
|
project.clone(),
|
||||||
|
files,
|
||||||
|
operations,
|
||||||
|
max_operations,
|
||||||
|
rng,
|
||||||
|
&mut cx,
|
||||||
|
)
|
||||||
|
.log_err()
|
||||||
|
.await;
|
||||||
log::info!("Host done");
|
log::info!("Host done");
|
||||||
|
|
||||||
self.project = Some(project);
|
self.project = Some(project);
|
||||||
(self, cx)
|
(self, cx)
|
||||||
}
|
}
|
||||||
|
@ -5726,244 +5795,292 @@ mod tests {
|
||||||
operations: Rc<Cell<usize>>,
|
operations: Rc<Cell<usize>>,
|
||||||
max_operations: usize,
|
max_operations: usize,
|
||||||
rng: Arc<Mutex<StdRng>>,
|
rng: Arc<Mutex<StdRng>>,
|
||||||
|
host_disconnected: Rc<AtomicBool>,
|
||||||
mut cx: TestAppContext,
|
mut cx: TestAppContext,
|
||||||
) -> (Self, TestAppContext) {
|
) -> (Self, TestAppContext) {
|
||||||
while operations.get() < max_operations {
|
async fn simulate_guest_internal(
|
||||||
let buffer = if self.buffers.is_empty() || rng.lock().gen() {
|
client: &mut TestClient,
|
||||||
let worktree = if let Some(worktree) = project.read_with(&cx, |project, cx| {
|
guest_id: usize,
|
||||||
project
|
project: ModelHandle<Project>,
|
||||||
.worktrees(&cx)
|
operations: Rc<Cell<usize>>,
|
||||||
.filter(|worktree| {
|
max_operations: usize,
|
||||||
let worktree = worktree.read(cx);
|
rng: Arc<Mutex<StdRng>>,
|
||||||
worktree.is_visible()
|
cx: &mut TestAppContext,
|
||||||
&& worktree.entries(false).any(|e| e.is_file())
|
) -> anyhow::Result<()> {
|
||||||
|
while operations.get() < max_operations {
|
||||||
|
let buffer = if client.buffers.is_empty() || rng.lock().gen() {
|
||||||
|
let worktree = if let Some(worktree) =
|
||||||
|
project.read_with(cx, |project, cx| {
|
||||||
|
project
|
||||||
|
.worktrees(&cx)
|
||||||
|
.filter(|worktree| {
|
||||||
|
let worktree = worktree.read(cx);
|
||||||
|
worktree.is_visible()
|
||||||
|
&& worktree.entries(false).any(|e| e.is_file())
|
||||||
|
})
|
||||||
|
.choose(&mut *rng.lock())
|
||||||
|
}) {
|
||||||
|
worktree
|
||||||
|
} else {
|
||||||
|
cx.background().simulate_random_delay().await;
|
||||||
|
continue;
|
||||||
|
};
|
||||||
|
|
||||||
|
operations.set(operations.get() + 1);
|
||||||
|
let (worktree_root_name, project_path) =
|
||||||
|
worktree.read_with(cx, |worktree, _| {
|
||||||
|
let entry = worktree
|
||||||
|
.entries(false)
|
||||||
|
.filter(|e| e.is_file())
|
||||||
|
.choose(&mut *rng.lock())
|
||||||
|
.unwrap();
|
||||||
|
(
|
||||||
|
worktree.root_name().to_string(),
|
||||||
|
(worktree.id(), entry.path.clone()),
|
||||||
|
)
|
||||||
|
});
|
||||||
|
log::info!(
|
||||||
|
"Guest {}: opening path {:?} in worktree {} ({})",
|
||||||
|
guest_id,
|
||||||
|
project_path.1,
|
||||||
|
project_path.0,
|
||||||
|
worktree_root_name,
|
||||||
|
);
|
||||||
|
let buffer = project
|
||||||
|
.update(cx, |project, cx| {
|
||||||
|
project.open_buffer(project_path.clone(), cx)
|
||||||
})
|
})
|
||||||
.choose(&mut *rng.lock())
|
.await?;
|
||||||
}) {
|
log::info!(
|
||||||
worktree
|
"Guest {}: opened path {:?} in worktree {} ({}) with buffer id {}",
|
||||||
|
guest_id,
|
||||||
|
project_path.1,
|
||||||
|
project_path.0,
|
||||||
|
worktree_root_name,
|
||||||
|
buffer.read_with(cx, |buffer, _| buffer.remote_id())
|
||||||
|
);
|
||||||
|
client.buffers.insert(buffer.clone());
|
||||||
|
buffer
|
||||||
} else {
|
} else {
|
||||||
cx.background().simulate_random_delay().await;
|
operations.set(operations.get() + 1);
|
||||||
continue;
|
|
||||||
|
client
|
||||||
|
.buffers
|
||||||
|
.iter()
|
||||||
|
.choose(&mut *rng.lock())
|
||||||
|
.unwrap()
|
||||||
|
.clone()
|
||||||
};
|
};
|
||||||
|
|
||||||
operations.set(operations.get() + 1);
|
let choice = rng.lock().gen_range(0..100);
|
||||||
let (worktree_root_name, project_path) =
|
match choice {
|
||||||
worktree.read_with(&cx, |worktree, _| {
|
0..=9 => {
|
||||||
let entry = worktree
|
cx.update(|cx| {
|
||||||
.entries(false)
|
log::info!(
|
||||||
.filter(|e| e.is_file())
|
"Guest {}: dropping buffer {:?}",
|
||||||
.choose(&mut *rng.lock())
|
guest_id,
|
||||||
.unwrap();
|
buffer.read(cx).file().unwrap().full_path(cx)
|
||||||
(
|
);
|
||||||
worktree.root_name().to_string(),
|
client.buffers.remove(&buffer);
|
||||||
(worktree.id(), entry.path.clone()),
|
drop(buffer);
|
||||||
)
|
});
|
||||||
});
|
}
|
||||||
log::info!(
|
10..=19 => {
|
||||||
"Guest {}: opening path {:?} in worktree {} ({})",
|
let completions = project.update(cx, |project, cx| {
|
||||||
guest_id,
|
log::info!(
|
||||||
project_path.1,
|
"Guest {}: requesting completions for buffer {} ({:?})",
|
||||||
project_path.0,
|
guest_id,
|
||||||
worktree_root_name,
|
buffer.read(cx).remote_id(),
|
||||||
);
|
buffer.read(cx).file().unwrap().full_path(cx)
|
||||||
let buffer = project
|
);
|
||||||
.update(&mut cx, |project, cx| {
|
let offset = rng.lock().gen_range(0..=buffer.read(cx).len());
|
||||||
project.open_buffer(project_path.clone(), cx)
|
project.completions(&buffer, offset, cx)
|
||||||
})
|
});
|
||||||
.await
|
let completions = cx.background().spawn(async move {
|
||||||
.unwrap();
|
completions
|
||||||
log::info!(
|
.await
|
||||||
"Guest {}: opened path {:?} in worktree {} ({}) with buffer id {}",
|
.map_err(|err| anyhow!("completions request failed: {:?}", err))
|
||||||
guest_id,
|
});
|
||||||
project_path.1,
|
if rng.lock().gen_bool(0.3) {
|
||||||
project_path.0,
|
log::info!("Guest {}: detaching completions request", guest_id);
|
||||||
worktree_root_name,
|
cx.update(|cx| completions.detach_and_log_err(cx));
|
||||||
buffer.read_with(&cx, |buffer, _| buffer.remote_id())
|
} else {
|
||||||
);
|
completions.await?;
|
||||||
self.buffers.insert(buffer.clone());
|
}
|
||||||
buffer
|
}
|
||||||
} else {
|
20..=29 => {
|
||||||
operations.set(operations.get() + 1);
|
let code_actions = project.update(cx, |project, cx| {
|
||||||
|
log::info!(
|
||||||
self.buffers
|
"Guest {}: requesting code actions for buffer {} ({:?})",
|
||||||
.iter()
|
guest_id,
|
||||||
.choose(&mut *rng.lock())
|
buffer.read(cx).remote_id(),
|
||||||
.unwrap()
|
buffer.read(cx).file().unwrap().full_path(cx)
|
||||||
.clone()
|
);
|
||||||
};
|
let range = buffer.read(cx).random_byte_range(0, &mut *rng.lock());
|
||||||
|
project.code_actions(&buffer, range, cx)
|
||||||
let choice = rng.lock().gen_range(0..100);
|
});
|
||||||
match choice {
|
let code_actions = cx.background().spawn(async move {
|
||||||
0..=9 => {
|
code_actions.await.map_err(|err| {
|
||||||
cx.update(|cx| {
|
anyhow!("code actions request failed: {:?}", err)
|
||||||
log::info!(
|
})
|
||||||
"Guest {}: dropping buffer {:?}",
|
});
|
||||||
guest_id,
|
if rng.lock().gen_bool(0.3) {
|
||||||
buffer.read(cx).file().unwrap().full_path(cx)
|
log::info!("Guest {}: detaching code actions request", guest_id);
|
||||||
);
|
cx.update(|cx| code_actions.detach_and_log_err(cx));
|
||||||
self.buffers.remove(&buffer);
|
} else {
|
||||||
drop(buffer);
|
code_actions.await?;
|
||||||
});
|
}
|
||||||
}
|
}
|
||||||
10..=19 => {
|
30..=39 if buffer.read_with(cx, |buffer, _| buffer.is_dirty()) => {
|
||||||
let completions = project.update(&mut cx, |project, cx| {
|
let (requested_version, save) = buffer.update(cx, |buffer, cx| {
|
||||||
log::info!(
|
log::info!(
|
||||||
"Guest {}: requesting completions for buffer {} ({:?})",
|
"Guest {}: saving buffer {} ({:?})",
|
||||||
guest_id,
|
guest_id,
|
||||||
buffer.read(cx).remote_id(),
|
buffer.remote_id(),
|
||||||
buffer.read(cx).file().unwrap().full_path(cx)
|
buffer.file().unwrap().full_path(cx)
|
||||||
);
|
);
|
||||||
let offset = rng.lock().gen_range(0..=buffer.read(cx).len());
|
(buffer.version(), buffer.save(cx))
|
||||||
project.completions(&buffer, offset, cx)
|
});
|
||||||
});
|
let save = cx.background().spawn(async move {
|
||||||
let completions = cx.background().spawn(async move {
|
let (saved_version, _) = save
|
||||||
completions.await.expect("completions request failed");
|
.await
|
||||||
});
|
.map_err(|err| anyhow!("save request failed: {:?}", err))?;
|
||||||
if rng.lock().gen_bool(0.3) {
|
assert!(saved_version.observed_all(&requested_version));
|
||||||
log::info!("Guest {}: detaching completions request", guest_id);
|
Ok::<_, anyhow::Error>(())
|
||||||
completions.detach();
|
});
|
||||||
} else {
|
if rng.lock().gen_bool(0.3) {
|
||||||
completions.await;
|
log::info!("Guest {}: detaching save request", guest_id);
|
||||||
|
cx.update(|cx| save.detach_and_log_err(cx));
|
||||||
|
} else {
|
||||||
|
save.await?;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
40..=44 => {
|
||||||
|
let prepare_rename = project.update(cx, |project, cx| {
|
||||||
|
log::info!(
|
||||||
|
"Guest {}: preparing rename for buffer {} ({:?})",
|
||||||
|
guest_id,
|
||||||
|
buffer.read(cx).remote_id(),
|
||||||
|
buffer.read(cx).file().unwrap().full_path(cx)
|
||||||
|
);
|
||||||
|
let offset = rng.lock().gen_range(0..=buffer.read(cx).len());
|
||||||
|
project.prepare_rename(buffer, offset, cx)
|
||||||
|
});
|
||||||
|
let prepare_rename = cx.background().spawn(async move {
|
||||||
|
prepare_rename.await.map_err(|err| {
|
||||||
|
anyhow!("prepare rename request failed: {:?}", err)
|
||||||
|
})
|
||||||
|
});
|
||||||
|
if rng.lock().gen_bool(0.3) {
|
||||||
|
log::info!("Guest {}: detaching prepare rename request", guest_id);
|
||||||
|
cx.update(|cx| prepare_rename.detach_and_log_err(cx));
|
||||||
|
} else {
|
||||||
|
prepare_rename.await?;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
45..=49 => {
|
||||||
|
let definitions = project.update(cx, |project, cx| {
|
||||||
|
log::info!(
|
||||||
|
"Guest {}: requesting definitions for buffer {} ({:?})",
|
||||||
|
guest_id,
|
||||||
|
buffer.read(cx).remote_id(),
|
||||||
|
buffer.read(cx).file().unwrap().full_path(cx)
|
||||||
|
);
|
||||||
|
let offset = rng.lock().gen_range(0..=buffer.read(cx).len());
|
||||||
|
project.definition(&buffer, offset, cx)
|
||||||
|
});
|
||||||
|
let definitions = cx.background().spawn(async move {
|
||||||
|
definitions
|
||||||
|
.await
|
||||||
|
.map_err(|err| anyhow!("definitions request failed: {:?}", err))
|
||||||
|
});
|
||||||
|
if rng.lock().gen_bool(0.3) {
|
||||||
|
log::info!("Guest {}: detaching definitions request", guest_id);
|
||||||
|
cx.update(|cx| definitions.detach_and_log_err(cx));
|
||||||
|
} else {
|
||||||
|
client
|
||||||
|
.buffers
|
||||||
|
.extend(definitions.await?.into_iter().map(|loc| loc.buffer));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
50..=54 => {
|
||||||
|
let highlights = project.update(cx, |project, cx| {
|
||||||
|
log::info!(
|
||||||
|
"Guest {}: requesting highlights for buffer {} ({:?})",
|
||||||
|
guest_id,
|
||||||
|
buffer.read(cx).remote_id(),
|
||||||
|
buffer.read(cx).file().unwrap().full_path(cx)
|
||||||
|
);
|
||||||
|
let offset = rng.lock().gen_range(0..=buffer.read(cx).len());
|
||||||
|
project.document_highlights(&buffer, offset, cx)
|
||||||
|
});
|
||||||
|
let highlights = cx.background().spawn(async move {
|
||||||
|
highlights
|
||||||
|
.await
|
||||||
|
.map_err(|err| anyhow!("highlights request failed: {:?}", err))
|
||||||
|
});
|
||||||
|
if rng.lock().gen_bool(0.3) {
|
||||||
|
log::info!("Guest {}: detaching highlights request", guest_id);
|
||||||
|
cx.update(|cx| highlights.detach_and_log_err(cx));
|
||||||
|
} else {
|
||||||
|
highlights.await?;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
55..=59 => {
|
||||||
|
let search = project.update(cx, |project, cx| {
|
||||||
|
let query = rng.lock().gen_range('a'..='z');
|
||||||
|
log::info!("Guest {}: project-wide search {:?}", guest_id, query);
|
||||||
|
project.search(SearchQuery::text(query, false, false), cx)
|
||||||
|
});
|
||||||
|
let search = cx.background().spawn(async move {
|
||||||
|
search
|
||||||
|
.await
|
||||||
|
.map_err(|err| anyhow!("search request failed: {:?}", err))
|
||||||
|
});
|
||||||
|
if rng.lock().gen_bool(0.3) {
|
||||||
|
log::info!("Guest {}: detaching search request", guest_id);
|
||||||
|
cx.update(|cx| search.detach_and_log_err(cx));
|
||||||
|
} else {
|
||||||
|
client.buffers.extend(search.await?.into_keys());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
_ => {
|
||||||
|
buffer.update(cx, |buffer, cx| {
|
||||||
|
log::info!(
|
||||||
|
"Guest {}: updating buffer {} ({:?})",
|
||||||
|
guest_id,
|
||||||
|
buffer.remote_id(),
|
||||||
|
buffer.file().unwrap().full_path(cx)
|
||||||
|
);
|
||||||
|
buffer.randomly_edit(&mut *rng.lock(), 5, cx)
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
20..=29 => {
|
cx.background().simulate_random_delay().await;
|
||||||
let code_actions = project.update(&mut cx, |project, cx| {
|
|
||||||
log::info!(
|
|
||||||
"Guest {}: requesting code actions for buffer {} ({:?})",
|
|
||||||
guest_id,
|
|
||||||
buffer.read(cx).remote_id(),
|
|
||||||
buffer.read(cx).file().unwrap().full_path(cx)
|
|
||||||
);
|
|
||||||
let range = buffer.read(cx).random_byte_range(0, &mut *rng.lock());
|
|
||||||
project.code_actions(&buffer, range, cx)
|
|
||||||
});
|
|
||||||
let code_actions = cx.background().spawn(async move {
|
|
||||||
code_actions.await.expect("code actions request failed");
|
|
||||||
});
|
|
||||||
if rng.lock().gen_bool(0.3) {
|
|
||||||
log::info!("Guest {}: detaching code actions request", guest_id);
|
|
||||||
code_actions.detach();
|
|
||||||
} else {
|
|
||||||
code_actions.await;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
30..=39 if buffer.read_with(&cx, |buffer, _| buffer.is_dirty()) => {
|
|
||||||
let (requested_version, save) = buffer.update(&mut cx, |buffer, cx| {
|
|
||||||
log::info!(
|
|
||||||
"Guest {}: saving buffer {} ({:?})",
|
|
||||||
guest_id,
|
|
||||||
buffer.remote_id(),
|
|
||||||
buffer.file().unwrap().full_path(cx)
|
|
||||||
);
|
|
||||||
(buffer.version(), buffer.save(cx))
|
|
||||||
});
|
|
||||||
let save = cx.background().spawn(async move {
|
|
||||||
let (saved_version, _) = save.await.expect("save request failed");
|
|
||||||
assert!(saved_version.observed_all(&requested_version));
|
|
||||||
});
|
|
||||||
if rng.lock().gen_bool(0.3) {
|
|
||||||
log::info!("Guest {}: detaching save request", guest_id);
|
|
||||||
save.detach();
|
|
||||||
} else {
|
|
||||||
save.await;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
40..=44 => {
|
|
||||||
let prepare_rename = project.update(&mut cx, |project, cx| {
|
|
||||||
log::info!(
|
|
||||||
"Guest {}: preparing rename for buffer {} ({:?})",
|
|
||||||
guest_id,
|
|
||||||
buffer.read(cx).remote_id(),
|
|
||||||
buffer.read(cx).file().unwrap().full_path(cx)
|
|
||||||
);
|
|
||||||
let offset = rng.lock().gen_range(0..=buffer.read(cx).len());
|
|
||||||
project.prepare_rename(buffer, offset, cx)
|
|
||||||
});
|
|
||||||
let prepare_rename = cx.background().spawn(async move {
|
|
||||||
prepare_rename.await.expect("prepare rename request failed");
|
|
||||||
});
|
|
||||||
if rng.lock().gen_bool(0.3) {
|
|
||||||
log::info!("Guest {}: detaching prepare rename request", guest_id);
|
|
||||||
prepare_rename.detach();
|
|
||||||
} else {
|
|
||||||
prepare_rename.await;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
45..=49 => {
|
|
||||||
let definitions = project.update(&mut cx, |project, cx| {
|
|
||||||
log::info!(
|
|
||||||
"Guest {}: requesting definitions for buffer {} ({:?})",
|
|
||||||
guest_id,
|
|
||||||
buffer.read(cx).remote_id(),
|
|
||||||
buffer.read(cx).file().unwrap().full_path(cx)
|
|
||||||
);
|
|
||||||
let offset = rng.lock().gen_range(0..=buffer.read(cx).len());
|
|
||||||
project.definition(&buffer, offset, cx)
|
|
||||||
});
|
|
||||||
let definitions = cx.background().spawn(async move {
|
|
||||||
definitions.await.expect("definitions request failed")
|
|
||||||
});
|
|
||||||
if rng.lock().gen_bool(0.3) {
|
|
||||||
log::info!("Guest {}: detaching definitions request", guest_id);
|
|
||||||
definitions.detach();
|
|
||||||
} else {
|
|
||||||
self.buffers
|
|
||||||
.extend(definitions.await.into_iter().map(|loc| loc.buffer));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
50..=54 => {
|
|
||||||
let highlights = project.update(&mut cx, |project, cx| {
|
|
||||||
log::info!(
|
|
||||||
"Guest {}: requesting highlights for buffer {} ({:?})",
|
|
||||||
guest_id,
|
|
||||||
buffer.read(cx).remote_id(),
|
|
||||||
buffer.read(cx).file().unwrap().full_path(cx)
|
|
||||||
);
|
|
||||||
let offset = rng.lock().gen_range(0..=buffer.read(cx).len());
|
|
||||||
project.document_highlights(&buffer, offset, cx)
|
|
||||||
});
|
|
||||||
let highlights = cx.background().spawn(async move {
|
|
||||||
highlights.await.expect("highlights request failed");
|
|
||||||
});
|
|
||||||
if rng.lock().gen_bool(0.3) {
|
|
||||||
log::info!("Guest {}: detaching highlights request", guest_id);
|
|
||||||
highlights.detach();
|
|
||||||
} else {
|
|
||||||
highlights.await;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
55..=59 => {
|
|
||||||
let search = project.update(&mut cx, |project, cx| {
|
|
||||||
let query = rng.lock().gen_range('a'..='z');
|
|
||||||
log::info!("Guest {}: project-wide search {:?}", guest_id, query);
|
|
||||||
project.search(SearchQuery::text(query, false, false), cx)
|
|
||||||
});
|
|
||||||
let search = cx
|
|
||||||
.background()
|
|
||||||
.spawn(async move { search.await.expect("search request failed") });
|
|
||||||
if rng.lock().gen_bool(0.3) {
|
|
||||||
log::info!("Guest {}: detaching search request", guest_id);
|
|
||||||
search.detach();
|
|
||||||
} else {
|
|
||||||
self.buffers.extend(search.await.into_keys());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
_ => {
|
|
||||||
buffer.update(&mut cx, |buffer, cx| {
|
|
||||||
log::info!(
|
|
||||||
"Guest {}: updating buffer {} ({:?})",
|
|
||||||
guest_id,
|
|
||||||
buffer.remote_id(),
|
|
||||||
buffer.file().unwrap().full_path(cx)
|
|
||||||
);
|
|
||||||
buffer.randomly_edit(&mut *rng.lock(), 5, cx)
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
cx.background().simulate_random_delay().await;
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
log::info!("Guest {} done", guest_id);
|
match simulate_guest_internal(
|
||||||
|
&mut self,
|
||||||
|
guest_id,
|
||||||
|
project.clone(),
|
||||||
|
operations,
|
||||||
|
max_operations,
|
||||||
|
rng,
|
||||||
|
&mut cx,
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
{
|
||||||
|
Ok(()) => log::info!("guest {} done", guest_id),
|
||||||
|
Err(err) => {
|
||||||
|
if host_disconnected.load(SeqCst) {
|
||||||
|
log::error!("guest {} simulation error - {:?}", guest_id, err);
|
||||||
|
} else {
|
||||||
|
panic!("guest {} simulation error - {:?}", guest_id, err);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
self.project = Some(project);
|
self.project = Some(project);
|
||||||
(self, cx)
|
(self, cx)
|
||||||
|
|
|
@ -244,6 +244,9 @@ impl Store {
|
||||||
language_servers: Default::default(),
|
language_servers: Default::default(),
|
||||||
},
|
},
|
||||||
);
|
);
|
||||||
|
if let Some(connection) = self.connections.get_mut(&host_connection_id) {
|
||||||
|
connection.projects.insert(project_id);
|
||||||
|
}
|
||||||
self.next_project_id += 1;
|
self.next_project_id += 1;
|
||||||
project_id
|
project_id
|
||||||
}
|
}
|
||||||
|
@ -266,9 +269,7 @@ impl Store {
|
||||||
.or_default()
|
.or_default()
|
||||||
.insert(project_id);
|
.insert(project_id);
|
||||||
}
|
}
|
||||||
if let Some(connection) = self.connections.get_mut(&project.host_connection_id) {
|
|
||||||
connection.projects.insert(project_id);
|
|
||||||
}
|
|
||||||
project.worktrees.insert(worktree_id, worktree);
|
project.worktrees.insert(worktree_id, worktree);
|
||||||
if let Ok(share) = project.share_mut() {
|
if let Ok(share) = project.share_mut() {
|
||||||
share.worktrees.insert(worktree_id, Default::default());
|
share.worktrees.insert(worktree_id, Default::default());
|
||||||
|
|
Loading…
Reference in a new issue