pub mod repository; use anyhow::{anyhow, Result}; use fsevent::EventStream; use futures::{future::BoxFuture, Stream, StreamExt}; use git2::Repository as LibGitRepository; use parking_lot::Mutex; use repository::GitRepository; use rope::Rope; use smol::io::{AsyncReadExt, AsyncWriteExt}; use std::io::Write; use std::sync::Arc; use std::{ io, os::unix::fs::MetadataExt, path::{Component, Path, PathBuf}, pin::Pin, time::{Duration, SystemTime}, }; use tempfile::NamedTempFile; use text::LineEnding; use util::ResultExt; #[cfg(any(test, feature = "test-support"))] use collections::{btree_map, BTreeMap}; #[cfg(any(test, feature = "test-support"))] use repository::{FakeGitRepositoryState, GitFileStatus}; #[cfg(any(test, feature = "test-support"))] use std::ffi::OsStr; #[cfg(any(test, feature = "test-support"))] use std::sync::Weak; #[async_trait::async_trait] pub trait Fs: Send + Sync { async fn create_dir(&self, path: &Path) -> Result<()>; async fn create_file(&self, path: &Path, options: CreateOptions) -> Result<()>; async fn copy_file(&self, source: &Path, target: &Path, options: CopyOptions) -> Result<()>; async fn rename(&self, source: &Path, target: &Path, options: RenameOptions) -> Result<()>; async fn remove_dir(&self, path: &Path, options: RemoveOptions) -> Result<()>; async fn remove_file(&self, path: &Path, options: RemoveOptions) -> Result<()>; async fn open_sync(&self, path: &Path) -> Result>; async fn load(&self, path: &Path) -> Result; async fn atomic_write(&self, path: PathBuf, text: String) -> Result<()>; async fn save(&self, path: &Path, text: &Rope, line_ending: LineEnding) -> Result<()>; async fn canonicalize(&self, path: &Path) -> Result; async fn is_file(&self, path: &Path) -> bool; async fn metadata(&self, path: &Path) -> Result>; async fn read_link(&self, path: &Path) -> Result; async fn read_dir( &self, path: &Path, ) -> Result>>>>; async fn watch( &self, path: &Path, latency: Duration, ) -> Pin>>>; fn open_repo(&self, abs_dot_git: &Path) -> Option>>; fn is_fake(&self) -> bool; #[cfg(any(test, feature = "test-support"))] fn as_fake(&self) -> &FakeFs; } #[derive(Copy, Clone, Default)] pub struct CreateOptions { pub overwrite: bool, pub ignore_if_exists: bool, } #[derive(Copy, Clone, Default)] pub struct CopyOptions { pub overwrite: bool, pub ignore_if_exists: bool, } #[derive(Copy, Clone, Default)] pub struct RenameOptions { pub overwrite: bool, pub ignore_if_exists: bool, } #[derive(Copy, Clone, Default)] pub struct RemoveOptions { pub recursive: bool, pub ignore_if_not_exists: bool, } #[derive(Clone, Debug)] pub struct Metadata { pub inode: u64, pub mtime: SystemTime, pub is_symlink: bool, pub is_dir: bool, } impl From for CreateOptions { fn from(options: lsp::CreateFileOptions) -> Self { Self { overwrite: options.overwrite.unwrap_or(false), ignore_if_exists: options.ignore_if_exists.unwrap_or(false), } } } impl From for RenameOptions { fn from(options: lsp::RenameFileOptions) -> Self { Self { overwrite: options.overwrite.unwrap_or(false), ignore_if_exists: options.ignore_if_exists.unwrap_or(false), } } } impl From for RemoveOptions { fn from(options: lsp::DeleteFileOptions) -> Self { Self { recursive: options.recursive.unwrap_or(false), ignore_if_not_exists: options.ignore_if_not_exists.unwrap_or(false), } } } pub struct RealFs; #[async_trait::async_trait] impl Fs for RealFs { async fn create_dir(&self, path: &Path) -> Result<()> { Ok(smol::fs::create_dir_all(path).await?) } async fn create_file(&self, path: &Path, options: CreateOptions) -> Result<()> { let mut open_options = smol::fs::OpenOptions::new(); open_options.write(true).create(true); if options.overwrite { open_options.truncate(true); } else if !options.ignore_if_exists { open_options.create_new(true); } open_options.open(path).await?; Ok(()) } async fn copy_file(&self, source: &Path, target: &Path, options: CopyOptions) -> Result<()> { if !options.overwrite && smol::fs::metadata(target).await.is_ok() { if options.ignore_if_exists { return Ok(()); } else { return Err(anyhow!("{target:?} already exists")); } } smol::fs::copy(source, target).await?; Ok(()) } async fn rename(&self, source: &Path, target: &Path, options: RenameOptions) -> Result<()> { if !options.overwrite && smol::fs::metadata(target).await.is_ok() { if options.ignore_if_exists { return Ok(()); } else { return Err(anyhow!("{target:?} already exists")); } } smol::fs::rename(source, target).await?; Ok(()) } async fn remove_dir(&self, path: &Path, options: RemoveOptions) -> Result<()> { let result = if options.recursive { smol::fs::remove_dir_all(path).await } else { smol::fs::remove_dir(path).await }; match result { Ok(()) => Ok(()), Err(err) if err.kind() == io::ErrorKind::NotFound && options.ignore_if_not_exists => { Ok(()) } Err(err) => Err(err)?, } } async fn remove_file(&self, path: &Path, options: RemoveOptions) -> Result<()> { match smol::fs::remove_file(path).await { Ok(()) => Ok(()), Err(err) if err.kind() == io::ErrorKind::NotFound && options.ignore_if_not_exists => { Ok(()) } Err(err) => Err(err)?, } } async fn open_sync(&self, path: &Path) -> Result> { Ok(Box::new(std::fs::File::open(path)?)) } async fn load(&self, path: &Path) -> Result { let mut file = smol::fs::File::open(path).await?; let mut text = String::new(); file.read_to_string(&mut text).await?; Ok(text) } async fn atomic_write(&self, path: PathBuf, data: String) -> Result<()> { smol::unblock(move || { let mut tmp_file = NamedTempFile::new()?; tmp_file.write_all(data.as_bytes())?; tmp_file.persist(path)?; Ok::<(), anyhow::Error>(()) }) .await?; Ok(()) } async fn save(&self, path: &Path, text: &Rope, line_ending: LineEnding) -> Result<()> { let buffer_size = text.summary().len.min(10 * 1024); if let Some(path) = path.parent() { self.create_dir(path).await?; } let file = smol::fs::File::create(path).await?; let mut writer = smol::io::BufWriter::with_capacity(buffer_size, file); for chunk in chunks(text, line_ending) { writer.write_all(chunk.as_bytes()).await?; } writer.flush().await?; Ok(()) } async fn canonicalize(&self, path: &Path) -> Result { Ok(smol::fs::canonicalize(path).await?) } async fn is_file(&self, path: &Path) -> bool { smol::fs::metadata(path) .await .map_or(false, |metadata| metadata.is_file()) } async fn metadata(&self, path: &Path) -> Result> { let symlink_metadata = match smol::fs::symlink_metadata(path).await { Ok(metadata) => metadata, Err(err) => { return match (err.kind(), err.raw_os_error()) { (io::ErrorKind::NotFound, _) => Ok(None), (io::ErrorKind::Other, Some(libc::ENOTDIR)) => Ok(None), _ => Err(anyhow::Error::new(err)), } } }; let is_symlink = symlink_metadata.file_type().is_symlink(); let metadata = if is_symlink { smol::fs::metadata(path).await? } else { symlink_metadata }; Ok(Some(Metadata { inode: metadata.ino(), mtime: metadata.modified().unwrap(), is_symlink, is_dir: metadata.file_type().is_dir(), })) } async fn read_link(&self, path: &Path) -> Result { let path = smol::fs::read_link(path).await?; Ok(path) } async fn read_dir( &self, path: &Path, ) -> Result>>>> { let result = smol::fs::read_dir(path).await?.map(|entry| match entry { Ok(entry) => Ok(entry.path()), Err(error) => Err(anyhow!("failed to read dir entry {:?}", error)), }); Ok(Box::pin(result)) } async fn watch( &self, path: &Path, latency: Duration, ) -> Pin>>> { let (tx, rx) = smol::channel::unbounded(); let (stream, handle) = EventStream::new(&[path], latency); std::thread::spawn(move || { stream.run(move |events| smol::block_on(tx.send(events)).is_ok()); }); Box::pin(rx.chain(futures::stream::once(async move { drop(handle); vec![] }))) } fn open_repo(&self, dotgit_path: &Path) -> Option>> { LibGitRepository::open(&dotgit_path) .log_err() .and_then::>, _>(|libgit_repository| { Some(Arc::new(Mutex::new(libgit_repository))) }) } fn is_fake(&self) -> bool { false } #[cfg(any(test, feature = "test-support"))] fn as_fake(&self) -> &FakeFs { panic!("called `RealFs::as_fake`") } } #[cfg(any(test, feature = "test-support"))] pub struct FakeFs { // Use an unfair lock to ensure tests are deterministic. state: Mutex, executor: Weak, } #[cfg(any(test, feature = "test-support"))] struct FakeFsState { root: Arc>, next_inode: u64, next_mtime: SystemTime, event_txs: Vec>>, events_paused: bool, buffered_events: Vec, metadata_call_count: usize, read_dir_call_count: usize, } #[cfg(any(test, feature = "test-support"))] #[derive(Debug)] enum FakeFsEntry { File { inode: u64, mtime: SystemTime, content: String, }, Dir { inode: u64, mtime: SystemTime, entries: BTreeMap>>, git_repo_state: Option>>, }, Symlink { target: PathBuf, }, } #[cfg(any(test, feature = "test-support"))] impl FakeFsState { fn read_path<'a>(&'a self, target: &Path) -> Result>> { Ok(self .try_read_path(target, true) .ok_or_else(|| anyhow!("path does not exist: {}", target.display()))? .0) } fn try_read_path<'a>( &'a self, target: &Path, follow_symlink: bool, ) -> Option<(Arc>, PathBuf)> { let mut path = target.to_path_buf(); let mut canonical_path = PathBuf::new(); let mut entry_stack = Vec::new(); 'outer: loop { let mut path_components = path.components().peekable(); while let Some(component) = path_components.next() { match component { Component::Prefix(_) => panic!("prefix paths aren't supported"), Component::RootDir => { entry_stack.clear(); entry_stack.push(self.root.clone()); canonical_path.clear(); canonical_path.push("/"); } Component::CurDir => {} Component::ParentDir => { entry_stack.pop()?; canonical_path.pop(); } Component::Normal(name) => { let current_entry = entry_stack.last().cloned()?; let current_entry = current_entry.lock(); if let FakeFsEntry::Dir { entries, .. } = &*current_entry { let entry = entries.get(name.to_str().unwrap()).cloned()?; if path_components.peek().is_some() || follow_symlink { let entry = entry.lock(); if let FakeFsEntry::Symlink { target, .. } = &*entry { let mut target = target.clone(); target.extend(path_components); path = target; continue 'outer; } } entry_stack.push(entry.clone()); canonical_path.push(name); } else { return None; } } } } break; } Some((entry_stack.pop()?, canonical_path)) } fn write_path(&self, path: &Path, callback: Fn) -> Result where Fn: FnOnce(btree_map::Entry>>) -> Result, { let path = normalize_path(path); let filename = path .file_name() .ok_or_else(|| anyhow!("cannot overwrite the root"))?; let parent_path = path.parent().unwrap(); let parent = self.read_path(parent_path)?; let mut parent = parent.lock(); let new_entry = parent .dir_entries(parent_path)? .entry(filename.to_str().unwrap().into()); callback(new_entry) } fn emit_event(&mut self, paths: I) where I: IntoIterator, T: Into, { self.buffered_events .extend(paths.into_iter().map(|path| fsevent::Event { event_id: 0, flags: fsevent::StreamFlags::empty(), path: path.into(), })); if !self.events_paused { self.flush_events(self.buffered_events.len()); } } fn flush_events(&mut self, mut count: usize) { count = count.min(self.buffered_events.len()); let events = self.buffered_events.drain(0..count).collect::>(); self.event_txs.retain(|tx| { let _ = tx.try_send(events.clone()); !tx.is_closed() }); } } #[cfg(any(test, feature = "test-support"))] lazy_static::lazy_static! { pub static ref FS_DOT_GIT: &'static OsStr = OsStr::new(".git"); } #[cfg(any(test, feature = "test-support"))] impl FakeFs { pub fn new(executor: Arc) -> Arc { Arc::new(Self { executor: Arc::downgrade(&executor), state: Mutex::new(FakeFsState { root: Arc::new(Mutex::new(FakeFsEntry::Dir { inode: 0, mtime: SystemTime::UNIX_EPOCH, entries: Default::default(), git_repo_state: None, })), next_mtime: SystemTime::UNIX_EPOCH, next_inode: 1, event_txs: Default::default(), buffered_events: Vec::new(), events_paused: false, read_dir_call_count: 0, metadata_call_count: 0, }), }) } pub async fn insert_file(&self, path: impl AsRef, content: String) { self.write_file_internal(path, content).unwrap() } pub async fn insert_symlink(&self, path: impl AsRef, target: PathBuf) { let mut state = self.state.lock(); let path = path.as_ref(); let file = Arc::new(Mutex::new(FakeFsEntry::Symlink { target })); state .write_path(path.as_ref(), move |e| match e { btree_map::Entry::Vacant(e) => { e.insert(file); Ok(()) } btree_map::Entry::Occupied(mut e) => { *e.get_mut() = file; Ok(()) } }) .unwrap(); state.emit_event(&[path]); } fn write_file_internal(&self, path: impl AsRef, content: String) -> Result<()> { let mut state = self.state.lock(); let path = path.as_ref(); let inode = state.next_inode; let mtime = state.next_mtime; state.next_inode += 1; state.next_mtime += Duration::from_nanos(1); let file = Arc::new(Mutex::new(FakeFsEntry::File { inode, mtime, content, })); state.write_path(path, move |entry| { match entry { btree_map::Entry::Vacant(e) => { e.insert(file); } btree_map::Entry::Occupied(mut e) => { *e.get_mut() = file; } } Ok(()) })?; state.emit_event(&[path]); Ok(()) } pub fn pause_events(&self) { self.state.lock().events_paused = true; } pub fn buffered_event_count(&self) -> usize { self.state.lock().buffered_events.len() } pub fn flush_events(&self, count: usize) { self.state.lock().flush_events(count); } #[must_use] pub fn insert_tree<'a>( &'a self, path: impl 'a + AsRef + Send, tree: serde_json::Value, ) -> futures::future::BoxFuture<'a, ()> { use futures::FutureExt as _; use serde_json::Value::*; async move { let path = path.as_ref(); match tree { Object(map) => { self.create_dir(path).await.unwrap(); for (name, contents) in map { let mut path = PathBuf::from(path); path.push(name); self.insert_tree(&path, contents).await; } } Null => { self.create_dir(path).await.unwrap(); } String(contents) => { self.insert_file(&path, contents).await; } _ => { panic!("JSON object must contain only objects, strings, or null"); } } } .boxed() } pub fn with_git_state(&self, dot_git: &Path, emit_git_event: bool, f: F) where F: FnOnce(&mut FakeGitRepositoryState), { let mut state = self.state.lock(); let entry = state.read_path(dot_git).unwrap(); let mut entry = entry.lock(); if let FakeFsEntry::Dir { git_repo_state, .. } = &mut *entry { let repo_state = git_repo_state.get_or_insert_with(Default::default); let mut repo_state = repo_state.lock(); f(&mut repo_state); if emit_git_event { state.emit_event([dot_git]); } } else { panic!("not a directory"); } } pub fn set_branch_name(&self, dot_git: &Path, branch: Option>) { self.with_git_state(dot_git, true, |state| { state.branch_name = branch.map(Into::into) }) } pub fn set_index_for_repo(&self, dot_git: &Path, head_state: &[(&Path, String)]) { self.with_git_state(dot_git, true, |state| { state.index_contents.clear(); state.index_contents.extend( head_state .iter() .map(|(path, content)| (path.to_path_buf(), content.clone())), ); }); } pub fn set_status_for_repo_via_working_copy_change( &self, dot_git: &Path, statuses: &[(&Path, GitFileStatus)], ) { self.with_git_state(dot_git, false, |state| { state.worktree_statuses.clear(); state.worktree_statuses.extend( statuses .iter() .map(|(path, content)| ((**path).into(), content.clone())), ); }); self.state.lock().emit_event( statuses .iter() .map(|(path, _)| dot_git.parent().unwrap().join(path)), ); } pub fn set_status_for_repo_via_git_operation( &self, dot_git: &Path, statuses: &[(&Path, GitFileStatus)], ) { self.with_git_state(dot_git, true, |state| { state.worktree_statuses.clear(); state.worktree_statuses.extend( statuses .iter() .map(|(path, content)| ((**path).into(), content.clone())), ); }); } pub fn paths(&self, include_dot_git: bool) -> Vec { let mut result = Vec::new(); let mut queue = collections::VecDeque::new(); queue.push_back((PathBuf::from("/"), self.state.lock().root.clone())); while let Some((path, entry)) = queue.pop_front() { if let FakeFsEntry::Dir { entries, .. } = &*entry.lock() { for (name, entry) in entries { queue.push_back((path.join(name), entry.clone())); } } if include_dot_git || !path .components() .any(|component| component.as_os_str() == *FS_DOT_GIT) { result.push(path); } } result } pub fn directories(&self, include_dot_git: bool) -> Vec { let mut result = Vec::new(); let mut queue = collections::VecDeque::new(); queue.push_back((PathBuf::from("/"), self.state.lock().root.clone())); while let Some((path, entry)) = queue.pop_front() { if let FakeFsEntry::Dir { entries, .. } = &*entry.lock() { for (name, entry) in entries { queue.push_back((path.join(name), entry.clone())); } if include_dot_git || !path .components() .any(|component| component.as_os_str() == *FS_DOT_GIT) { result.push(path); } } } result } pub fn files(&self) -> Vec { let mut result = Vec::new(); let mut queue = collections::VecDeque::new(); queue.push_back((PathBuf::from("/"), self.state.lock().root.clone())); while let Some((path, entry)) = queue.pop_front() { let e = entry.lock(); match &*e { FakeFsEntry::File { .. } => result.push(path), FakeFsEntry::Dir { entries, .. } => { for (name, entry) in entries { queue.push_back((path.join(name), entry.clone())); } } FakeFsEntry::Symlink { .. } => {} } } result } /// How many `read_dir` calls have been issued. pub fn read_dir_call_count(&self) -> usize { self.state.lock().read_dir_call_count } /// How many `metadata` calls have been issued. pub fn metadata_call_count(&self) -> usize { self.state.lock().metadata_call_count } async fn simulate_random_delay(&self) { self.executor .upgrade() .expect("executor has been dropped") .simulate_random_delay() .await; } } #[cfg(any(test, feature = "test-support"))] impl FakeFsEntry { fn is_file(&self) -> bool { matches!(self, Self::File { .. }) } fn is_symlink(&self) -> bool { matches!(self, Self::Symlink { .. }) } fn file_content(&self, path: &Path) -> Result<&String> { if let Self::File { content, .. } = self { Ok(content) } else { Err(anyhow!("not a file: {}", path.display())) } } fn set_file_content(&mut self, path: &Path, new_content: String) -> Result<()> { if let Self::File { content, mtime, .. } = self { *mtime = SystemTime::now(); *content = new_content; Ok(()) } else { Err(anyhow!("not a file: {}", path.display())) } } fn dir_entries( &mut self, path: &Path, ) -> Result<&mut BTreeMap>>> { if let Self::Dir { entries, .. } = self { Ok(entries) } else { Err(anyhow!("not a directory: {}", path.display())) } } } #[cfg(any(test, feature = "test-support"))] #[async_trait::async_trait] impl Fs for FakeFs { async fn create_dir(&self, path: &Path) -> Result<()> { self.simulate_random_delay().await; let mut created_dirs = Vec::new(); let mut cur_path = PathBuf::new(); for component in path.components() { let mut state = self.state.lock(); cur_path.push(component); if cur_path == Path::new("/") { continue; } let inode = state.next_inode; let mtime = state.next_mtime; state.next_mtime += Duration::from_nanos(1); state.next_inode += 1; state.write_path(&cur_path, |entry| { entry.or_insert_with(|| { created_dirs.push(cur_path.clone()); Arc::new(Mutex::new(FakeFsEntry::Dir { inode, mtime, entries: Default::default(), git_repo_state: None, })) }); Ok(()) })? } self.state.lock().emit_event(&created_dirs); Ok(()) } async fn create_file(&self, path: &Path, options: CreateOptions) -> Result<()> { self.simulate_random_delay().await; let mut state = self.state.lock(); let inode = state.next_inode; let mtime = state.next_mtime; state.next_mtime += Duration::from_nanos(1); state.next_inode += 1; let file = Arc::new(Mutex::new(FakeFsEntry::File { inode, mtime, content: String::new(), })); state.write_path(path, |entry| { match entry { btree_map::Entry::Occupied(mut e) => { if options.overwrite { *e.get_mut() = file; } else if !options.ignore_if_exists { return Err(anyhow!("path already exists: {}", path.display())); } } btree_map::Entry::Vacant(e) => { e.insert(file); } } Ok(()) })?; state.emit_event(&[path]); Ok(()) } async fn rename(&self, old_path: &Path, new_path: &Path, options: RenameOptions) -> Result<()> { self.simulate_random_delay().await; let old_path = normalize_path(old_path); let new_path = normalize_path(new_path); let mut state = self.state.lock(); let moved_entry = state.write_path(&old_path, |e| { if let btree_map::Entry::Occupied(e) = e { Ok(e.get().clone()) } else { Err(anyhow!("path does not exist: {}", &old_path.display())) } })?; state.write_path(&new_path, |e| { match e { btree_map::Entry::Occupied(mut e) => { if options.overwrite { *e.get_mut() = moved_entry; } else if !options.ignore_if_exists { return Err(anyhow!("path already exists: {}", new_path.display())); } } btree_map::Entry::Vacant(e) => { e.insert(moved_entry); } } Ok(()) })?; state .write_path(&old_path, |e| { if let btree_map::Entry::Occupied(e) = e { Ok(e.remove()) } else { unreachable!() } }) .unwrap(); state.emit_event(&[old_path, new_path]); Ok(()) } async fn copy_file(&self, source: &Path, target: &Path, options: CopyOptions) -> Result<()> { self.simulate_random_delay().await; let source = normalize_path(source); let target = normalize_path(target); let mut state = self.state.lock(); let mtime = state.next_mtime; let inode = util::post_inc(&mut state.next_inode); state.next_mtime += Duration::from_nanos(1); let source_entry = state.read_path(&source)?; let content = source_entry.lock().file_content(&source)?.clone(); let entry = state.write_path(&target, |e| match e { btree_map::Entry::Occupied(e) => { if options.overwrite { Ok(Some(e.get().clone())) } else if !options.ignore_if_exists { return Err(anyhow!("{target:?} already exists")); } else { Ok(None) } } btree_map::Entry::Vacant(e) => Ok(Some( e.insert(Arc::new(Mutex::new(FakeFsEntry::File { inode, mtime, content: String::new(), }))) .clone(), )), })?; if let Some(entry) = entry { entry.lock().set_file_content(&target, content)?; } state.emit_event(&[target]); Ok(()) } async fn remove_dir(&self, path: &Path, options: RemoveOptions) -> Result<()> { self.simulate_random_delay().await; let path = normalize_path(path); let parent_path = path .parent() .ok_or_else(|| anyhow!("cannot remove the root"))?; let base_name = path.file_name().unwrap(); let mut state = self.state.lock(); let parent_entry = state.read_path(parent_path)?; let mut parent_entry = parent_entry.lock(); let entry = parent_entry .dir_entries(parent_path)? .entry(base_name.to_str().unwrap().into()); match entry { btree_map::Entry::Vacant(_) => { if !options.ignore_if_not_exists { return Err(anyhow!("{path:?} does not exist")); } } btree_map::Entry::Occupied(e) => { { let mut entry = e.get().lock(); let children = entry.dir_entries(&path)?; if !options.recursive && !children.is_empty() { return Err(anyhow!("{path:?} is not empty")); } } e.remove(); } } state.emit_event(&[path]); Ok(()) } async fn remove_file(&self, path: &Path, options: RemoveOptions) -> Result<()> { self.simulate_random_delay().await; let path = normalize_path(path); let parent_path = path .parent() .ok_or_else(|| anyhow!("cannot remove the root"))?; let base_name = path.file_name().unwrap(); let mut state = self.state.lock(); let parent_entry = state.read_path(parent_path)?; let mut parent_entry = parent_entry.lock(); let entry = parent_entry .dir_entries(parent_path)? .entry(base_name.to_str().unwrap().into()); match entry { btree_map::Entry::Vacant(_) => { if !options.ignore_if_not_exists { return Err(anyhow!("{path:?} does not exist")); } } btree_map::Entry::Occupied(e) => { e.get().lock().file_content(&path)?; e.remove(); } } state.emit_event(&[path]); Ok(()) } async fn open_sync(&self, path: &Path) -> Result> { let text = self.load(path).await?; Ok(Box::new(io::Cursor::new(text))) } async fn load(&self, path: &Path) -> Result { let path = normalize_path(path); self.simulate_random_delay().await; let state = self.state.lock(); let entry = state.read_path(&path)?; let entry = entry.lock(); entry.file_content(&path).cloned() } async fn atomic_write(&self, path: PathBuf, data: String) -> Result<()> { self.simulate_random_delay().await; let path = normalize_path(path.as_path()); self.write_file_internal(path, data.to_string())?; Ok(()) } async fn save(&self, path: &Path, text: &Rope, line_ending: LineEnding) -> Result<()> { self.simulate_random_delay().await; let path = normalize_path(path); let content = chunks(text, line_ending).collect(); if let Some(path) = path.parent() { self.create_dir(path).await?; } self.write_file_internal(path, content)?; Ok(()) } async fn canonicalize(&self, path: &Path) -> Result { let path = normalize_path(path); self.simulate_random_delay().await; let state = self.state.lock(); if let Some((_, canonical_path)) = state.try_read_path(&path, true) { Ok(canonical_path) } else { Err(anyhow!("path does not exist: {}", path.display())) } } async fn is_file(&self, path: &Path) -> bool { let path = normalize_path(path); self.simulate_random_delay().await; let state = self.state.lock(); if let Some((entry, _)) = state.try_read_path(&path, true) { entry.lock().is_file() } else { false } } async fn metadata(&self, path: &Path) -> Result> { self.simulate_random_delay().await; let path = normalize_path(path); let mut state = self.state.lock(); state.metadata_call_count += 1; if let Some((mut entry, _)) = state.try_read_path(&path, false) { let is_symlink = entry.lock().is_symlink(); if is_symlink { if let Some(e) = state.try_read_path(&path, true).map(|e| e.0) { entry = e; } else { return Ok(None); } } let entry = entry.lock(); Ok(Some(match &*entry { FakeFsEntry::File { inode, mtime, .. } => Metadata { inode: *inode, mtime: *mtime, is_dir: false, is_symlink, }, FakeFsEntry::Dir { inode, mtime, .. } => Metadata { inode: *inode, mtime: *mtime, is_dir: true, is_symlink, }, FakeFsEntry::Symlink { .. } => unreachable!(), })) } else { Ok(None) } } async fn read_link(&self, path: &Path) -> Result { self.simulate_random_delay().await; let path = normalize_path(path); let state = self.state.lock(); if let Some((entry, _)) = state.try_read_path(&path, false) { let entry = entry.lock(); if let FakeFsEntry::Symlink { target } = &*entry { Ok(target.clone()) } else { Err(anyhow!("not a symlink: {}", path.display())) } } else { Err(anyhow!("path does not exist: {}", path.display())) } } async fn read_dir( &self, path: &Path, ) -> Result>>>> { self.simulate_random_delay().await; let path = normalize_path(path); let mut state = self.state.lock(); state.read_dir_call_count += 1; let entry = state.read_path(&path)?; let mut entry = entry.lock(); let children = entry.dir_entries(&path)?; let paths = children .keys() .map(|file_name| Ok(path.join(file_name))) .collect::>(); Ok(Box::pin(futures::stream::iter(paths))) } async fn watch( &self, path: &Path, _: Duration, ) -> Pin>>> { self.simulate_random_delay().await; let (tx, rx) = smol::channel::unbounded(); self.state.lock().event_txs.push(tx); let path = path.to_path_buf(); let executor = self.executor.clone(); Box::pin(futures::StreamExt::filter(rx, move |events| { let result = events.iter().any(|event| event.path.starts_with(&path)); let executor = executor.clone(); async move { if let Some(executor) = executor.clone().upgrade() { executor.simulate_random_delay().await; } result } })) } fn open_repo(&self, abs_dot_git: &Path) -> Option>> { let state = self.state.lock(); let entry = state.read_path(abs_dot_git).unwrap(); let mut entry = entry.lock(); if let FakeFsEntry::Dir { git_repo_state, .. } = &mut *entry { let state = git_repo_state .get_or_insert_with(|| Arc::new(Mutex::new(FakeGitRepositoryState::default()))) .clone(); Some(repository::FakeGitRepository::open(state)) } else { None } } fn is_fake(&self) -> bool { true } #[cfg(any(test, feature = "test-support"))] fn as_fake(&self) -> &FakeFs { self } } fn chunks(rope: &Rope, line_ending: LineEnding) -> impl Iterator { rope.chunks().flat_map(move |chunk| { let mut newline = false; chunk.split('\n').flat_map(move |line| { let ending = if newline { Some(line_ending.as_str()) } else { None }; newline = true; ending.into_iter().chain([line]) }) }) } pub fn normalize_path(path: &Path) -> PathBuf { let mut components = path.components().peekable(); let mut ret = if let Some(c @ Component::Prefix(..)) = components.peek().cloned() { components.next(); PathBuf::from(c.as_os_str()) } else { PathBuf::new() }; for component in components { match component { Component::Prefix(..) => unreachable!(), Component::RootDir => { ret.push(component.as_os_str()); } Component::CurDir => {} Component::ParentDir => { ret.pop(); } Component::Normal(c) => { ret.push(c); } } } ret } pub fn copy_recursive<'a>( fs: &'a dyn Fs, source: &'a Path, target: &'a Path, options: CopyOptions, ) -> BoxFuture<'a, Result<()>> { use futures::future::FutureExt; async move { let metadata = fs .metadata(source) .await? .ok_or_else(|| anyhow!("path does not exist: {}", source.display()))?; if metadata.is_dir { if !options.overwrite && fs.metadata(target).await.is_ok() { if options.ignore_if_exists { return Ok(()); } else { return Err(anyhow!("{target:?} already exists")); } } let _ = fs .remove_dir( target, RemoveOptions { recursive: true, ignore_if_not_exists: true, }, ) .await; fs.create_dir(target).await?; let mut children = fs.read_dir(source).await?; while let Some(child_path) = children.next().await { if let Ok(child_path) = child_path { if let Some(file_name) = child_path.file_name() { let child_target_path = target.join(file_name); copy_recursive(fs, &child_path, &child_target_path, options).await?; } } } Ok(()) } else { fs.copy_file(source, target, options).await } } .boxed() } #[cfg(test)] mod tests { use super::*; use gpui::TestAppContext; use serde_json::json; #[gpui::test] async fn test_fake_fs(cx: &mut TestAppContext) { let fs = FakeFs::new(cx.background()); fs.insert_tree( "/root", json!({ "dir1": { "a": "A", "b": "B" }, "dir2": { "c": "C", "dir3": { "d": "D" } } }), ) .await; assert_eq!( fs.files(), vec![ PathBuf::from("/root/dir1/a"), PathBuf::from("/root/dir1/b"), PathBuf::from("/root/dir2/c"), PathBuf::from("/root/dir2/dir3/d"), ] ); fs.insert_symlink("/root/dir2/link-to-dir3", "./dir3".into()) .await; assert_eq!( fs.canonicalize("/root/dir2/link-to-dir3".as_ref()) .await .unwrap(), PathBuf::from("/root/dir2/dir3"), ); assert_eq!( fs.canonicalize("/root/dir2/link-to-dir3/d".as_ref()) .await .unwrap(), PathBuf::from("/root/dir2/dir3/d"), ); assert_eq!( fs.load("/root/dir2/link-to-dir3/d".as_ref()).await.unwrap(), "D", ); } }