use crate::changelog::Changelog; use crate::config::{Config, ConfigError, ConfigParseError}; use crate::dirstate::DirstateParents; use crate::dirstate_tree::dirstate_map::DirstateMapWriteMode; use crate::dirstate_tree::on_disk::Docket as DirstateDocket; use crate::dirstate_tree::owning::OwningDirstateMap; use crate::errors::HgResultExt; use crate::errors::{HgError, IoResultExt}; use crate::lock::{try_with_lock_no_wait, LockError}; use crate::manifest::{Manifest, Manifestlog}; use crate::revlog::filelog::Filelog; use crate::revlog::RevlogError; use crate::utils::debug::debug_wait_for_file_or_print; use crate::utils::files::get_path_from_bytes; use crate::utils::hg_path::HgPath; use crate::utils::SliceExt; use crate::vfs::{is_dir, is_file, Vfs}; use crate::DirstateError; use crate::{requirements, NodePrefix, UncheckedRevision}; use std::cell::{Ref, RefCell, RefMut}; use std::collections::HashSet; use std::io::Seek; use std::io::SeekFrom; use std::io::Write as IoWrite; use std::path::{Path, PathBuf}; const V2_MAX_READ_ATTEMPTS: usize = 5; type DirstateMapIdentity = (Option, Option>, usize); /// A repository on disk pub struct Repo { working_directory: PathBuf, dot_hg: PathBuf, store: PathBuf, requirements: HashSet, config: Config, dirstate_parents: LazyCell, dirstate_map: LazyCell, changelog: LazyCell, manifestlog: LazyCell, } #[derive(Debug, derive_more::From)] pub enum RepoError { NotFound { at: PathBuf, }, #[from] ConfigParseError(ConfigParseError), #[from] Other(HgError), } impl From for RepoError { fn from(error: ConfigError) -> Self { match error { ConfigError::Parse(error) => error.into(), ConfigError::Other(error) => error.into(), } } } impl Repo { /// tries to find nearest repository root in current working directory or /// its ancestors pub fn find_repo_root() -> Result { let current_directory = crate::utils::current_dir()?; // ancestors() is inclusive: it first yields `current_directory` // as-is. for ancestor in current_directory.ancestors() { if is_dir(ancestor.join(".hg"))? { return Ok(ancestor.to_path_buf()); } } Err(RepoError::NotFound { at: current_directory, }) } /// Find a repository, either at the given path (which must contain a `.hg` /// sub-directory) or by searching the current directory and its /// ancestors. /// /// A method with two very different "modes" like this usually a code smell /// to make two methods instead, but in this case an `Option` is what rhg /// sub-commands get from Clap for the `-R` / `--repository` CLI argument. /// Having two methods would just move that `if` to almost all callers. pub fn find( config: &Config, explicit_path: Option, ) -> Result { if let Some(root) = explicit_path { if is_dir(root.join(".hg"))? { Self::new_at_path(root, config) } else if is_file(&root)? { Err(HgError::unsupported("bundle repository").into()) } else { Err(RepoError::NotFound { at: root }) } } else { let root = Self::find_repo_root()?; Self::new_at_path(root, config) } } /// To be called after checking that `.hg` is a sub-directory fn new_at_path( working_directory: PathBuf, config: &Config, ) -> Result { let dot_hg = working_directory.join(".hg"); let mut repo_config_files = vec![dot_hg.join("hgrc"), dot_hg.join("hgrc-not-shared")]; let hg_vfs = Vfs { base: &dot_hg }; let mut reqs = requirements::load_if_exists(hg_vfs)?; let relative = reqs.contains(requirements::RELATIVE_SHARED_REQUIREMENT); let shared = reqs.contains(requirements::SHARED_REQUIREMENT) || relative; // From `mercurial/localrepo.py`: // // if .hg/requires contains the sharesafe requirement, it means // there exists a `.hg/store/requires` too and we should read it // NOTE: presence of SHARESAFE_REQUIREMENT imply that store requirement // is present. We never write SHARESAFE_REQUIREMENT for a repo if store // is not present, refer checkrequirementscompat() for that // // However, if SHARESAFE_REQUIREMENT is not present, it means that the // repository was shared the old way. We check the share source // .hg/requires for SHARESAFE_REQUIREMENT to detect whether the // current repository needs to be reshared let share_safe = reqs.contains(requirements::SHARESAFE_REQUIREMENT); let store_path; if !shared { store_path = dot_hg.join("store"); } else { let bytes = hg_vfs.read("sharedpath")?; let mut shared_path = get_path_from_bytes(bytes.trim_end_matches(|b| b == b'\n')) .to_owned(); if relative { shared_path = dot_hg.join(shared_path) } if !is_dir(&shared_path)? { return Err(HgError::corrupted(format!( ".hg/sharedpath points to nonexistent directory {}", shared_path.display() )) .into()); } store_path = shared_path.join("store"); let source_is_share_safe = requirements::load(Vfs { base: &shared_path })? .contains(requirements::SHARESAFE_REQUIREMENT); if share_safe != source_is_share_safe { return Err(HgError::unsupported("share-safe mismatch").into()); } if share_safe { repo_config_files.insert(0, shared_path.join("hgrc")) } } if share_safe { reqs.extend(requirements::load(Vfs { base: &store_path })?); } let repo_config = if std::env::var_os("HGRCSKIPREPO").is_none() { config.combine_with_repo(&repo_config_files)? } else { config.clone() }; let repo = Self { requirements: reqs, working_directory, store: store_path, dot_hg, config: repo_config, dirstate_parents: LazyCell::new(), dirstate_map: LazyCell::new(), changelog: LazyCell::new(), manifestlog: LazyCell::new(), }; requirements::check(&repo)?; Ok(repo) } pub fn working_directory_path(&self) -> &Path { &self.working_directory } pub fn requirements(&self) -> &HashSet { &self.requirements } pub fn config(&self) -> &Config { &self.config } /// For accessing repository files (in `.hg`), except for the store /// (`.hg/store`). pub fn hg_vfs(&self) -> Vfs<'_> { Vfs { base: &self.dot_hg } } /// For accessing repository store files (in `.hg/store`) pub fn store_vfs(&self) -> Vfs<'_> { Vfs { base: &self.store } } /// For accessing the working copy pub fn working_directory_vfs(&self) -> Vfs<'_> { Vfs { base: &self.working_directory, } } pub fn try_with_wlock_no_wait( &self, f: impl FnOnce() -> R, ) -> Result { try_with_lock_no_wait(self.hg_vfs(), "wlock", f) } /// Whether this repo should use dirstate-v2. /// The presence of `dirstate-v2` in the requirements does not mean that /// the on-disk dirstate is necessarily in version 2. In most cases, /// a dirstate-v2 file will indeed be found, but in rare cases (like the /// upgrade mechanism being cut short), the on-disk version will be a /// v1 file. /// Semantically, having a requirement only means that a client cannot /// properly understand or properly update the repo if it lacks the support /// for the required feature, but not that that feature is actually used /// in all occasions. pub fn use_dirstate_v2(&self) -> bool { self.requirements .contains(requirements::DIRSTATE_V2_REQUIREMENT) } pub fn has_sparse(&self) -> bool { self.requirements.contains(requirements::SPARSE_REQUIREMENT) } pub fn has_narrow(&self) -> bool { self.requirements.contains(requirements::NARROW_REQUIREMENT) } pub fn has_nodemap(&self) -> bool { self.requirements .contains(requirements::NODEMAP_REQUIREMENT) } fn dirstate_file_contents(&self) -> Result, HgError> { Ok(self .hg_vfs() .read("dirstate") .io_not_found_as_none()? .unwrap_or_default()) } fn dirstate_identity(&self) -> Result, HgError> { use std::os::unix::fs::MetadataExt; Ok(self .hg_vfs() .symlink_metadata("dirstate") .io_not_found_as_none()? .map(|meta| meta.ino())) } pub fn dirstate_parents(&self) -> Result { Ok(*self .dirstate_parents .get_or_init(|| self.read_dirstate_parents())?) } fn read_dirstate_parents(&self) -> Result { let dirstate = self.dirstate_file_contents()?; let parents = if dirstate.is_empty() { DirstateParents::NULL } else if self.use_dirstate_v2() { let docket_res = crate::dirstate_tree::on_disk::read_docket(&dirstate); match docket_res { Ok(docket) => docket.parents(), Err(_) => { log::info!( "Parsing dirstate docket failed, \ falling back to dirstate-v1" ); *crate::dirstate::parsers::parse_dirstate_parents( &dirstate, )? } } } else { *crate::dirstate::parsers::parse_dirstate_parents(&dirstate)? }; self.dirstate_parents.set(parents); Ok(parents) } /// Returns the information read from the dirstate docket necessary to /// check if the data file has been updated/deleted by another process /// since we last read the dirstate. /// Namely, the inode, data file uuid and the data size. fn get_dirstate_data_file_integrity( &self, ) -> Result { assert!( self.use_dirstate_v2(), "accessing dirstate data file ID without dirstate-v2" ); // Get the identity before the contents since we could have a race // between the two. Having an identity that is too old is fine, but // one that is younger than the content change is bad. let identity = self.dirstate_identity()?; let dirstate = self.dirstate_file_contents()?; if dirstate.is_empty() { self.dirstate_parents.set(DirstateParents::NULL); Ok((identity, None, 0)) } else { let docket_res = crate::dirstate_tree::on_disk::read_docket(&dirstate); match docket_res { Ok(docket) => { self.dirstate_parents.set(docket.parents()); Ok(( identity, Some(docket.uuid.to_owned()), docket.data_size(), )) } Err(_) => { log::info!( "Parsing dirstate docket failed, \ falling back to dirstate-v1" ); let parents = *crate::dirstate::parsers::parse_dirstate_parents( &dirstate, )?; self.dirstate_parents.set(parents); Ok((identity, None, 0)) } } } } fn new_dirstate_map(&self) -> Result { if self.use_dirstate_v2() { // The v2 dirstate is split into a docket and a data file. // Since we don't always take the `wlock` to read it // (like in `hg status`), it is susceptible to races. // A simple retry method should be enough since full rewrites // only happen when too much garbage data is present and // this race is unlikely. let mut tries = 0; while tries < V2_MAX_READ_ATTEMPTS { tries += 1; match self.read_docket_and_data_file() { Ok(m) => { return Ok(m); } Err(e) => match e { DirstateError::Common(HgError::RaceDetected( context, )) => { log::info!( "dirstate read race detected {} (retry {}/{})", context, tries, V2_MAX_READ_ATTEMPTS, ); continue; } _ => { log::info!( "Reading dirstate v2 failed, \ falling back to v1" ); return self.new_dirstate_map_v1(); } }, } } let error = HgError::abort( format!("dirstate read race happened {tries} times in a row"), 255, None, ); Err(DirstateError::Common(error)) } else { self.new_dirstate_map_v1() } } fn new_dirstate_map_v1(&self) -> Result { debug_wait_for_file_or_print(self.config(), "dirstate.pre-read-file"); let identity = self.dirstate_identity()?; let dirstate_file_contents = self.dirstate_file_contents()?; if dirstate_file_contents.is_empty() { self.dirstate_parents.set(DirstateParents::NULL); Ok(OwningDirstateMap::new_empty(Vec::new())) } else { let (map, parents) = OwningDirstateMap::new_v1(dirstate_file_contents, identity)?; self.dirstate_parents.set(parents); Ok(map) } } fn read_docket_and_data_file( &self, ) -> Result { debug_wait_for_file_or_print(self.config(), "dirstate.pre-read-file"); let dirstate_file_contents = self.dirstate_file_contents()?; let identity = self.dirstate_identity()?; if dirstate_file_contents.is_empty() { self.dirstate_parents.set(DirstateParents::NULL); return Ok(OwningDirstateMap::new_empty(Vec::new())); } let docket = crate::dirstate_tree::on_disk::read_docket( &dirstate_file_contents, )?; debug_wait_for_file_or_print( self.config(), "dirstate.post-docket-read-file", ); self.dirstate_parents.set(docket.parents()); let uuid = docket.uuid.to_owned(); let data_size = docket.data_size(); let context = "between reading dirstate docket and data file"; let race_error = HgError::RaceDetected(context.into()); let metadata = docket.tree_metadata(); let mut map = if crate::vfs::is_on_nfs_mount(docket.data_filename()) { // Don't mmap on NFS to prevent `SIGBUS` error on deletion let contents = self.hg_vfs().read(docket.data_filename()); let contents = match contents { Ok(c) => c, Err(HgError::IoError { error, context }) => { match error.raw_os_error().expect("real os error") { // 2 = ENOENT, No such file or directory // 116 = ESTALE, Stale NFS file handle // // TODO match on `error.kind()` when // `ErrorKind::StaleNetworkFileHandle` is stable. 2 | 116 => { // Race where the data file was deleted right after // we read the docket, try again return Err(race_error.into()); } _ => { return Err( HgError::IoError { error, context }.into() ) } } } Err(e) => return Err(e.into()), }; OwningDirstateMap::new_v2( contents, data_size, metadata, uuid, identity, ) } else { match self .hg_vfs() .mmap_open(docket.data_filename()) .io_not_found_as_none() { Ok(Some(data_mmap)) => OwningDirstateMap::new_v2( data_mmap, data_size, metadata, uuid, identity, ), Ok(None) => { // Race where the data file was deleted right after we // read the docket, try again return Err(race_error.into()); } Err(e) => return Err(e.into()), } }?; let write_mode_config = self .config() .get_str(b"devel", b"dirstate.v2.data_update_mode") .unwrap_or(Some("auto")) .unwrap_or("auto"); // don't bother for devel options let write_mode = match write_mode_config { "auto" => DirstateMapWriteMode::Auto, "force-new" => DirstateMapWriteMode::ForceNewDataFile, "force-append" => DirstateMapWriteMode::ForceAppend, _ => DirstateMapWriteMode::Auto, }; map.with_dmap_mut(|m| m.set_write_mode(write_mode)); Ok(map) } pub fn dirstate_map( &self, ) -> Result, DirstateError> { self.dirstate_map.get_or_init(|| self.new_dirstate_map()) } pub fn dirstate_map_mut( &self, ) -> Result, DirstateError> { self.dirstate_map .get_mut_or_init(|| self.new_dirstate_map()) } fn new_changelog(&self) -> Result { Changelog::open(&self.store_vfs(), self.has_nodemap()) } pub fn changelog(&self) -> Result, HgError> { self.changelog.get_or_init(|| self.new_changelog()) } pub fn changelog_mut(&self) -> Result, HgError> { self.changelog.get_mut_or_init(|| self.new_changelog()) } fn new_manifestlog(&self) -> Result { Manifestlog::open(&self.store_vfs(), self.has_nodemap()) } pub fn manifestlog(&self) -> Result, HgError> { self.manifestlog.get_or_init(|| self.new_manifestlog()) } pub fn manifestlog_mut(&self) -> Result, HgError> { self.manifestlog.get_mut_or_init(|| self.new_manifestlog()) } /// Returns the manifest of the *changeset* with the given node ID pub fn manifest_for_node( &self, node: impl Into, ) -> Result { self.manifestlog()?.data_for_node( self.changelog()? .data_for_node(node.into())? .manifest_node()? .into(), ) } /// Returns the manifest of the *changeset* with the given revision number pub fn manifest_for_rev( &self, revision: UncheckedRevision, ) -> Result { self.manifestlog()?.data_for_node( self.changelog()? .data_for_rev(revision)? .manifest_node()? .into(), ) } pub fn has_subrepos(&self) -> Result { if let Some(entry) = self.dirstate_map()?.get(HgPath::new(".hgsub"))? { Ok(entry.tracked()) } else { Ok(false) } } pub fn filelog(&self, path: &HgPath) -> Result { Filelog::open(self, path) } /// Write to disk any updates that were made through `dirstate_map_mut`. /// /// The "wlock" must be held while calling this. /// See for example `try_with_wlock_no_wait`. /// /// TODO: have a `WritableRepo` type only accessible while holding the /// lock? pub fn write_dirstate(&self) -> Result<(), DirstateError> { let map = self.dirstate_map()?; // TODO: Maintain a `DirstateMap::dirty` flag, and return early here if // it’s unset let parents = self.dirstate_parents()?; let (packed_dirstate, old_uuid_to_remove) = if self.use_dirstate_v2() { let (identity, uuid, data_size) = self.get_dirstate_data_file_integrity()?; let identity_changed = identity != map.old_identity(); let uuid_changed = uuid.as_deref() != map.old_uuid(); let data_length_changed = data_size != map.old_data_size(); if identity_changed || uuid_changed || data_length_changed { // If any of identity, uuid or length have changed since // last disk read, don't write. // This is fine because either we're in a command that doesn't // write anything too important (like `hg status`), or we're in // `hg add` and we're supposed to have taken the lock before // reading anyway. // // TODO complain loudly if we've changed anything important // without taking the lock. // (see `hg help config.format.use-dirstate-tracked-hint`) log::debug!( "dirstate has changed since last read, not updating." ); return Ok(()); } let uuid_opt = map.old_uuid(); let write_mode = if uuid_opt.is_some() { DirstateMapWriteMode::Auto } else { DirstateMapWriteMode::ForceNewDataFile }; let (data, tree_metadata, append, old_data_size) = map.pack_v2(write_mode)?; // Reuse the uuid, or generate a new one, keeping the old for // deletion. let (uuid, old_uuid) = match uuid_opt { Some(uuid) => { let as_str = std::str::from_utf8(uuid) .map_err(|_| { HgError::corrupted( "non-UTF-8 dirstate data file ID", ) })? .to_owned(); if append { (as_str, None) } else { (DirstateDocket::new_uid(), Some(as_str)) } } None => (DirstateDocket::new_uid(), None), }; let data_filename = format!("dirstate.{}", uuid); let data_filename = self.hg_vfs().join(data_filename); let mut options = std::fs::OpenOptions::new(); options.write(true); // Why are we not using the O_APPEND flag when appending? // // - O_APPEND makes it trickier to deal with garbage at the end of // the file, left by a previous uncommitted transaction. By // starting the write at [old_data_size] we make sure we erase // all such garbage. // // - O_APPEND requires to special-case 0-byte writes, whereas we // don't need that. // // - Some OSes have bugs in implementation O_APPEND: // revlog.py talks about a Solaris bug, but we also saw some ZFS // bug: https://github.com/openzfs/zfs/pull/3124, // https://github.com/openzfs/zfs/issues/13370 // if !append { log::trace!("creating a new dirstate data file"); options.create_new(true); } else { log::trace!("appending to the dirstate data file"); } let data_size = (|| { // TODO: loop and try another random ID if !append and this // returns `ErrorKind::AlreadyExists`? Collision chance of two // random IDs is one in 2**32 let mut file = options.open(&data_filename)?; if append { file.seek(SeekFrom::Start(old_data_size as u64))?; } file.write_all(&data)?; file.flush()?; file.stream_position() })() .when_writing_file(&data_filename)?; let packed_dirstate = DirstateDocket::serialize( parents, tree_metadata, data_size, uuid.as_bytes(), ) .map_err(|_: std::num::TryFromIntError| { HgError::corrupted("overflow in dirstate docket serialization") })?; (packed_dirstate, old_uuid) } else { let identity = self.dirstate_identity()?; if identity != map.old_identity() { // If identity changed since last disk read, don't write. // This is fine because either we're in a command that doesn't // write anything too important (like `hg status`), or we're in // `hg add` and we're supposed to have taken the lock before // reading anyway. // // TODO complain loudly if we've changed anything important // without taking the lock. // (see `hg help config.format.use-dirstate-tracked-hint`) log::debug!( "dirstate has changed since last read, not updating." ); return Ok(()); } (map.pack_v1(parents)?, None) }; let vfs = self.hg_vfs(); vfs.atomic_write("dirstate", &packed_dirstate)?; if let Some(uuid) = old_uuid_to_remove { // Remove the old data file after the new docket pointing to the // new data file was written. vfs.remove_file(format!("dirstate.{}", uuid))?; } Ok(()) } } /// Lazily-initialized component of `Repo` with interior mutability /// /// This differs from `OnceCell` in that the value can still be "deinitialized" /// later by setting its inner `Option` to `None`. It also takes the /// initialization function as an argument when the value is requested, not /// when the instance is created. struct LazyCell { value: RefCell>, } impl LazyCell { fn new() -> Self { Self { value: RefCell::new(None), } } fn set(&self, value: T) { *self.value.borrow_mut() = Some(value) } fn get_or_init( &self, init: impl Fn() -> Result, ) -> Result, E> { let mut borrowed = self.value.borrow(); if borrowed.is_none() { drop(borrowed); // Only use `borrow_mut` if it is really needed to avoid panic in // case there is another outstanding borrow but mutation is not // needed. *self.value.borrow_mut() = Some(init()?); borrowed = self.value.borrow() } Ok(Ref::map(borrowed, |option| option.as_ref().unwrap())) } fn get_mut_or_init( &self, init: impl Fn() -> Result, ) -> Result, E> { let mut borrowed = self.value.borrow_mut(); if borrowed.is_none() { *borrowed = Some(init()?); } Ok(RefMut::map(borrowed, |option| option.as_mut().unwrap())) } }