|
|
# -*- coding: utf-8 -*-
|
|
|
|
|
|
# Copyright (C) 2014-2020 RhodeCode GmbH
|
|
|
#
|
|
|
# This program is free software: you can redistribute it and/or modify
|
|
|
# it under the terms of the GNU Affero General Public License, version 3
|
|
|
# (only), as published by the Free Software Foundation.
|
|
|
#
|
|
|
# This program is distributed in the hope that it will be useful,
|
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
# GNU General Public License for more details.
|
|
|
#
|
|
|
# You should have received a copy of the GNU Affero General Public License
|
|
|
# along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
|
#
|
|
|
# This program is dual-licensed. If you wish to learn more about the
|
|
|
# RhodeCode Enterprise Edition, including its added features, Support services,
|
|
|
# and proprietary license terms, please see https://rhodecode.com/licenses/
|
|
|
|
|
|
"""
|
|
|
HG repository module
|
|
|
"""
|
|
|
import os
|
|
|
import logging
|
|
|
import binascii
|
|
|
import urllib
|
|
|
|
|
|
from zope.cachedescriptors.property import Lazy as LazyProperty
|
|
|
|
|
|
from rhodecode.lib.compat import OrderedDict
|
|
|
from rhodecode.lib.datelib import (
|
|
|
date_to_timestamp_plus_offset, utcdate_fromtimestamp, makedate)
|
|
|
from rhodecode.lib.utils import safe_unicode, safe_str
|
|
|
from rhodecode.lib.utils2 import CachedProperty
|
|
|
from rhodecode.lib.vcs import connection, exceptions
|
|
|
from rhodecode.lib.vcs.backends.base import (
|
|
|
BaseRepository, CollectionGenerator, Config, MergeResponse,
|
|
|
MergeFailureReason, Reference, BasePathPermissionChecker)
|
|
|
from rhodecode.lib.vcs.backends.hg.commit import MercurialCommit
|
|
|
from rhodecode.lib.vcs.backends.hg.diff import MercurialDiff
|
|
|
from rhodecode.lib.vcs.backends.hg.inmemory import MercurialInMemoryCommit
|
|
|
from rhodecode.lib.vcs.exceptions import (
|
|
|
EmptyRepositoryError, RepositoryError, TagAlreadyExistError,
|
|
|
TagDoesNotExistError, CommitDoesNotExistError, SubrepoMergeError, UnresolvedFilesInRepo)
|
|
|
from rhodecode.lib.vcs.compat import configparser
|
|
|
|
|
|
hexlify = binascii.hexlify
|
|
|
nullid = "\0" * 20
|
|
|
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
class MercurialRepository(BaseRepository):
|
|
|
"""
|
|
|
Mercurial repository backend
|
|
|
"""
|
|
|
DEFAULT_BRANCH_NAME = 'default'
|
|
|
|
|
|
def __init__(self, repo_path, config=None, create=False, src_url=None,
|
|
|
do_workspace_checkout=False, with_wire=None, bare=False):
|
|
|
"""
|
|
|
Raises RepositoryError if repository could not be find at the given
|
|
|
``repo_path``.
|
|
|
|
|
|
:param repo_path: local path of the repository
|
|
|
:param config: config object containing the repo configuration
|
|
|
:param create=False: if set to True, would try to create repository if
|
|
|
it does not exist rather than raising exception
|
|
|
:param src_url=None: would try to clone repository from given location
|
|
|
:param do_workspace_checkout=False: sets update of working copy after
|
|
|
making a clone
|
|
|
:param bare: not used, compatible with other VCS
|
|
|
"""
|
|
|
|
|
|
self.path = safe_str(os.path.abspath(repo_path))
|
|
|
# mercurial since 4.4.X requires certain configuration to be present
|
|
|
# because sometimes we init the repos with config we need to meet
|
|
|
# special requirements
|
|
|
self.config = config if config else self.get_default_config(
|
|
|
default=[('extensions', 'largefiles', '1')])
|
|
|
self.with_wire = with_wire or {"cache": False} # default should not use cache
|
|
|
|
|
|
self._init_repo(create, src_url, do_workspace_checkout)
|
|
|
|
|
|
# caches
|
|
|
self._commit_ids = {}
|
|
|
|
|
|
@LazyProperty
|
|
|
def _remote(self):
|
|
|
repo_id = self.path
|
|
|
return connection.Hg(self.path, repo_id, self.config, with_wire=self.with_wire)
|
|
|
|
|
|
@CachedProperty
|
|
|
def commit_ids(self):
|
|
|
"""
|
|
|
Returns list of commit ids, in ascending order. Being lazy
|
|
|
attribute allows external tools to inject shas from cache.
|
|
|
"""
|
|
|
commit_ids = self._get_all_commit_ids()
|
|
|
self._rebuild_cache(commit_ids)
|
|
|
return commit_ids
|
|
|
|
|
|
def _rebuild_cache(self, commit_ids):
|
|
|
self._commit_ids = dict((commit_id, index)
|
|
|
for index, commit_id in enumerate(commit_ids))
|
|
|
|
|
|
@CachedProperty
|
|
|
def branches(self):
|
|
|
return self._get_branches()
|
|
|
|
|
|
@CachedProperty
|
|
|
def branches_closed(self):
|
|
|
return self._get_branches(active=False, closed=True)
|
|
|
|
|
|
@CachedProperty
|
|
|
def branches_all(self):
|
|
|
all_branches = {}
|
|
|
all_branches.update(self.branches)
|
|
|
all_branches.update(self.branches_closed)
|
|
|
return all_branches
|
|
|
|
|
|
def _get_branches(self, active=True, closed=False):
|
|
|
"""
|
|
|
Gets branches for this repository
|
|
|
Returns only not closed active branches by default
|
|
|
|
|
|
:param active: return also active branches
|
|
|
:param closed: return also closed branches
|
|
|
|
|
|
"""
|
|
|
if self.is_empty():
|
|
|
return {}
|
|
|
|
|
|
def get_name(ctx):
|
|
|
return ctx[0]
|
|
|
|
|
|
_branches = [(safe_unicode(n), hexlify(h),) for n, h in
|
|
|
self._remote.branches(active, closed).items()]
|
|
|
|
|
|
return OrderedDict(sorted(_branches, key=get_name, reverse=False))
|
|
|
|
|
|
@CachedProperty
|
|
|
def tags(self):
|
|
|
"""
|
|
|
Gets tags for this repository
|
|
|
"""
|
|
|
return self._get_tags()
|
|
|
|
|
|
def _get_tags(self):
|
|
|
if self.is_empty():
|
|
|
return {}
|
|
|
|
|
|
def get_name(ctx):
|
|
|
return ctx[0]
|
|
|
|
|
|
_tags = [(safe_unicode(n), hexlify(h),) for n, h in
|
|
|
self._remote.tags().items()]
|
|
|
|
|
|
return OrderedDict(sorted(_tags, key=get_name, reverse=True))
|
|
|
|
|
|
def tag(self, name, user, commit_id=None, message=None, date=None, **kwargs):
|
|
|
"""
|
|
|
Creates and returns a tag for the given ``commit_id``.
|
|
|
|
|
|
:param name: name for new tag
|
|
|
:param user: full username, i.e.: "Joe Doe <joe.doe@example.com>"
|
|
|
:param commit_id: commit id for which new tag would be created
|
|
|
:param message: message of the tag's commit
|
|
|
:param date: date of tag's commit
|
|
|
|
|
|
:raises TagAlreadyExistError: if tag with same name already exists
|
|
|
"""
|
|
|
if name in self.tags:
|
|
|
raise TagAlreadyExistError("Tag %s already exists" % name)
|
|
|
|
|
|
commit = self.get_commit(commit_id=commit_id)
|
|
|
local = kwargs.setdefault('local', False)
|
|
|
|
|
|
if message is None:
|
|
|
message = "Added tag %s for commit %s" % (name, commit.short_id)
|
|
|
|
|
|
date, tz = date_to_timestamp_plus_offset(date)
|
|
|
|
|
|
self._remote.tag(name, commit.raw_id, message, local, user, date, tz)
|
|
|
self._remote.invalidate_vcs_cache()
|
|
|
|
|
|
# Reinitialize tags
|
|
|
self._invalidate_prop_cache('tags')
|
|
|
tag_id = self.tags[name]
|
|
|
|
|
|
return self.get_commit(commit_id=tag_id)
|
|
|
|
|
|
def remove_tag(self, name, user, message=None, date=None):
|
|
|
"""
|
|
|
Removes tag with the given `name`.
|
|
|
|
|
|
:param name: name of the tag to be removed
|
|
|
:param user: full username, i.e.: "Joe Doe <joe.doe@example.com>"
|
|
|
:param message: message of the tag's removal commit
|
|
|
:param date: date of tag's removal commit
|
|
|
|
|
|
:raises TagDoesNotExistError: if tag with given name does not exists
|
|
|
"""
|
|
|
if name not in self.tags:
|
|
|
raise TagDoesNotExistError("Tag %s does not exist" % name)
|
|
|
|
|
|
if message is None:
|
|
|
message = "Removed tag %s" % name
|
|
|
local = False
|
|
|
|
|
|
date, tz = date_to_timestamp_plus_offset(date)
|
|
|
|
|
|
self._remote.tag(name, nullid, message, local, user, date, tz)
|
|
|
self._remote.invalidate_vcs_cache()
|
|
|
self._invalidate_prop_cache('tags')
|
|
|
|
|
|
@LazyProperty
|
|
|
def bookmarks(self):
|
|
|
"""
|
|
|
Gets bookmarks for this repository
|
|
|
"""
|
|
|
return self._get_bookmarks()
|
|
|
|
|
|
def _get_bookmarks(self):
|
|
|
if self.is_empty():
|
|
|
return {}
|
|
|
|
|
|
def get_name(ctx):
|
|
|
return ctx[0]
|
|
|
|
|
|
_bookmarks = [
|
|
|
(safe_unicode(n), hexlify(h)) for n, h in
|
|
|
self._remote.bookmarks().items()]
|
|
|
|
|
|
return OrderedDict(sorted(_bookmarks, key=get_name))
|
|
|
|
|
|
def _get_all_commit_ids(self):
|
|
|
return self._remote.get_all_commit_ids('visible')
|
|
|
|
|
|
def get_diff(
|
|
|
self, commit1, commit2, path='', ignore_whitespace=False,
|
|
|
context=3, path1=None):
|
|
|
"""
|
|
|
Returns (git like) *diff*, as plain text. Shows changes introduced by
|
|
|
`commit2` since `commit1`.
|
|
|
|
|
|
:param commit1: Entry point from which diff is shown. Can be
|
|
|
``self.EMPTY_COMMIT`` - in this case, patch showing all
|
|
|
the changes since empty state of the repository until `commit2`
|
|
|
:param commit2: Until which commit changes should be shown.
|
|
|
:param ignore_whitespace: If set to ``True``, would not show whitespace
|
|
|
changes. Defaults to ``False``.
|
|
|
:param context: How many lines before/after changed lines should be
|
|
|
shown. Defaults to ``3``.
|
|
|
"""
|
|
|
self._validate_diff_commits(commit1, commit2)
|
|
|
if path1 is not None and path1 != path:
|
|
|
raise ValueError("Diff of two different paths not supported.")
|
|
|
|
|
|
if path:
|
|
|
file_filter = [self.path, path]
|
|
|
else:
|
|
|
file_filter = None
|
|
|
|
|
|
diff = self._remote.diff(
|
|
|
commit1.raw_id, commit2.raw_id, file_filter=file_filter,
|
|
|
opt_git=True, opt_ignorews=ignore_whitespace,
|
|
|
context=context)
|
|
|
return MercurialDiff(diff)
|
|
|
|
|
|
def strip(self, commit_id, branch=None):
|
|
|
self._remote.strip(commit_id, update=False, backup="none")
|
|
|
|
|
|
self._remote.invalidate_vcs_cache()
|
|
|
# clear cache
|
|
|
self._invalidate_prop_cache('commit_ids')
|
|
|
|
|
|
return len(self.commit_ids)
|
|
|
|
|
|
def verify(self):
|
|
|
verify = self._remote.verify()
|
|
|
|
|
|
self._remote.invalidate_vcs_cache()
|
|
|
return verify
|
|
|
|
|
|
def hg_update_cache(self):
|
|
|
update_cache = self._remote.hg_update_cache()
|
|
|
|
|
|
self._remote.invalidate_vcs_cache()
|
|
|
return update_cache
|
|
|
|
|
|
def hg_rebuild_fn_cache(self):
|
|
|
update_cache = self._remote.hg_rebuild_fn_cache()
|
|
|
|
|
|
self._remote.invalidate_vcs_cache()
|
|
|
return update_cache
|
|
|
|
|
|
def get_common_ancestor(self, commit_id1, commit_id2, repo2):
|
|
|
if commit_id1 == commit_id2:
|
|
|
return commit_id1
|
|
|
|
|
|
ancestors = self._remote.revs_from_revspec(
|
|
|
"ancestor(id(%s), id(%s))", commit_id1, commit_id2,
|
|
|
other_path=repo2.path)
|
|
|
return repo2[ancestors[0]].raw_id if ancestors else None
|
|
|
|
|
|
def compare(self, commit_id1, commit_id2, repo2, merge, pre_load=None):
|
|
|
if commit_id1 == commit_id2:
|
|
|
commits = []
|
|
|
else:
|
|
|
if merge:
|
|
|
indexes = self._remote.revs_from_revspec(
|
|
|
"ancestors(id(%s)) - ancestors(id(%s)) - id(%s)",
|
|
|
commit_id2, commit_id1, commit_id1, other_path=repo2.path)
|
|
|
else:
|
|
|
indexes = self._remote.revs_from_revspec(
|
|
|
"id(%s)..id(%s) - id(%s)", commit_id1, commit_id2,
|
|
|
commit_id1, other_path=repo2.path)
|
|
|
|
|
|
commits = [repo2.get_commit(commit_idx=idx, pre_load=pre_load)
|
|
|
for idx in indexes]
|
|
|
|
|
|
return commits
|
|
|
|
|
|
@staticmethod
|
|
|
def check_url(url, config):
|
|
|
"""
|
|
|
Function will check given url and try to verify if it's a valid
|
|
|
link. Sometimes it may happened that mercurial will issue basic
|
|
|
auth request that can cause whole API to hang when used from python
|
|
|
or other external calls.
|
|
|
|
|
|
On failures it'll raise urllib2.HTTPError, exception is also thrown
|
|
|
when the return code is non 200
|
|
|
"""
|
|
|
# check first if it's not an local url
|
|
|
if os.path.isdir(url) or url.startswith('file:'):
|
|
|
return True
|
|
|
|
|
|
# Request the _remote to verify the url
|
|
|
return connection.Hg.check_url(url, config.serialize())
|
|
|
|
|
|
@staticmethod
|
|
|
def is_valid_repository(path):
|
|
|
return os.path.isdir(os.path.join(path, '.hg'))
|
|
|
|
|
|
def _init_repo(self, create, src_url=None, do_workspace_checkout=False):
|
|
|
"""
|
|
|
Function will check for mercurial repository in given path. If there
|
|
|
is no repository in that path it will raise an exception unless
|
|
|
`create` parameter is set to True - in that case repository would
|
|
|
be created.
|
|
|
|
|
|
If `src_url` is given, would try to clone repository from the
|
|
|
location at given clone_point. Additionally it'll make update to
|
|
|
working copy accordingly to `do_workspace_checkout` flag.
|
|
|
"""
|
|
|
if create and os.path.exists(self.path):
|
|
|
raise RepositoryError(
|
|
|
"Cannot create repository at %s, location already exist"
|
|
|
% self.path)
|
|
|
|
|
|
if src_url:
|
|
|
url = str(self._get_url(src_url))
|
|
|
MercurialRepository.check_url(url, self.config)
|
|
|
|
|
|
self._remote.clone(url, self.path, do_workspace_checkout)
|
|
|
|
|
|
# Don't try to create if we've already cloned repo
|
|
|
create = False
|
|
|
|
|
|
if create:
|
|
|
os.makedirs(self.path, mode=0o755)
|
|
|
self._remote.localrepository(create)
|
|
|
|
|
|
@LazyProperty
|
|
|
def in_memory_commit(self):
|
|
|
return MercurialInMemoryCommit(self)
|
|
|
|
|
|
@LazyProperty
|
|
|
def description(self):
|
|
|
description = self._remote.get_config_value(
|
|
|
'web', 'description', untrusted=True)
|
|
|
return safe_unicode(description or self.DEFAULT_DESCRIPTION)
|
|
|
|
|
|
@LazyProperty
|
|
|
def contact(self):
|
|
|
contact = (
|
|
|
self._remote.get_config_value("web", "contact") or
|
|
|
self._remote.get_config_value("ui", "username"))
|
|
|
return safe_unicode(contact or self.DEFAULT_CONTACT)
|
|
|
|
|
|
@LazyProperty
|
|
|
def last_change(self):
|
|
|
"""
|
|
|
Returns last change made on this repository as
|
|
|
`datetime.datetime` object.
|
|
|
"""
|
|
|
try:
|
|
|
return self.get_commit().date
|
|
|
except RepositoryError:
|
|
|
tzoffset = makedate()[1]
|
|
|
return utcdate_fromtimestamp(self._get_fs_mtime(), tzoffset)
|
|
|
|
|
|
def _get_fs_mtime(self):
|
|
|
# fallback to filesystem
|
|
|
cl_path = os.path.join(self.path, '.hg', "00changelog.i")
|
|
|
st_path = os.path.join(self.path, '.hg', "store")
|
|
|
if os.path.exists(cl_path):
|
|
|
return os.stat(cl_path).st_mtime
|
|
|
else:
|
|
|
return os.stat(st_path).st_mtime
|
|
|
|
|
|
def _get_url(self, url):
|
|
|
"""
|
|
|
Returns normalized url. If schema is not given, would fall
|
|
|
to filesystem
|
|
|
(``file:///``) schema.
|
|
|
"""
|
|
|
url = url.encode('utf8')
|
|
|
if url != 'default' and '://' not in url:
|
|
|
url = "file:" + urllib.pathname2url(url)
|
|
|
return url
|
|
|
|
|
|
def get_hook_location(self):
|
|
|
"""
|
|
|
returns absolute path to location where hooks are stored
|
|
|
"""
|
|
|
return os.path.join(self.path, '.hg', '.hgrc')
|
|
|
|
|
|
def get_commit(self, commit_id=None, commit_idx=None, pre_load=None,
|
|
|
translate_tag=None, maybe_unreachable=False):
|
|
|
"""
|
|
|
Returns ``MercurialCommit`` object representing repository's
|
|
|
commit at the given `commit_id` or `commit_idx`.
|
|
|
"""
|
|
|
if self.is_empty():
|
|
|
raise EmptyRepositoryError("There are no commits yet")
|
|
|
|
|
|
if commit_id is not None:
|
|
|
self._validate_commit_id(commit_id)
|
|
|
try:
|
|
|
# we have cached idx, use it without contacting the remote
|
|
|
idx = self._commit_ids[commit_id]
|
|
|
return MercurialCommit(self, commit_id, idx, pre_load=pre_load)
|
|
|
except KeyError:
|
|
|
pass
|
|
|
|
|
|
elif commit_idx is not None:
|
|
|
self._validate_commit_idx(commit_idx)
|
|
|
try:
|
|
|
_commit_id = self.commit_ids[commit_idx]
|
|
|
if commit_idx < 0:
|
|
|
commit_idx = self.commit_ids.index(_commit_id)
|
|
|
|
|
|
return MercurialCommit(self, _commit_id, commit_idx, pre_load=pre_load)
|
|
|
except IndexError:
|
|
|
commit_id = commit_idx
|
|
|
else:
|
|
|
commit_id = "tip"
|
|
|
|
|
|
if isinstance(commit_id, unicode):
|
|
|
commit_id = safe_str(commit_id)
|
|
|
|
|
|
try:
|
|
|
raw_id, idx = self._remote.lookup(commit_id, both=True)
|
|
|
except CommitDoesNotExistError:
|
|
|
msg = "Commit {} does not exist for `{}`".format(
|
|
|
*map(safe_str, [commit_id, self.name]))
|
|
|
raise CommitDoesNotExistError(msg)
|
|
|
|
|
|
return MercurialCommit(self, raw_id, idx, pre_load=pre_load)
|
|
|
|
|
|
def get_commits(
|
|
|
self, start_id=None, end_id=None, start_date=None, end_date=None,
|
|
|
branch_name=None, show_hidden=False, pre_load=None, translate_tags=None):
|
|
|
"""
|
|
|
Returns generator of ``MercurialCommit`` objects from start to end
|
|
|
(both are inclusive)
|
|
|
|
|
|
:param start_id: None, str(commit_id)
|
|
|
:param end_id: None, str(commit_id)
|
|
|
:param start_date: if specified, commits with commit date less than
|
|
|
``start_date`` would be filtered out from returned set
|
|
|
:param end_date: if specified, commits with commit date greater than
|
|
|
``end_date`` would be filtered out from returned set
|
|
|
:param branch_name: if specified, commits not reachable from given
|
|
|
branch would be filtered out from returned set
|
|
|
:param show_hidden: Show hidden commits such as obsolete or hidden from
|
|
|
Mercurial evolve
|
|
|
:raise BranchDoesNotExistError: If given ``branch_name`` does not
|
|
|
exist.
|
|
|
:raise CommitDoesNotExistError: If commit for given ``start`` or
|
|
|
``end`` could not be found.
|
|
|
"""
|
|
|
# actually we should check now if it's not an empty repo
|
|
|
if self.is_empty():
|
|
|
raise EmptyRepositoryError("There are no commits yet")
|
|
|
self._validate_branch_name(branch_name)
|
|
|
|
|
|
branch_ancestors = False
|
|
|
if start_id is not None:
|
|
|
self._validate_commit_id(start_id)
|
|
|
c_start = self.get_commit(commit_id=start_id)
|
|
|
start_pos = self._commit_ids[c_start.raw_id]
|
|
|
else:
|
|
|
start_pos = None
|
|
|
|
|
|
if end_id is not None:
|
|
|
self._validate_commit_id(end_id)
|
|
|
c_end = self.get_commit(commit_id=end_id)
|
|
|
end_pos = max(0, self._commit_ids[c_end.raw_id])
|
|
|
else:
|
|
|
end_pos = None
|
|
|
|
|
|
if None not in [start_id, end_id] and start_pos > end_pos:
|
|
|
raise RepositoryError(
|
|
|
"Start commit '%s' cannot be after end commit '%s'" %
|
|
|
(start_id, end_id))
|
|
|
|
|
|
if end_pos is not None:
|
|
|
end_pos += 1
|
|
|
|
|
|
commit_filter = []
|
|
|
|
|
|
if branch_name and not branch_ancestors:
|
|
|
commit_filter.append('branch("%s")' % (branch_name,))
|
|
|
elif branch_name and branch_ancestors:
|
|
|
commit_filter.append('ancestors(branch("%s"))' % (branch_name,))
|
|
|
|
|
|
if start_date and not end_date:
|
|
|
commit_filter.append('date(">%s")' % (start_date,))
|
|
|
if end_date and not start_date:
|
|
|
commit_filter.append('date("<%s")' % (end_date,))
|
|
|
if start_date and end_date:
|
|
|
commit_filter.append(
|
|
|
'date(">%s") and date("<%s")' % (start_date, end_date))
|
|
|
|
|
|
if not show_hidden:
|
|
|
commit_filter.append('not obsolete()')
|
|
|
commit_filter.append('not hidden()')
|
|
|
|
|
|
# TODO: johbo: Figure out a simpler way for this solution
|
|
|
collection_generator = CollectionGenerator
|
|
|
if commit_filter:
|
|
|
commit_filter = ' and '.join(map(safe_str, commit_filter))
|
|
|
revisions = self._remote.rev_range([commit_filter])
|
|
|
collection_generator = MercurialIndexBasedCollectionGenerator
|
|
|
else:
|
|
|
revisions = self.commit_ids
|
|
|
|
|
|
if start_pos or end_pos:
|
|
|
revisions = revisions[start_pos:end_pos]
|
|
|
|
|
|
return collection_generator(self, revisions, pre_load=pre_load)
|
|
|
|
|
|
def pull(self, url, commit_ids=None):
|
|
|
"""
|
|
|
Pull changes from external location.
|
|
|
|
|
|
:param commit_ids: Optional. Can be set to a list of commit ids
|
|
|
which shall be pulled from the other repository.
|
|
|
"""
|
|
|
url = self._get_url(url)
|
|
|
self._remote.pull(url, commit_ids=commit_ids)
|
|
|
self._remote.invalidate_vcs_cache()
|
|
|
|
|
|
def fetch(self, url, commit_ids=None):
|
|
|
"""
|
|
|
Backward compatibility with GIT fetch==pull
|
|
|
"""
|
|
|
return self.pull(url, commit_ids=commit_ids)
|
|
|
|
|
|
def push(self, url):
|
|
|
url = self._get_url(url)
|
|
|
self._remote.sync_push(url)
|
|
|
|
|
|
def _local_clone(self, clone_path):
|
|
|
"""
|
|
|
Create a local clone of the current repo.
|
|
|
"""
|
|
|
self._remote.clone(self.path, clone_path, update_after_clone=True,
|
|
|
hooks=False)
|
|
|
|
|
|
def _update(self, revision, clean=False):
|
|
|
"""
|
|
|
Update the working copy to the specified revision.
|
|
|
"""
|
|
|
log.debug('Doing checkout to commit: `%s` for %s', revision, self)
|
|
|
self._remote.update(revision, clean=clean)
|
|
|
|
|
|
def _identify(self):
|
|
|
"""
|
|
|
Return the current state of the working directory.
|
|
|
"""
|
|
|
return self._remote.identify().strip().rstrip('+')
|
|
|
|
|
|
def _heads(self, branch=None):
|
|
|
"""
|
|
|
Return the commit ids of the repository heads.
|
|
|
"""
|
|
|
return self._remote.heads(branch=branch).strip().split(' ')
|
|
|
|
|
|
def _ancestor(self, revision1, revision2):
|
|
|
"""
|
|
|
Return the common ancestor of the two revisions.
|
|
|
"""
|
|
|
return self._remote.ancestor(revision1, revision2)
|
|
|
|
|
|
def _local_push(
|
|
|
self, revision, repository_path, push_branches=False,
|
|
|
enable_hooks=False):
|
|
|
"""
|
|
|
Push the given revision to the specified repository.
|
|
|
|
|
|
:param push_branches: allow to create branches in the target repo.
|
|
|
"""
|
|
|
self._remote.push(
|
|
|
[revision], repository_path, hooks=enable_hooks,
|
|
|
push_branches=push_branches)
|
|
|
|
|
|
def _local_merge(self, target_ref, merge_message, user_name, user_email,
|
|
|
source_ref, use_rebase=False, dry_run=False):
|
|
|
"""
|
|
|
Merge the given source_revision into the checked out revision.
|
|
|
|
|
|
Returns the commit id of the merge and a boolean indicating if the
|
|
|
commit needs to be pushed.
|
|
|
"""
|
|
|
self._update(target_ref.commit_id, clean=True)
|
|
|
|
|
|
ancestor = self._ancestor(target_ref.commit_id, source_ref.commit_id)
|
|
|
is_the_same_branch = self._is_the_same_branch(target_ref, source_ref)
|
|
|
|
|
|
if ancestor == source_ref.commit_id:
|
|
|
# Nothing to do, the changes were already integrated
|
|
|
return target_ref.commit_id, False
|
|
|
|
|
|
elif ancestor == target_ref.commit_id and is_the_same_branch:
|
|
|
# In this case we should force a commit message
|
|
|
return source_ref.commit_id, True
|
|
|
|
|
|
unresolved = None
|
|
|
if use_rebase:
|
|
|
try:
|
|
|
bookmark_name = 'rcbook%s%s' % (source_ref.commit_id,
|
|
|
target_ref.commit_id)
|
|
|
self.bookmark(bookmark_name, revision=source_ref.commit_id)
|
|
|
self._remote.rebase(
|
|
|
source=source_ref.commit_id, dest=target_ref.commit_id)
|
|
|
self._remote.invalidate_vcs_cache()
|
|
|
self._update(bookmark_name, clean=True)
|
|
|
return self._identify(), True
|
|
|
except RepositoryError as e:
|
|
|
# The rebase-abort may raise another exception which 'hides'
|
|
|
# the original one, therefore we log it here.
|
|
|
log.exception('Error while rebasing shadow repo during merge.')
|
|
|
if 'unresolved conflicts' in safe_str(e):
|
|
|
unresolved = self._remote.get_unresolved_files()
|
|
|
log.debug('unresolved files: %s', unresolved)
|
|
|
|
|
|
# Cleanup any rebase leftovers
|
|
|
self._remote.invalidate_vcs_cache()
|
|
|
self._remote.rebase(abort=True)
|
|
|
self._remote.invalidate_vcs_cache()
|
|
|
self._remote.update(clean=True)
|
|
|
if unresolved:
|
|
|
raise UnresolvedFilesInRepo(unresolved)
|
|
|
else:
|
|
|
raise
|
|
|
else:
|
|
|
try:
|
|
|
self._remote.merge(source_ref.commit_id)
|
|
|
self._remote.invalidate_vcs_cache()
|
|
|
self._remote.commit(
|
|
|
message=safe_str(merge_message),
|
|
|
username=safe_str('%s <%s>' % (user_name, user_email)))
|
|
|
self._remote.invalidate_vcs_cache()
|
|
|
return self._identify(), True
|
|
|
except RepositoryError as e:
|
|
|
# The merge-abort may raise another exception which 'hides'
|
|
|
# the original one, therefore we log it here.
|
|
|
log.exception('Error while merging shadow repo during merge.')
|
|
|
if 'unresolved merge conflicts' in safe_str(e):
|
|
|
unresolved = self._remote.get_unresolved_files()
|
|
|
log.debug('unresolved files: %s', unresolved)
|
|
|
|
|
|
# Cleanup any merge leftovers
|
|
|
self._remote.update(clean=True)
|
|
|
if unresolved:
|
|
|
raise UnresolvedFilesInRepo(unresolved)
|
|
|
else:
|
|
|
raise
|
|
|
|
|
|
def _local_close(self, target_ref, user_name, user_email,
|
|
|
source_ref, close_message=''):
|
|
|
"""
|
|
|
Close the branch of the given source_revision
|
|
|
|
|
|
Returns the commit id of the close and a boolean indicating if the
|
|
|
commit needs to be pushed.
|
|
|
"""
|
|
|
self._update(source_ref.commit_id)
|
|
|
message = close_message or "Closing branch: `{}`".format(source_ref.name)
|
|
|
try:
|
|
|
self._remote.commit(
|
|
|
message=safe_str(message),
|
|
|
username=safe_str('%s <%s>' % (user_name, user_email)),
|
|
|
close_branch=True)
|
|
|
self._remote.invalidate_vcs_cache()
|
|
|
return self._identify(), True
|
|
|
except RepositoryError:
|
|
|
# Cleanup any commit leftovers
|
|
|
self._remote.update(clean=True)
|
|
|
raise
|
|
|
|
|
|
def _is_the_same_branch(self, target_ref, source_ref):
|
|
|
return (
|
|
|
self._get_branch_name(target_ref) ==
|
|
|
self._get_branch_name(source_ref))
|
|
|
|
|
|
def _get_branch_name(self, ref):
|
|
|
if ref.type == 'branch':
|
|
|
return ref.name
|
|
|
return self._remote.ctx_branch(ref.commit_id)
|
|
|
|
|
|
def _maybe_prepare_merge_workspace(
|
|
|
self, repo_id, workspace_id, unused_target_ref, unused_source_ref):
|
|
|
shadow_repository_path = self._get_shadow_repository_path(
|
|
|
self.path, repo_id, workspace_id)
|
|
|
if not os.path.exists(shadow_repository_path):
|
|
|
self._local_clone(shadow_repository_path)
|
|
|
log.debug(
|
|
|
'Prepared shadow repository in %s', shadow_repository_path)
|
|
|
|
|
|
return shadow_repository_path
|
|
|
|
|
|
def _merge_repo(self, repo_id, workspace_id, target_ref,
|
|
|
source_repo, source_ref, merge_message,
|
|
|
merger_name, merger_email, dry_run=False,
|
|
|
use_rebase=False, close_branch=False):
|
|
|
|
|
|
log.debug('Executing merge_repo with %s strategy, dry_run mode:%s',
|
|
|
'rebase' if use_rebase else 'merge', dry_run)
|
|
|
if target_ref.commit_id not in self._heads():
|
|
|
return MergeResponse(
|
|
|
False, False, None, MergeFailureReason.TARGET_IS_NOT_HEAD,
|
|
|
metadata={'target_ref': target_ref})
|
|
|
|
|
|
try:
|
|
|
if target_ref.type == 'branch' and len(self._heads(target_ref.name)) != 1:
|
|
|
heads = '\n,'.join(self._heads(target_ref.name))
|
|
|
metadata = {
|
|
|
'target_ref': target_ref,
|
|
|
'source_ref': source_ref,
|
|
|
'heads': heads
|
|
|
}
|
|
|
return MergeResponse(
|
|
|
False, False, None,
|
|
|
MergeFailureReason.HG_TARGET_HAS_MULTIPLE_HEADS,
|
|
|
metadata=metadata)
|
|
|
except CommitDoesNotExistError:
|
|
|
log.exception('Failure when looking up branch heads on hg target')
|
|
|
return MergeResponse(
|
|
|
False, False, None, MergeFailureReason.MISSING_TARGET_REF,
|
|
|
metadata={'target_ref': target_ref})
|
|
|
|
|
|
shadow_repository_path = self._maybe_prepare_merge_workspace(
|
|
|
repo_id, workspace_id, target_ref, source_ref)
|
|
|
shadow_repo = self.get_shadow_instance(shadow_repository_path)
|
|
|
|
|
|
log.debug('Pulling in target reference %s', target_ref)
|
|
|
self._validate_pull_reference(target_ref)
|
|
|
shadow_repo._local_pull(self.path, target_ref)
|
|
|
|
|
|
try:
|
|
|
log.debug('Pulling in source reference %s', source_ref)
|
|
|
source_repo._validate_pull_reference(source_ref)
|
|
|
shadow_repo._local_pull(source_repo.path, source_ref)
|
|
|
except CommitDoesNotExistError:
|
|
|
log.exception('Failure when doing local pull on hg shadow repo')
|
|
|
return MergeResponse(
|
|
|
False, False, None, MergeFailureReason.MISSING_SOURCE_REF,
|
|
|
metadata={'source_ref': source_ref})
|
|
|
|
|
|
merge_ref = None
|
|
|
merge_commit_id = None
|
|
|
close_commit_id = None
|
|
|
merge_failure_reason = MergeFailureReason.NONE
|
|
|
metadata = {}
|
|
|
|
|
|
# enforce that close branch should be used only in case we source from
|
|
|
# an actual Branch
|
|
|
close_branch = close_branch and source_ref.type == 'branch'
|
|
|
|
|
|
# don't allow to close branch if source and target are the same
|
|
|
close_branch = close_branch and source_ref.name != target_ref.name
|
|
|
|
|
|
needs_push_on_close = False
|
|
|
if close_branch and not use_rebase and not dry_run:
|
|
|
try:
|
|
|
close_commit_id, needs_push_on_close = shadow_repo._local_close(
|
|
|
target_ref, merger_name, merger_email, source_ref)
|
|
|
merge_possible = True
|
|
|
except RepositoryError:
|
|
|
log.exception('Failure when doing close branch on '
|
|
|
'shadow repo: %s', shadow_repo)
|
|
|
merge_possible = False
|
|
|
merge_failure_reason = MergeFailureReason.MERGE_FAILED
|
|
|
else:
|
|
|
merge_possible = True
|
|
|
|
|
|
needs_push = False
|
|
|
if merge_possible:
|
|
|
try:
|
|
|
merge_commit_id, needs_push = shadow_repo._local_merge(
|
|
|
target_ref, merge_message, merger_name, merger_email,
|
|
|
source_ref, use_rebase=use_rebase, dry_run=dry_run)
|
|
|
merge_possible = True
|
|
|
|
|
|
# read the state of the close action, if it
|
|
|
# maybe required a push
|
|
|
needs_push = needs_push or needs_push_on_close
|
|
|
|
|
|
# Set a bookmark pointing to the merge commit. This bookmark
|
|
|
# may be used to easily identify the last successful merge
|
|
|
# commit in the shadow repository.
|
|
|
shadow_repo.bookmark('pr-merge', revision=merge_commit_id)
|
|
|
merge_ref = Reference('book', 'pr-merge', merge_commit_id)
|
|
|
except SubrepoMergeError:
|
|
|
log.exception(
|
|
|
'Subrepo merge error during local merge on hg shadow repo.')
|
|
|
merge_possible = False
|
|
|
merge_failure_reason = MergeFailureReason.SUBREPO_MERGE_FAILED
|
|
|
needs_push = False
|
|
|
except RepositoryError as e:
|
|
|
log.exception('Failure when doing local merge on hg shadow repo')
|
|
|
if isinstance(e, UnresolvedFilesInRepo):
|
|
|
metadata['unresolved_files'] = '\n* conflict: ' + ('\n * conflict: '.join(e.args[0]))
|
|
|
|
|
|
merge_possible = False
|
|
|
merge_failure_reason = MergeFailureReason.MERGE_FAILED
|
|
|
needs_push = False
|
|
|
|
|
|
if merge_possible and not dry_run:
|
|
|
if needs_push:
|
|
|
# In case the target is a bookmark, update it, so after pushing
|
|
|
# the bookmarks is also updated in the target.
|
|
|
if target_ref.type == 'book':
|
|
|
shadow_repo.bookmark(
|
|
|
target_ref.name, revision=merge_commit_id)
|
|
|
try:
|
|
|
shadow_repo_with_hooks = self.get_shadow_instance(
|
|
|
shadow_repository_path,
|
|
|
enable_hooks=True)
|
|
|
# This is the actual merge action, we push from shadow
|
|
|
# into origin.
|
|
|
# Note: the push_branches option will push any new branch
|
|
|
# defined in the source repository to the target. This may
|
|
|
# be dangerous as branches are permanent in Mercurial.
|
|
|
# This feature was requested in issue #441.
|
|
|
shadow_repo_with_hooks._local_push(
|
|
|
merge_commit_id, self.path, push_branches=True,
|
|
|
enable_hooks=True)
|
|
|
|
|
|
# maybe we also need to push the close_commit_id
|
|
|
if close_commit_id:
|
|
|
shadow_repo_with_hooks._local_push(
|
|
|
close_commit_id, self.path, push_branches=True,
|
|
|
enable_hooks=True)
|
|
|
merge_succeeded = True
|
|
|
except RepositoryError:
|
|
|
log.exception(
|
|
|
'Failure when doing local push from the shadow '
|
|
|
'repository to the target repository at %s.', self.path)
|
|
|
merge_succeeded = False
|
|
|
merge_failure_reason = MergeFailureReason.PUSH_FAILED
|
|
|
metadata['target'] = 'hg shadow repo'
|
|
|
metadata['merge_commit'] = merge_commit_id
|
|
|
else:
|
|
|
merge_succeeded = True
|
|
|
else:
|
|
|
merge_succeeded = False
|
|
|
|
|
|
return MergeResponse(
|
|
|
merge_possible, merge_succeeded, merge_ref, merge_failure_reason,
|
|
|
metadata=metadata)
|
|
|
|
|
|
def get_shadow_instance(self, shadow_repository_path, enable_hooks=False, cache=False):
|
|
|
config = self.config.copy()
|
|
|
if not enable_hooks:
|
|
|
config.clear_section('hooks')
|
|
|
return MercurialRepository(shadow_repository_path, config, with_wire={"cache": cache})
|
|
|
|
|
|
def _validate_pull_reference(self, reference):
|
|
|
if not (reference.name in self.bookmarks or
|
|
|
reference.name in self.branches or
|
|
|
self.get_commit(reference.commit_id)):
|
|
|
raise CommitDoesNotExistError(
|
|
|
'Unknown branch, bookmark or commit id')
|
|
|
|
|
|
def _local_pull(self, repository_path, reference):
|
|
|
"""
|
|
|
Fetch a branch, bookmark or commit from a local repository.
|
|
|
"""
|
|
|
repository_path = os.path.abspath(repository_path)
|
|
|
if repository_path == self.path:
|
|
|
raise ValueError('Cannot pull from the same repository')
|
|
|
|
|
|
reference_type_to_option_name = {
|
|
|
'book': 'bookmark',
|
|
|
'branch': 'branch',
|
|
|
}
|
|
|
option_name = reference_type_to_option_name.get(
|
|
|
reference.type, 'revision')
|
|
|
|
|
|
if option_name == 'revision':
|
|
|
ref = reference.commit_id
|
|
|
else:
|
|
|
ref = reference.name
|
|
|
|
|
|
options = {option_name: [ref]}
|
|
|
self._remote.pull_cmd(repository_path, hooks=False, **options)
|
|
|
self._remote.invalidate_vcs_cache()
|
|
|
|
|
|
def bookmark(self, bookmark, revision=None):
|
|
|
if isinstance(bookmark, unicode):
|
|
|
bookmark = safe_str(bookmark)
|
|
|
self._remote.bookmark(bookmark, revision=revision)
|
|
|
self._remote.invalidate_vcs_cache()
|
|
|
|
|
|
def get_path_permissions(self, username):
|
|
|
hgacl_file = os.path.join(self.path, '.hg/hgacl')
|
|
|
|
|
|
def read_patterns(suffix):
|
|
|
svalue = None
|
|
|
for section, option in [
|
|
|
('narrowacl', username + suffix),
|
|
|
('narrowacl', 'default' + suffix),
|
|
|
('narrowhgacl', username + suffix),
|
|
|
('narrowhgacl', 'default' + suffix)
|
|
|
]:
|
|
|
try:
|
|
|
svalue = hgacl.get(section, option)
|
|
|
break # stop at the first value we find
|
|
|
except configparser.NoOptionError:
|
|
|
pass
|
|
|
if not svalue:
|
|
|
return None
|
|
|
result = ['/']
|
|
|
for pattern in svalue.split():
|
|
|
result.append(pattern)
|
|
|
if '*' not in pattern and '?' not in pattern:
|
|
|
result.append(pattern + '/*')
|
|
|
return result
|
|
|
|
|
|
if os.path.exists(hgacl_file):
|
|
|
try:
|
|
|
hgacl = configparser.RawConfigParser()
|
|
|
hgacl.read(hgacl_file)
|
|
|
|
|
|
includes = read_patterns('.includes')
|
|
|
excludes = read_patterns('.excludes')
|
|
|
return BasePathPermissionChecker.create_from_patterns(
|
|
|
includes, excludes)
|
|
|
except BaseException as e:
|
|
|
msg = 'Cannot read ACL settings from {} on {}: {}'.format(
|
|
|
hgacl_file, self.name, e)
|
|
|
raise exceptions.RepositoryRequirementError(msg)
|
|
|
else:
|
|
|
return None
|
|
|
|
|
|
|
|
|
class MercurialIndexBasedCollectionGenerator(CollectionGenerator):
|
|
|
|
|
|
def _commit_factory(self, commit_id):
|
|
|
return self.repo.get_commit(
|
|
|
commit_idx=commit_id, pre_load=self.pre_load)
|
|
|
|