base.py
1981 lines
| 62.7 KiB
| text/x-python
|
PythonLexer
r5608 | # Copyright (C) 2014-2024 RhodeCode GmbH | |||
r1 | # | |||
# This program is free software: you can redistribute it and/or modify | ||||
# it under the terms of the GNU Affero General Public License, version 3 | ||||
# (only), as published by the Free Software Foundation. | ||||
# | ||||
# This program is distributed in the hope that it will be useful, | ||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of | ||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||||
# GNU General Public License for more details. | ||||
# | ||||
# You should have received a copy of the GNU Affero General Public License | ||||
# along with this program. If not, see <http://www.gnu.org/licenses/>. | ||||
# | ||||
# This program is dual-licensed. If you wish to learn more about the | ||||
# RhodeCode Enterprise Edition, including its added features, Support services, | ||||
# and proprietary license terms, please see https://rhodecode.com/licenses/ | ||||
""" | ||||
Base module for all VCS systems | ||||
""" | ||||
r5647 | ||||
r3339 | import os | |||
import re | ||||
import time | ||||
import shutil | ||||
r1 | import datetime | |||
r2619 | import fnmatch | |||
r1 | import itertools | |||
import logging | ||||
r5074 | import dataclasses | |||
r1 | import warnings | |||
from zope.cachedescriptors.property import Lazy as LazyProperty | ||||
r3743 | ||||
r1 | ||||
r3842 | import rhodecode | |||
r3339 | from rhodecode.translation import lazy_ugettext | |||
r5065 | from rhodecode.lib.utils2 import safe_str, CachedProperty | |||
r1 | from rhodecode.lib.vcs.utils import author_name, author_email | |||
from rhodecode.lib.vcs.conf import settings | ||||
from rhodecode.lib.vcs.exceptions import ( | ||||
r5647 | CommitError, | |||
EmptyRepositoryError, | ||||
NodeAlreadyAddedError, | ||||
NodeAlreadyChangedError, | ||||
NodeAlreadyExistsError, | ||||
NodeAlreadyRemovedError, | ||||
NodeDoesNotExistError, | ||||
NodeNotChangedError, | ||||
VCSError, | ||||
ImproperArchiveTypeError, | ||||
BranchDoesNotExistError, | ||||
CommitDoesNotExistError, | ||||
RepositoryError, | ||||
) | ||||
r1 | ||||
log = logging.getLogger(__name__) | ||||
r5647 | EMPTY_COMMIT_ID = "0" * 40 | |||
r1 | ||||
r4509 | ||||
r5074 | @dataclasses.dataclass | |||
class Reference: | ||||
type: str | ||||
name: str | ||||
commit_id: str | ||||
r4509 | ||||
r5074 | def __iter__(self): | |||
yield self.type | ||||
yield self.name | ||||
yield self.commit_id | ||||
r4509 | ||||
@property | ||||
def branch(self): | ||||
r5647 | if self.type == "branch": | |||
r4509 | return self.name | |||
@property | ||||
def bookmark(self): | ||||
r5647 | if self.type == "book": | |||
r4509 | return self.name | |||
r1 | ||||
r4653 | @property | |||
r5065 | def to_str(self): | |||
r4653 | return reference_to_unicode(self) | |||
r5074 | def asdict(self): | |||
r5647 | return dict(type=self.type, name=self.name, commit_id=self.commit_id) | |||
r1 | ||||
r5074 | ||||
def unicode_to_reference(raw: str): | ||||
r4519 | """ | |||
Convert a unicode (or string) to a reference object. | ||||
If unicode evaluates to False it returns None. | ||||
""" | ||||
if raw: | ||||
r5647 | refs = raw.split(":") | |||
r4519 | return Reference(*refs) | |||
else: | ||||
return None | ||||
r5065 | def reference_to_unicode(ref: Reference): | |||
r4519 | """ | |||
Convert a reference object to unicode. | ||||
If reference is None it returns None. | ||||
""" | ||||
if ref: | ||||
r5647 | return ":".join(ref) | |||
r4519 | else: | |||
return None | ||||
r1 | class MergeFailureReason(object): | |||
""" | ||||
Enumeration with all the reasons why the server side merge could fail. | ||||
DO NOT change the number of the reasons, as they may be stored in the | ||||
database. | ||||
Changing the name of a reason is acceptable and encouraged to deprecate old | ||||
reasons. | ||||
""" | ||||
# Everything went well. | ||||
NONE = 0 | ||||
# An unexpected exception was raised. Check the logs for more details. | ||||
UNKNOWN = 1 | ||||
# The merge was not successful, there are conflicts. | ||||
MERGE_FAILED = 2 | ||||
# The merge succeeded but we could not push it to the target repository. | ||||
PUSH_FAILED = 3 | ||||
# The specified target is not a head in the target repository. | ||||
TARGET_IS_NOT_HEAD = 4 | ||||
# The source repository contains more branches than the target. Pushing | ||||
# the merge will create additional branches in the target. | ||||
HG_SOURCE_HAS_MORE_BRANCHES = 5 | ||||
# The target reference has multiple heads. That does not allow to correctly | ||||
# identify the target location. This could only happen for mercurial | ||||
# branches. | ||||
HG_TARGET_HAS_MULTIPLE_HEADS = 6 | ||||
# The target repository is locked | ||||
TARGET_IS_LOCKED = 7 | ||||
Martin Bornhold
|
r1081 | # Deprecated, use MISSING_TARGET_REF or MISSING_SOURCE_REF instead. | ||
r1 | # A involved commit could not be found. | |||
Martin Bornhold
|
r1081 | _DEPRECATED_MISSING_COMMIT = 8 | ||
r1 | ||||
Martin Bornhold
|
r1067 | # The target repo reference is missing. | ||
MISSING_TARGET_REF = 9 | ||||
# The source repo reference is missing. | ||||
MISSING_SOURCE_REF = 10 | ||||
Martin Bornhold
|
r1106 | # The merge was not successful, there are conflicts related to sub | ||
# repositories. | ||||
SUBREPO_MERGE_FAILED = 11 | ||||
r1 | ||||
Martin Bornhold
|
r1069 | class UpdateFailureReason(object): | ||
""" | ||||
Enumeration with all the reasons why the pull request update could fail. | ||||
DO NOT change the number of the reasons, as they may be stored in the | ||||
database. | ||||
Changing the name of a reason is acceptable and encouraged to deprecate old | ||||
reasons. | ||||
""" | ||||
# Everything went well. | ||||
NONE = 0 | ||||
# An unexpected exception was raised. Check the logs for more details. | ||||
UNKNOWN = 1 | ||||
# The pull request is up to date. | ||||
NO_CHANGE = 2 | ||||
# The pull request has a reference type that is not supported for update. | ||||
r1687 | WRONG_REF_TYPE = 3 | |||
Martin Bornhold
|
r1069 | |||
# Update failed because the target reference is missing. | ||||
MISSING_TARGET_REF = 4 | ||||
# Update failed because the source reference is missing. | ||||
MISSING_SOURCE_REF = 5 | ||||
r3339 | class MergeResponse(object): | |||
# uses .format(**metadata) for variables | ||||
MERGE_STATUS_MESSAGES = { | ||||
r5647 | MergeFailureReason.NONE: lazy_ugettext("This pull request can be automatically merged."), | |||
r3339 | MergeFailureReason.UNKNOWN: lazy_ugettext( | |||
r5647 | "This pull request cannot be merged because of an unhandled exception. " "{exception}" | |||
), | ||||
r3339 | MergeFailureReason.MERGE_FAILED: lazy_ugettext( | |||
r5647 | "This pull request cannot be merged because of merge conflicts. {unresolved_files}" | |||
), | ||||
r3339 | MergeFailureReason.PUSH_FAILED: lazy_ugettext( | |||
r5647 | "This pull request could not be merged because push to " "target:`{target}@{merge_commit}` failed." | |||
), | ||||
r3339 | MergeFailureReason.TARGET_IS_NOT_HEAD: lazy_ugettext( | |||
r5647 | "This pull request cannot be merged because the target " "`{target_ref.name}` is not a head." | |||
), | ||||
r3339 | MergeFailureReason.HG_SOURCE_HAS_MORE_BRANCHES: lazy_ugettext( | |||
r5647 | "This pull request cannot be merged because the source contains " "more branches than the target." | |||
), | ||||
r3339 | MergeFailureReason.HG_TARGET_HAS_MULTIPLE_HEADS: lazy_ugettext( | |||
r5647 | "This pull request cannot be merged because the target `{target_ref.name}` " | |||
"has multiple heads: `{heads}`." | ||||
), | ||||
r3339 | MergeFailureReason.TARGET_IS_LOCKED: lazy_ugettext( | |||
r5647 | "This pull request cannot be merged because the target repository is " "locked by {locked_by}." | |||
), | ||||
r3339 | MergeFailureReason.MISSING_TARGET_REF: lazy_ugettext( | |||
r5647 | "This pull request cannot be merged because the target " "reference `{target_ref.name}` is missing." | |||
), | ||||
r3339 | MergeFailureReason.MISSING_SOURCE_REF: lazy_ugettext( | |||
r5647 | "This pull request cannot be merged because the source " "reference `{source_ref.name}` is missing." | |||
), | ||||
r3339 | MergeFailureReason.SUBREPO_MERGE_FAILED: lazy_ugettext( | |||
r5647 | "This pull request cannot be merged because of conflicts related " "to sub repositories." | |||
), | ||||
r3339 | # Deprecations | |||
MergeFailureReason._DEPRECATED_MISSING_COMMIT: lazy_ugettext( | ||||
r5647 | "This pull request cannot be merged because the target or the " "source reference is missing." | |||
), | ||||
r3339 | } | |||
r5178 | def __init__(self, possible, executed, merge_ref: Reference, failure_reason, metadata=None): | |||
r3339 | self.possible = possible | |||
self.executed = executed | ||||
self.merge_ref = merge_ref | ||||
self.failure_reason = failure_reason | ||||
self.metadata = metadata or {} | ||||
def __repr__(self): | ||||
r5647 | return f"<MergeResponse:{self.label} {self.failure_reason}>" | |||
r3339 | ||||
def __eq__(self, other): | ||||
same_instance = isinstance(other, self.__class__) | ||||
r5647 | return ( | |||
same_instance | ||||
and self.possible == other.possible | ||||
and self.executed == other.executed | ||||
and self.failure_reason == other.failure_reason | ||||
) | ||||
r3339 | ||||
@property | ||||
def label(self): | ||||
r5647 | label_dict = dict((v, k) for k, v in MergeFailureReason.__dict__.items() if not k.startswith("_")) | |||
r3339 | return label_dict.get(self.failure_reason) | |||
@property | ||||
def merge_status_message(self): | ||||
""" | ||||
Return a human friendly error message for the given merge status code. | ||||
""" | ||||
r5065 | msg = safe_str(self.MERGE_STATUS_MESSAGES[self.failure_reason]) | |||
r4299 | ||||
r3339 | try: | |||
return msg.format(**self.metadata) | ||||
except Exception: | ||||
r5647 | log.exception("Failed to format %s message", self) | |||
r3339 | return msg | |||
def asdict(self): | ||||
data = {} | ||||
r5647 | for k in ["possible", "executed", "merge_ref", "failure_reason", "merge_status_message"]: | |||
r3339 | data[k] = getattr(self, k) | |||
return data | ||||
r4317 | class TargetRefMissing(ValueError): | |||
pass | ||||
class SourceRefMissing(ValueError): | ||||
pass | ||||
r5647 | class BaseCommit: | |||
r1 | """ | |||
Each backend should implement it's commit representation. | ||||
**Attributes** | ||||
``repository`` | ||||
repository object within which commit exists | ||||
``id`` | ||||
The commit id, may be ``raw_id`` or i.e. for mercurial's tip | ||||
just ``tip``. | ||||
``raw_id`` | ||||
raw commit representation (i.e. full 40 length sha for git | ||||
backend) | ||||
``short_id`` | ||||
shortened (if apply) version of ``raw_id``; it would be simple | ||||
shortcut for ``raw_id[:12]`` for git/mercurial backends or same | ||||
as ``raw_id`` for subversion | ||||
``idx`` | ||||
commit index | ||||
``files`` | ||||
list of ``FileNode`` (``Node`` with NodeKind.FILE) objects | ||||
``dirs`` | ||||
list of ``DirNode`` (``Node`` with NodeKind.DIR) objects | ||||
``nodes`` | ||||
combined list of ``Node`` objects | ||||
``author`` | ||||
author of the commit, as unicode | ||||
``message`` | ||||
message of the commit, as unicode | ||||
``parents`` | ||||
list of parent commits | ||||
""" | ||||
r5647 | ||||
r4536 | repository = None | |||
branch = None | ||||
r1 | ||||
""" | ||||
Depending on the backend this should be set to the branch name of the | ||||
commit. Backends not supporting branches on commits should leave this | ||||
value as ``None``. | ||||
""" | ||||
r5647 | _ARCHIVE_PREFIX_TEMPLATE = "{repo_name}-{short_id}" | |||
r1 | """ | |||
This template is used to generate a default prefix for repository archives | ||||
if no prefix has been specified. | ||||
""" | ||||
def __repr__(self): | ||||
return self.__str__() | ||||
r5065 | def __str__(self): | |||
r5647 | return f"<{self.__class__.__name__} at {self.idx}:{self.short_id}>" | |||
r1 | ||||
def __eq__(self, other): | ||||
r985 | same_instance = isinstance(other, self.__class__) | |||
return same_instance and self.raw_id == other.raw_id | ||||
r1 | ||||
def __json__(self): | ||||
parents = [] | ||||
try: | ||||
for parent in self.parents: | ||||
r5647 | parents.append({"raw_id": parent.raw_id}) | |||
r1 | except NotImplementedError: | |||
# empty commit doesn't have parents implemented | ||||
pass | ||||
return { | ||||
r5647 | "short_id": self.short_id, | |||
"raw_id": self.raw_id, | ||||
"revision": self.idx, | ||||
"message": self.message, | ||||
"date": self.date, | ||||
"author": self.author, | ||||
"parents": parents, | ||||
"branch": self.branch, | ||||
r1 | } | |||
Bartłomiej Wołyńczyk
|
r2685 | def __getstate__(self): | ||
d = self.__dict__.copy() | ||||
r5647 | d.pop("_remote", None) | |||
d.pop("repository", None) | ||||
Bartłomiej Wołyńczyk
|
r2685 | return d | ||
r5074 | def get_remote(self): | |||
return self._remote | ||||
r4305 | def serialize(self): | |||
return self.__json__() | ||||
r2337 | def _get_refs(self): | |||
return { | ||||
r5647 | "branches": [self.branch] if self.branch else [], | |||
"bookmarks": getattr(self, "bookmarks", []), | ||||
"tags": self.tags, | ||||
r2337 | } | |||
r1 | @LazyProperty | |||
def last(self): | ||||
""" | ||||
``True`` if this is last commit in repository, ``False`` | ||||
otherwise; trying to access this attribute while there is no | ||||
commits would raise `EmptyRepositoryError` | ||||
""" | ||||
if self.repository is None: | ||||
raise CommitError("Cannot check if it's most recent commit") | ||||
return self.raw_id == self.repository.commit_ids[-1] | ||||
@LazyProperty | ||||
def parents(self): | ||||
""" | ||||
Returns list of parent commits. | ||||
""" | ||||
raise NotImplementedError | ||||
r3124 | @LazyProperty | |||
def first_parent(self): | ||||
""" | ||||
Returns list of parent commits. | ||||
""" | ||||
return self.parents[0] if self.parents else EmptyCommit() | ||||
r1 | @property | |||
def merge(self): | ||||
""" | ||||
Returns boolean if commit is a merge. | ||||
""" | ||||
return len(self.parents) > 1 | ||||
@LazyProperty | ||||
def children(self): | ||||
""" | ||||
Returns list of child commits. | ||||
""" | ||||
raise NotImplementedError | ||||
@LazyProperty | ||||
def id(self): | ||||
""" | ||||
Returns string identifying this commit. | ||||
""" | ||||
raise NotImplementedError | ||||
@LazyProperty | ||||
def raw_id(self): | ||||
""" | ||||
Returns raw string identifying this commit. | ||||
""" | ||||
raise NotImplementedError | ||||
@LazyProperty | ||||
def short_id(self): | ||||
""" | ||||
Returns shortened version of ``raw_id`` attribute, as string, | ||||
identifying this commit, useful for presentation to users. | ||||
""" | ||||
raise NotImplementedError | ||||
@LazyProperty | ||||
def idx(self): | ||||
""" | ||||
Returns integer identifying this commit. | ||||
""" | ||||
raise NotImplementedError | ||||
@LazyProperty | ||||
def committer(self): | ||||
""" | ||||
Returns committer for this commit | ||||
""" | ||||
raise NotImplementedError | ||||
@LazyProperty | ||||
def committer_name(self): | ||||
""" | ||||
Returns committer name for this commit | ||||
""" | ||||
return author_name(self.committer) | ||||
@LazyProperty | ||||
def committer_email(self): | ||||
""" | ||||
Returns committer email address for this commit | ||||
""" | ||||
return author_email(self.committer) | ||||
@LazyProperty | ||||
def author(self): | ||||
""" | ||||
Returns author for this commit | ||||
""" | ||||
raise NotImplementedError | ||||
@LazyProperty | ||||
def author_name(self): | ||||
""" | ||||
Returns author name for this commit | ||||
""" | ||||
return author_name(self.author) | ||||
@LazyProperty | ||||
def author_email(self): | ||||
""" | ||||
Returns author email address for this commit | ||||
""" | ||||
return author_email(self.author) | ||||
r5074 | def get_file_mode(self, path: bytes): | |||
r1 | """ | |||
Returns stat mode of the file at `path`. | ||||
""" | ||||
raise NotImplementedError | ||||
r5647 | def is_link(self, path: bytes): | |||
r1 | """ | |||
Returns ``True`` if given `path` is a symlink | ||||
""" | ||||
raise NotImplementedError | ||||
r3896 | def is_node_binary(self, path): | |||
""" | ||||
Returns ``True`` is given path is a binary file | ||||
""" | ||||
raise NotImplementedError | ||||
r5074 | def node_md5_hash(self, path): | |||
""" | ||||
Returns md5 hash of a node data | ||||
""" | ||||
raise NotImplementedError | ||||
def get_file_content(self, path) -> bytes: | ||||
r1 | """ | |||
Returns content of the file at the given `path`. | ||||
""" | ||||
raise NotImplementedError | ||||
r3895 | def get_file_content_streamed(self, path): | |||
""" | ||||
returns a streaming response from vcsserver with file content | ||||
""" | ||||
raise NotImplementedError | ||||
r1 | def get_file_size(self, path): | |||
""" | ||||
Returns size of the file at the given `path`. | ||||
""" | ||||
raise NotImplementedError | ||||
r5647 | def get_path_commit(self, path: bytes, pre_load=None): | |||
r1 | """ | |||
Returns last commit of the file at the given `path`. | ||||
""" | ||||
r3275 | commits = self.get_path_history(path, limit=1, pre_load=pre_load) | |||
r1179 | if not commits: | |||
raise RepositoryError( | ||||
r5647 | f"Failed to fetch history for path {path}. Please check if such path exists in your repository" | |||
) | ||||
r1179 | return commits[0] | |||
r1 | ||||
r5647 | def get_path_history(self, path: bytes, limit=None, pre_load=None): | |||
r1 | """ | |||
Returns history of file as reversed list of :class:`BaseCommit` | ||||
objects for which file at given `path` has been modified. | ||||
r5647 | :param path: file path or dir path | |||
r1 | :param limit: Optional. Allows to limit the size of the returned | |||
history. This is intended as a hint to the underlying backend, so | ||||
that it can apply optimizations depending on the limit. | ||||
:param pre_load: Optional. List of commit attributes to load. | ||||
""" | ||||
raise NotImplementedError | ||||
r5647 | def get_file_annotate(self, path: bytes, pre_load=None): | |||
r1 | """ | |||
Returns a generator of four element tuples with | ||||
lineno, sha, commit lazy loader and line | ||||
r5647 | :param path: file path | |||
r1 | :param pre_load: Optional. List of commit attributes to load. | |||
""" | ||||
raise NotImplementedError | ||||
r5647 | def get_nodes(self, path: bytes, pre_load=None): | |||
r1 | """ | |||
Returns combined ``DirNode`` and ``FileNode`` objects list representing | ||||
state of commit at the given ``path``. | ||||
:raises ``CommitError``: if node at the given ``path`` is not | ||||
instance of ``DirNode`` | ||||
""" | ||||
raise NotImplementedError | ||||
r5647 | def get_node(self, path: bytes, pre_load=None): | |||
r1 | """ | |||
Returns ``Node`` object from the given ``path``. | ||||
:raises ``NodeDoesNotExistError``: if there is no node at the given | ||||
``path`` | ||||
""" | ||||
raise NotImplementedError | ||||
r5647 | def get_largefile_node(self, path: bytes): | |||
r1 | """ | |||
r1577 | Returns the path to largefile from Mercurial/Git-lfs storage. | |||
or None if it's not a largefile node | ||||
r1 | """ | |||
r1577 | return None | |||
r1 | ||||
r5647 | def archive_repo( | |||
self, | ||||
archive_name_key, | ||||
kind="tgz", | ||||
subrepos=None, | ||||
archive_dir_name=None, | ||||
write_metadata=False, | ||||
mtime=None, | ||||
archive_at_path="/", | ||||
cache_config=None, | ||||
): | ||||
r1 | """ | |||
Creates an archive containing the contents of the repository. | ||||
r5074 | :param archive_name_key: unique key under this archive should be generated | |||
r5135 | :param kind: one of the following: ``"tbz2"``, ``"tgz"``, ``"zip"``. | |||
r4534 | :param archive_dir_name: name of root directory in archive. | |||
r1 | Default is repository name and commit's short_id joined with dash: | |||
``"{repo_name}-{short_id}"``. | ||||
:param write_metadata: write a metadata file into archive. | ||||
:param mtime: custom modification time for archive creation, defaults | ||||
to time.time() if not given. | ||||
r3709 | :param archive_at_path: pack files at this path (default '/') | |||
r5074 | :param cache_config: config spec to send to vcsserver to configure the backend to store files | |||
r1 | ||||
:raise VCSError: If prefix has a problem. | ||||
""" | ||||
r5074 | cache_config = cache_config or {} | |||
r3736 | allowed_kinds = [x[0] for x in settings.ARCHIVE_SPECS] | |||
r1 | if kind not in allowed_kinds: | |||
r5647 | raise ImproperArchiveTypeError(f"Archive kind ({kind}) not supported use one of {allowed_kinds}") | |||
r1 | ||||
r4534 | archive_dir_name = self._validate_archive_prefix(archive_dir_name) | |||
r3727 | mtime = mtime is not None or time.mktime(self.date.timetuple()) | |||
r4536 | commit_id = self.raw_id | |||
r1 | ||||
r4536 | return self.repository._remote.archive_repo( | |||
r5647 | archive_name_key, kind, mtime, archive_at_path, archive_dir_name, commit_id, cache_config | |||
) | ||||
r1 | ||||
r4534 | def _validate_archive_prefix(self, archive_dir_name): | |||
if archive_dir_name is None: | ||||
archive_dir_name = self._ARCHIVE_PREFIX_TEMPLATE.format( | ||||
r5647 | repo_name=safe_str(self.repository.name), short_id=self.short_id | |||
) | ||||
r4534 | elif not isinstance(archive_dir_name, str): | |||
r5074 | raise ValueError(f"archive_dir_name is not str object but: {type(archive_dir_name)}") | |||
r5647 | elif archive_dir_name.startswith("/"): | |||
r1 | raise VCSError("Prefix cannot start with leading slash") | |||
r5647 | elif archive_dir_name.strip() == "": | |||
r1 | raise VCSError("Prefix cannot be empty") | |||
r5074 | elif not archive_dir_name.isascii(): | |||
raise VCSError("Prefix cannot contain non ascii characters") | ||||
r4534 | return archive_dir_name | |||
r1 | ||||
@LazyProperty | ||||
def root(self): | ||||
""" | ||||
Returns ``RootNode`` object for this commit. | ||||
""" | ||||
r5647 | return self.get_node(b"") | |||
r1 | ||||
def next(self, branch=None): | ||||
""" | ||||
Returns next commit from current, if branch is gives it will return | ||||
next commit belonging to this branch | ||||
:param branch: show commits within the given named branch | ||||
""" | ||||
r4906 | indexes = range(self.idx + 1, self.repository.count()) | |||
r1 | return self._find_next(indexes, branch) | |||
def prev(self, branch=None): | ||||
""" | ||||
Returns previous commit from current, if branch is gives it will | ||||
return previous commit belonging to this branch | ||||
:param branch: show commit within the given named branch | ||||
""" | ||||
r4906 | indexes = range(self.idx - 1, -1, -1) | |||
r1 | return self._find_next(indexes, branch) | |||
def _find_next(self, indexes, branch=None): | ||||
if branch and self.branch != branch: | ||||
r5647 | raise VCSError("Branch option used on commit not belonging " "to that branch") | |||
r1 | ||||
for next_idx in indexes: | ||||
commit = self.repository.get_commit(commit_idx=next_idx) | ||||
if branch and branch != commit.branch: | ||||
continue | ||||
return commit | ||||
raise CommitDoesNotExistError | ||||
def diff(self, ignore_whitespace=True, context=3): | ||||
""" | ||||
Returns a `Diff` object representing the change made by this commit. | ||||
""" | ||||
r3124 | parent = self.first_parent | |||
r5647 | diff = self.repository.get_diff(parent, self, ignore_whitespace=ignore_whitespace, context=context) | |||
r1 | return diff | |||
@LazyProperty | ||||
def size(self): | ||||
""" | ||||
Returns total number of bytes from contents of all filenodes. | ||||
""" | ||||
r5091 | return sum(node.size for node in self.get_filenodes_generator()) | |||
r1 | ||||
r5647 | def walk(self, top_url=b""): | |||
r1 | """ | |||
Similar to os.walk method. Insted of filesystem it walks through | ||||
commit starting at given ``topurl``. Returns generator of tuples | ||||
r5074 | (top_node, dirnodes, filenodes). | |||
r1 | """ | |||
r5074 | from rhodecode.lib.vcs.nodes import DirNode | |||
r5647 | if isinstance(top_url, DirNode): | |||
top_node = top_url | ||||
r5074 | else: | |||
r5647 | top_node = self.get_node(top_url) | |||
r5074 | ||||
has_default_pre_load = False | ||||
if isinstance(top_node, DirNode): | ||||
# used to inject as we walk same defaults as given top_node | ||||
default_pre_load = top_node.default_pre_load | ||||
has_default_pre_load = True | ||||
if not top_node.is_dir(): | ||||
r1 | return | |||
r5074 | yield top_node, top_node.dirs, top_node.files | |||
for dir_node in top_node.dirs: | ||||
if has_default_pre_load: | ||||
dir_node.default_pre_load = default_pre_load | ||||
r5091 | yield from self.walk(dir_node) | |||
r1 | ||||
def get_filenodes_generator(self): | ||||
""" | ||||
Returns generator that yields *all* file nodes. | ||||
""" | ||||
for topnode, dirs, files in self.walk(): | ||||
r5091 | yield from files | |||
r1 | ||||
# | ||||
# Utilities for sub classes to support consistent behavior | ||||
# | ||||
def no_node_at_path(self, path): | ||||
return NodeDoesNotExistError( | ||||
r5647 | f"There is no file nor directory at the given path: " f"`{safe_str(path)}` at commit {self.short_id}" | |||
) | ||||
r1 | ||||
r5647 | @classmethod | |||
def _fix_path(cls, path: bytes) -> bytes: | ||||
r1 | """ | |||
r5647 | Paths are stored without trailing slash so we need to get rid off it if needed. | |||
It also validates that path is a bytestring for support of mixed encodings... | ||||
r1 | """ | |||
r5647 | ||||
if not isinstance(path, bytes): | ||||
raise ValueError(f'path=`{safe_str(path)}` must be bytes') | ||||
return path.rstrip(b"/") | ||||
r1 | ||||
# | ||||
# Deprecated API based on changesets | ||||
# | ||||
@property | ||||
def revision(self): | ||||
warnings.warn("Use idx instead", DeprecationWarning) | ||||
return self.idx | ||||
@revision.setter | ||||
def revision(self, value): | ||||
warnings.warn("Use idx instead", DeprecationWarning) | ||||
self.idx = value | ||||
def get_file_changeset(self, path): | ||||
r3275 | warnings.warn("Use get_path_commit instead", DeprecationWarning) | |||
return self.get_path_commit(path) | ||||
r1 | ||||
r5647 | class BaseRepository(object): | |||
""" | ||||
Base Repository for final backends | ||||
.. attribute:: DEFAULT_BRANCH_NAME | ||||
name of default branch (i.e. "trunk" for svn, "master" for git etc. | ||||
.. attribute:: commit_ids | ||||
list of all available commit ids, in ascending order | ||||
.. attribute:: path | ||||
absolute path to the repository | ||||
.. attribute:: bookmarks | ||||
Mapping from name to :term:`Commit ID` of the bookmark. Empty in case | ||||
there are no bookmarks or the backend implementation does not support | ||||
bookmarks. | ||||
.. attribute:: tags | ||||
Mapping from name to :term:`Commit ID` of the tag. | ||||
""" | ||||
DEFAULT_BRANCH_NAME = None | ||||
DEFAULT_CONTACT = "Unknown" | ||||
DEFAULT_DESCRIPTION = "unknown" | ||||
EMPTY_COMMIT_ID = "0" * 40 | ||||
COMMIT_ID_PAT = re.compile(r"[0-9a-fA-F]{40}") | ||||
path = None | ||||
_is_empty = None | ||||
_commit_ids = {} | ||||
def __init__(self, repo_path, config=None, create=False, **kwargs): | ||||
""" | ||||
Initializes repository. Raises RepositoryError if repository could | ||||
not be find at the given ``repo_path`` or directory at ``repo_path`` | ||||
exists and ``create`` is set to True. | ||||
:param repo_path: local path of the repository | ||||
:param config: repository configuration | ||||
:param create=False: if set to True, would try to create repository. | ||||
:param src_url=None: if set, should be proper url from which repository | ||||
would be cloned; requires ``create`` parameter to be set to True - | ||||
raises RepositoryError if src_url is set and create evaluates to | ||||
False | ||||
""" | ||||
raise NotImplementedError | ||||
def __repr__(self): | ||||
return f"<{self.__class__.__name__} at {self.path}>" | ||||
def __len__(self): | ||||
return self.count() | ||||
def __eq__(self, other): | ||||
same_instance = isinstance(other, self.__class__) | ||||
return same_instance and other.path == self.path | ||||
def __ne__(self, other): | ||||
return not self.__eq__(other) | ||||
@classmethod | ||||
def get_create_shadow_cache_pr_path(cls, db_repo): | ||||
path = db_repo.cached_diffs_dir | ||||
if not os.path.exists(path): | ||||
os.makedirs(path, 0o755) | ||||
return path | ||||
@classmethod | ||||
def get_default_config(cls, default=None): | ||||
config = Config() | ||||
if default and isinstance(default, list): | ||||
for section, key, val in default: | ||||
config.set(section, key, val) | ||||
return config | ||||
@LazyProperty | ||||
def _remote(self): | ||||
raise NotImplementedError | ||||
def _heads(self, branch=None): | ||||
return [] | ||||
@LazyProperty | ||||
def EMPTY_COMMIT(self): | ||||
return EmptyCommit(self.EMPTY_COMMIT_ID) | ||||
@LazyProperty | ||||
def alias(self): | ||||
for k, v in settings.BACKENDS.items(): | ||||
if v.split(".")[-1] == str(self.__class__.__name__): | ||||
return k | ||||
@LazyProperty | ||||
def name(self): | ||||
return safe_str(os.path.basename(self.path)) | ||||
@LazyProperty | ||||
def description(self): | ||||
raise NotImplementedError | ||||
def refs(self): | ||||
""" | ||||
returns a `dict` with branches, bookmarks, tags, and closed_branches | ||||
for this repository | ||||
""" | ||||
return dict( | ||||
branches=self.branches, branches_closed=self.branches_closed, tags=self.tags, bookmarks=self.bookmarks | ||||
) | ||||
@LazyProperty | ||||
def branches(self): | ||||
""" | ||||
A `dict` which maps branch names to commit ids. | ||||
""" | ||||
raise NotImplementedError | ||||
@LazyProperty | ||||
def branches_closed(self): | ||||
""" | ||||
A `dict` which maps tags names to commit ids. | ||||
""" | ||||
raise NotImplementedError | ||||
@LazyProperty | ||||
def bookmarks(self): | ||||
""" | ||||
A `dict` which maps tags names to commit ids. | ||||
""" | ||||
raise NotImplementedError | ||||
@LazyProperty | ||||
def tags(self): | ||||
""" | ||||
A `dict` which maps tags names to commit ids. | ||||
""" | ||||
raise NotImplementedError | ||||
@LazyProperty | ||||
def size(self): | ||||
""" | ||||
Returns combined size in bytes for all repository files | ||||
""" | ||||
tip = self.get_commit() | ||||
return tip.size | ||||
def size_at_commit(self, commit_id): | ||||
commit = self.get_commit(commit_id) | ||||
return commit.size | ||||
def _check_for_empty(self): | ||||
no_commits = len(self._commit_ids) == 0 | ||||
if no_commits: | ||||
# check on remote to be sure | ||||
return self._remote.is_empty() | ||||
else: | ||||
return False | ||||
def is_empty(self): | ||||
if rhodecode.is_test: | ||||
return self._check_for_empty() | ||||
if self._is_empty is None: | ||||
# cache empty for production, but not tests | ||||
self._is_empty = self._check_for_empty() | ||||
return self._is_empty | ||||
@staticmethod | ||||
def check_url(url, config): | ||||
""" | ||||
Function will check given url and try to verify if it's a valid | ||||
link. | ||||
""" | ||||
raise NotImplementedError | ||||
@staticmethod | ||||
def is_valid_repository(path): | ||||
""" | ||||
Check if given `path` contains a valid repository of this backend | ||||
""" | ||||
raise NotImplementedError | ||||
# ========================================================================== | ||||
# COMMITS | ||||
# ========================================================================== | ||||
@CachedProperty | ||||
def commit_ids(self): | ||||
raise NotImplementedError | ||||
def append_commit_id(self, commit_id): | ||||
if commit_id not in self.commit_ids: | ||||
self._rebuild_cache(self.commit_ids + [commit_id]) | ||||
# clear cache | ||||
self._invalidate_prop_cache("commit_ids") | ||||
self._is_empty = False | ||||
def get_commit( | ||||
self, | ||||
commit_id=None, | ||||
commit_idx=None, | ||||
pre_load=None, | ||||
translate_tag=None, | ||||
maybe_unreachable=False, | ||||
reference_obj=None, | ||||
) -> BaseCommit: | ||||
""" | ||||
Returns instance of `BaseCommit` class. If `commit_id` and `commit_idx` | ||||
are both None, most recent commit is returned. | ||||
:param pre_load: Optional. List of commit attributes to load. | ||||
:raises ``EmptyRepositoryError``: if there are no commits | ||||
""" | ||||
raise NotImplementedError | ||||
def __iter__(self): | ||||
for commit_id in self.commit_ids: | ||||
yield self.get_commit(commit_id=commit_id) | ||||
def get_commits( | ||||
self, | ||||
start_id=None, | ||||
end_id=None, | ||||
start_date=None, | ||||
end_date=None, | ||||
branch_name=None, | ||||
show_hidden=False, | ||||
pre_load=None, | ||||
translate_tags=None, | ||||
): | ||||
""" | ||||
Returns iterator of `BaseCommit` objects from start to end | ||||
not inclusive. This should behave just like a list, ie. end is not | ||||
inclusive. | ||||
:param start_id: None or str, must be a valid commit id | ||||
:param end_id: None or str, must be a valid commit id | ||||
:param start_date: | ||||
:param end_date: | ||||
:param branch_name: | ||||
:param show_hidden: | ||||
:param pre_load: | ||||
:param translate_tags: | ||||
""" | ||||
raise NotImplementedError | ||||
def __getitem__(self, key): | ||||
""" | ||||
Allows index based access to the commit objects of this repository. | ||||
""" | ||||
pre_load = ["author", "branch", "date", "message", "parents"] | ||||
if isinstance(key, slice): | ||||
return self._get_range(key, pre_load) | ||||
return self.get_commit(commit_idx=key, pre_load=pre_load) | ||||
def _get_range(self, slice_obj, pre_load): | ||||
for commit_id in self.commit_ids.__getitem__(slice_obj): | ||||
yield self.get_commit(commit_id=commit_id, pre_load=pre_load) | ||||
def count(self): | ||||
return len(self.commit_ids) | ||||
def tag(self, name, user, commit_id=None, message=None, date=None, **opts): | ||||
""" | ||||
Creates and returns a tag for the given ``commit_id``. | ||||
:param name: name for new tag | ||||
:param user: full username, i.e.: "Joe Doe <joe.doe@example.com>" | ||||
:param commit_id: commit id for which new tag would be created | ||||
:param message: message of the tag's commit | ||||
:param date: date of tag's commit | ||||
:raises TagAlreadyExistError: if tag with same name already exists | ||||
""" | ||||
raise NotImplementedError | ||||
def remove_tag(self, name, user, message=None, date=None): | ||||
""" | ||||
Removes tag with the given ``name``. | ||||
:param name: name of the tag to be removed | ||||
:param user: full username, i.e.: "Joe Doe <joe.doe@example.com>" | ||||
:param message: message of the tag's removal commit | ||||
:param date: date of tag's removal commit | ||||
:raises TagDoesNotExistError: if tag with given name does not exists | ||||
""" | ||||
raise NotImplementedError | ||||
def get_diff(self, commit1, commit2, path=None, ignore_whitespace=False, context=3, path1=None): | ||||
""" | ||||
Returns (git like) *diff*, as plain text. Shows changes introduced by | ||||
`commit2` since `commit1`. | ||||
:param commit1: Entry point from which diff is shown. Can be | ||||
``self.EMPTY_COMMIT`` - in this case, patch showing all | ||||
the changes since empty state of the repository until `commit2` | ||||
:param commit2: Until which commit changes should be shown. | ||||
:param path: Can be set to a path of a file to create a diff of that | ||||
file. If `path1` is also set, this value is only associated to | ||||
`commit2`. | ||||
:param ignore_whitespace: If set to ``True``, would not show whitespace | ||||
changes. Defaults to ``False``. | ||||
:param context: How many lines before/after changed lines should be | ||||
shown. Defaults to ``3``. | ||||
:param path1: Can be set to a path to associate with `commit1`. This | ||||
parameter works only for backends which support diff generation for | ||||
different paths. Other backends will raise a `ValueError` if `path1` | ||||
is set and has a different value than `path`. | ||||
:param file_path: filter this diff by given path pattern | ||||
""" | ||||
raise NotImplementedError | ||||
def strip(self, commit_id, branch=None): | ||||
""" | ||||
Strip given commit_id from the repository | ||||
""" | ||||
raise NotImplementedError | ||||
def get_common_ancestor(self, commit_id1, commit_id2, repo2): | ||||
""" | ||||
Return a latest common ancestor commit if one exists for this repo | ||||
`commit_id1` vs `commit_id2` from `repo2`. | ||||
:param commit_id1: Commit it from this repository to use as a | ||||
target for the comparison. | ||||
:param commit_id2: Source commit id to use for comparison. | ||||
:param repo2: Source repository to use for comparison. | ||||
""" | ||||
raise NotImplementedError | ||||
def compare(self, commit_id1, commit_id2, repo2, merge, pre_load=None): | ||||
""" | ||||
Compare this repository's revision `commit_id1` with `commit_id2`. | ||||
Returns a tuple(commits, ancestor) that would be merged from | ||||
`commit_id2`. Doing a normal compare (``merge=False``), ``None`` | ||||
will be returned as ancestor. | ||||
:param commit_id1: Commit it from this repository to use as a | ||||
target for the comparison. | ||||
:param commit_id2: Source commit id to use for comparison. | ||||
:param repo2: Source repository to use for comparison. | ||||
:param merge: If set to ``True`` will do a merge compare which also | ||||
returns the common ancestor. | ||||
:param pre_load: Optional. List of commit attributes to load. | ||||
""" | ||||
raise NotImplementedError | ||||
def merge( | ||||
self, | ||||
repo_id, | ||||
workspace_id, | ||||
target_ref, | ||||
source_repo, | ||||
source_ref, | ||||
user_name="", | ||||
user_email="", | ||||
message="", | ||||
dry_run=False, | ||||
use_rebase=False, | ||||
close_branch=False, | ||||
): | ||||
""" | ||||
Merge the revisions specified in `source_ref` from `source_repo` | ||||
onto the `target_ref` of this repository. | ||||
`source_ref` and `target_ref` are named tupls with the following | ||||
fields `type`, `name` and `commit_id`. | ||||
Returns a MergeResponse named tuple with the following fields | ||||
'possible', 'executed', 'source_commit', 'target_commit', | ||||
'merge_commit'. | ||||
:param repo_id: `repo_id` target repo id. | ||||
:param workspace_id: `workspace_id` unique identifier. | ||||
:param target_ref: `target_ref` points to the commit on top of which | ||||
the `source_ref` should be merged. | ||||
:param source_repo: The repository that contains the commits to be | ||||
merged. | ||||
:param source_ref: `source_ref` points to the topmost commit from | ||||
the `source_repo` which should be merged. | ||||
:param user_name: Merge commit `user_name`. | ||||
:param user_email: Merge commit `user_email`. | ||||
:param message: Merge commit `message`. | ||||
:param dry_run: If `True` the merge will not take place. | ||||
:param use_rebase: If `True` commits from the source will be rebased | ||||
on top of the target instead of being merged. | ||||
:param close_branch: If `True` branch will be close before merging it | ||||
""" | ||||
if dry_run: | ||||
message = message or settings.MERGE_DRY_RUN_MESSAGE | ||||
user_email = user_email or settings.MERGE_DRY_RUN_EMAIL | ||||
user_name = user_name or settings.MERGE_DRY_RUN_USER | ||||
else: | ||||
if not user_name: | ||||
raise ValueError("user_name cannot be empty") | ||||
if not user_email: | ||||
raise ValueError("user_email cannot be empty") | ||||
if not message: | ||||
raise ValueError("message cannot be empty") | ||||
try: | ||||
return self._merge_repo( | ||||
repo_id, | ||||
workspace_id, | ||||
target_ref, | ||||
source_repo, | ||||
source_ref, | ||||
message, | ||||
user_name, | ||||
user_email, | ||||
dry_run=dry_run, | ||||
use_rebase=use_rebase, | ||||
close_branch=close_branch, | ||||
) | ||||
except RepositoryError as exc: | ||||
log.exception("Unexpected failure when running merge, dry-run=%s", dry_run) | ||||
return MergeResponse(False, False, None, MergeFailureReason.UNKNOWN, metadata={"exception": str(exc)}) | ||||
def _merge_repo( | ||||
self, | ||||
repo_id, | ||||
workspace_id, | ||||
target_ref, | ||||
source_repo, | ||||
source_ref, | ||||
merge_message, | ||||
merger_name, | ||||
merger_email, | ||||
dry_run=False, | ||||
use_rebase=False, | ||||
close_branch=False, | ||||
): | ||||
"""Internal implementation of merge.""" | ||||
raise NotImplementedError | ||||
def _maybe_prepare_merge_workspace(self, repo_id, workspace_id, target_ref, source_ref): | ||||
""" | ||||
Create the merge workspace. | ||||
:param workspace_id: `workspace_id` unique identifier. | ||||
""" | ||||
raise NotImplementedError | ||||
@classmethod | ||||
def _get_legacy_shadow_repository_path(cls, repo_path, workspace_id): | ||||
""" | ||||
Legacy version that was used before. We still need it for | ||||
backward compat | ||||
""" | ||||
return os.path.join(os.path.dirname(repo_path), f".__shadow_{os.path.basename(repo_path)}_{workspace_id}") | ||||
@classmethod | ||||
def _get_shadow_repository_path(cls, repo_path, repo_id, workspace_id): | ||||
# The name of the shadow repository must start with '.', so it is | ||||
# skipped by 'rhodecode.lib.utils.get_filesystem_repos'. | ||||
legacy_repository_path = cls._get_legacy_shadow_repository_path(repo_path, workspace_id) | ||||
if os.path.exists(legacy_repository_path): | ||||
return legacy_repository_path | ||||
else: | ||||
return os.path.join(os.path.dirname(repo_path), f".__shadow_repo_{repo_id}_{workspace_id}") | ||||
def cleanup_merge_workspace(self, repo_id, workspace_id): | ||||
""" | ||||
Remove merge workspace. | ||||
This function MUST not fail in case there is no workspace associated to | ||||
the given `workspace_id`. | ||||
:param workspace_id: `workspace_id` unique identifier. | ||||
""" | ||||
shadow_repository_path = self._get_shadow_repository_path(self.path, repo_id, workspace_id) | ||||
shadow_repository_path_del = "{}.{}.delete".format(shadow_repository_path, time.time()) | ||||
# move the shadow repo, so it never conflicts with the one used. | ||||
# we use this method because shutil.rmtree had some edge case problems | ||||
# removing symlinked repositories | ||||
if not os.path.isdir(shadow_repository_path): | ||||
return | ||||
shutil.move(shadow_repository_path, shadow_repository_path_del) | ||||
try: | ||||
shutil.rmtree(shadow_repository_path_del, ignore_errors=False) | ||||
except Exception: | ||||
log.exception("Failed to gracefully remove shadow repo under %s", shadow_repository_path_del) | ||||
shutil.rmtree(shadow_repository_path_del, ignore_errors=True) | ||||
# ========== # | ||||
# COMMIT API # | ||||
# ========== # | ||||
@LazyProperty | ||||
def in_memory_commit(self): | ||||
""" | ||||
Returns :class:`InMemoryCommit` object for this repository. | ||||
""" | ||||
raise NotImplementedError | ||||
# ======================== # | ||||
# UTILITIES FOR SUBCLASSES # | ||||
# ======================== # | ||||
def _validate_diff_commits(self, commit1, commit2): | ||||
""" | ||||
Validates that the given commits are related to this repository. | ||||
Intended as a utility for sub classes to have a consistent validation | ||||
of input parameters in methods like :meth:`get_diff`. | ||||
""" | ||||
self._validate_commit(commit1) | ||||
self._validate_commit(commit2) | ||||
if isinstance(commit1, EmptyCommit) and isinstance(commit2, EmptyCommit): | ||||
raise ValueError("Cannot compare two empty commits") | ||||
def _validate_commit(self, commit): | ||||
if not isinstance(commit, BaseCommit): | ||||
raise TypeError("%s is not of type BaseCommit" % repr(commit)) | ||||
if commit.repository != self and not isinstance(commit, EmptyCommit): | ||||
raise ValueError( | ||||
"Commit %s must be a valid commit from this repository %s, " | ||||
"related to this repository instead %s." % (commit, self, commit.repository) | ||||
) | ||||
def _validate_commit_id(self, commit_id): | ||||
if not isinstance(commit_id, str): | ||||
raise TypeError(f"commit_id must be a string value got {type(commit_id)} instead") | ||||
def _validate_commit_idx(self, commit_idx): | ||||
if not isinstance(commit_idx, int): | ||||
raise TypeError(f"commit_idx must be a numeric value, got {type(commit_idx)}") | ||||
def _validate_branch_name(self, branch_name): | ||||
if branch_name and branch_name not in self.branches_all: | ||||
msg = f"Branch {branch_name} not found in {self}" | ||||
raise BranchDoesNotExistError(msg) | ||||
# | ||||
# Supporting deprecated API parts | ||||
# TODO: johbo: consider to move this into a mixin | ||||
# | ||||
@property | ||||
def EMPTY_CHANGESET(self): | ||||
warnings.warn("Use EMPTY_COMMIT or EMPTY_COMMIT_ID instead", DeprecationWarning) | ||||
return self.EMPTY_COMMIT_ID | ||||
@property | ||||
def revisions(self): | ||||
warnings.warn("Use commits attribute instead", DeprecationWarning) | ||||
return self.commit_ids | ||||
@revisions.setter | ||||
def revisions(self, value): | ||||
warnings.warn("Use commits attribute instead", DeprecationWarning) | ||||
self.commit_ids = value | ||||
def get_changeset(self, revision=None, pre_load=None): | ||||
warnings.warn("Use get_commit instead", DeprecationWarning) | ||||
commit_id = None | ||||
commit_idx = None | ||||
if isinstance(revision, str): | ||||
commit_id = revision | ||||
else: | ||||
commit_idx = revision | ||||
return self.get_commit(commit_id=commit_id, commit_idx=commit_idx, pre_load=pre_load) | ||||
def get_changesets(self, start=None, end=None, start_date=None, end_date=None, branch_name=None, pre_load=None): | ||||
warnings.warn("Use get_commits instead", DeprecationWarning) | ||||
start_id = self._revision_to_commit(start) | ||||
end_id = self._revision_to_commit(end) | ||||
return self.get_commits( | ||||
start_id=start_id, | ||||
end_id=end_id, | ||||
start_date=start_date, | ||||
end_date=end_date, | ||||
branch_name=branch_name, | ||||
pre_load=pre_load, | ||||
) | ||||
def _revision_to_commit(self, revision): | ||||
""" | ||||
Translates a revision to a commit_id | ||||
Helps to support the old changeset based API which allows to use | ||||
commit ids and commit indices interchangeable. | ||||
""" | ||||
if revision is None: | ||||
return revision | ||||
if isinstance(revision, str): | ||||
commit_id = revision | ||||
else: | ||||
commit_id = self.commit_ids[revision] | ||||
return commit_id | ||||
@property | ||||
def in_memory_changeset(self): | ||||
warnings.warn("Use in_memory_commit instead", DeprecationWarning) | ||||
return self.in_memory_commit | ||||
def get_path_permissions(self, username): | ||||
""" | ||||
Returns a path permission checker or None if not supported | ||||
:param username: session user name | ||||
:return: an instance of BasePathPermissionChecker or None | ||||
""" | ||||
return None | ||||
def install_hooks(self, force=False): | ||||
return self._remote.install_hooks(force) | ||||
def get_hooks_info(self): | ||||
return self._remote.get_hooks_info() | ||||
def vcsserver_invalidate_cache(self, delete=False): | ||||
return self._remote.vcsserver_invalidate_cache(delete) | ||||
r1 | class BaseChangesetClass(type): | |||
def __instancecheck__(self, instance): | ||||
return isinstance(instance, BaseCommit) | ||||
r5074 | class BaseChangeset(BaseCommit, metaclass=BaseChangesetClass): | |||
r1 | def __new__(cls, *args, **kwargs): | |||
r5647 | warnings.warn("Use BaseCommit instead of BaseChangeset", DeprecationWarning) | |||
r5091 | return super().__new__(cls, *args, **kwargs) | |||
r1 | ||||
class BaseInMemoryCommit(object): | ||||
""" | ||||
Represents differences between repository's state (most recent head) and | ||||
changes made *in place*. | ||||
**Attributes** | ||||
``repository`` | ||||
repository object for this in-memory-commit | ||||
``added`` | ||||
list of ``FileNode`` objects marked as *added* | ||||
``changed`` | ||||
list of ``FileNode`` objects marked as *changed* | ||||
``removed`` | ||||
r5647 | list of ``FileNode`` objects marked to be *removed* | |||
r1 | ||||
``parents`` | ||||
list of :class:`BaseCommit` instances representing parents of | ||||
in-memory commit. Should always be 2-element sequence. | ||||
""" | ||||
def __init__(self, repository): | ||||
self.repository = repository | ||||
self.added = [] | ||||
self.changed = [] | ||||
self.removed = [] | ||||
self.parents = [] | ||||
def add(self, *filenodes): | ||||
""" | ||||
Marks given ``FileNode`` objects as *to be committed*. | ||||
:raises ``NodeAlreadyExistsError``: if node with same path exists at | ||||
latest commit | ||||
:raises ``NodeAlreadyAddedError``: if node with same path is already | ||||
marked as *added* | ||||
""" | ||||
# Check if not already marked as *added* first | ||||
for node in filenodes: | ||||
if node.path in (n.path for n in self.added): | ||||
r5647 | raise NodeAlreadyAddedError(f"Such FileNode {node.path} is already marked for addition") | |||
r1 | for node in filenodes: | |||
self.added.append(node) | ||||
def change(self, *filenodes): | ||||
""" | ||||
Marks given ``FileNode`` objects to be *changed* in next commit. | ||||
:raises ``EmptyRepositoryError``: if there are no commits yet | ||||
:raises ``NodeAlreadyExistsError``: if node with same path is already | ||||
marked to be *changed* | ||||
:raises ``NodeAlreadyRemovedError``: if node with same path is already | ||||
marked to be *removed* | ||||
:raises ``NodeDoesNotExistError``: if node doesn't exist in latest | ||||
commit | ||||
:raises ``NodeNotChangedError``: if node hasn't really be changed | ||||
""" | ||||
for node in filenodes: | ||||
if node.path in (n.path for n in self.removed): | ||||
r5647 | raise NodeAlreadyRemovedError("Node at %s is already marked as removed" % node.path) | |||
r1 | try: | |||
self.repository.get_commit() | ||||
except EmptyRepositoryError: | ||||
r5647 | raise EmptyRepositoryError("Nothing to change - try to *add* new nodes rather than " "changing them") | |||
r1 | for node in filenodes: | |||
if node.path in (n.path for n in self.changed): | ||||
r5647 | raise NodeAlreadyChangedError("Node at '%s' is already marked as changed" % node.path) | |||
r1 | self.changed.append(node) | |||
def remove(self, *filenodes): | ||||
""" | ||||
r5647 | Marks given ``FileNode`` objects to be *removed* in next commit. | |||
r1 | ||||
:raises ``NodeAlreadyRemovedError``: if node has been already marked to | ||||
be *removed* | ||||
:raises ``NodeAlreadyChangedError``: if node has been already marked to | ||||
be *changed* | ||||
""" | ||||
for node in filenodes: | ||||
if node.path in (n.path for n in self.removed): | ||||
r5647 | raise NodeAlreadyRemovedError("Node is already marked to for removal at %s" % node.path) | |||
r1 | if node.path in (n.path for n in self.changed): | |||
r5647 | raise NodeAlreadyChangedError("Node is already marked to be changed at %s" % node.path) | |||
r1 | # We only mark node as *removed* - real removal is done by | |||
# commit method | ||||
self.removed.append(node) | ||||
def reset(self): | ||||
""" | ||||
Resets this instance to initial state (cleans ``added``, ``changed`` | ||||
and ``removed`` lists). | ||||
""" | ||||
self.added = [] | ||||
self.changed = [] | ||||
self.removed = [] | ||||
self.parents = [] | ||||
def get_ipaths(self): | ||||
""" | ||||
Returns generator of paths from nodes marked as added, changed or | ||||
removed. | ||||
""" | ||||
for node in itertools.chain(self.added, self.changed, self.removed): | ||||
yield node.path | ||||
def get_paths(self): | ||||
""" | ||||
Returns list of paths from nodes marked as added, changed or removed. | ||||
""" | ||||
return list(self.get_ipaths()) | ||||
def check_integrity(self, parents=None): | ||||
""" | ||||
Checks in-memory commit's integrity. Also, sets parents if not | ||||
already set. | ||||
:raises CommitError: if any error occurs (i.e. | ||||
``NodeDoesNotExistError``). | ||||
""" | ||||
if not self.parents: | ||||
parents = parents or [] | ||||
if len(parents) == 0: | ||||
try: | ||||
parents = [self.repository.get_commit(), None] | ||||
except EmptyRepositoryError: | ||||
parents = [None, None] | ||||
elif len(parents) == 1: | ||||
parents += [None] | ||||
self.parents = parents | ||||
# Local parents, only if not None | ||||
parents = [p for p in self.parents if p] | ||||
# Check nodes marked as added | ||||
for p in parents: | ||||
for node in self.added: | ||||
try: | ||||
r5647 | p.get_node(node.bytes_path) | |||
r1 | except NodeDoesNotExistError: | |||
pass | ||||
else: | ||||
r5647 | raise NodeAlreadyExistsError(f"Node `{node.path}` already exists at {p}") | |||
r1 | ||||
# Check nodes marked as changed | ||||
missing = set(self.changed) | ||||
not_changed = set(self.changed) | ||||
if self.changed and not parents: | ||||
raise NodeDoesNotExistError(str(self.changed[0].path)) | ||||
for p in parents: | ||||
for node in self.changed: | ||||
try: | ||||
r5647 | old = p.get_node(node.bytes_path) | |||
r1 | missing.remove(node) | |||
# if content actually changed, remove node from not_changed | ||||
if old.content != node.content: | ||||
not_changed.remove(node) | ||||
except NodeDoesNotExistError: | ||||
pass | ||||
if self.changed and missing: | ||||
r5647 | raise NodeDoesNotExistError(f"Node `{node.path}` marked as modified but missing in parents: {parents}") | |||
r1 | ||||
if self.changed and not_changed: | ||||
raise NodeNotChangedError( | ||||
r5647 | "Node `%s` wasn't actually changed (parents: %s)" % (not_changed.pop().path, parents) | |||
) | ||||
r1 | ||||
# Check nodes marked as removed | ||||
if self.removed and not parents: | ||||
raise NodeDoesNotExistError( | ||||
r5647 | "Cannot remove node at %s as there " "were no parents specified" % self.removed[0].path | |||
) | ||||
r1 | really_removed = set() | |||
for p in parents: | ||||
for node in self.removed: | ||||
try: | ||||
r5647 | p.get_node(node.bytes_path) | |||
r1 | really_removed.add(node) | |||
except CommitError: | ||||
pass | ||||
not_removed = set(self.removed) - really_removed | ||||
if not_removed: | ||||
# TODO: johbo: This code branch does not seem to be covered | ||||
raise NodeDoesNotExistError( | ||||
r5647 | "Cannot remove node at %s from " "following parents: %s" % (not_removed, parents) | |||
) | ||||
r1 | ||||
r3743 | def commit(self, message, author, parents=None, branch=None, date=None, **kwargs): | |||
r1 | """ | |||
Performs in-memory commit (doesn't check workdir in any way) and | ||||
returns newly created :class:`BaseCommit`. Updates repository's | ||||
attribute `commits`. | ||||
.. note:: | ||||
While overriding this method each backend's should call | ||||
``self.check_integrity(parents)`` in the first place. | ||||
:param message: message of the commit | ||||
:param author: full username, i.e. "Joe Doe <joe.doe@example.com>" | ||||
:param parents: single parent or sequence of parents from which commit | ||||
would be derived | ||||
:param date: ``datetime.datetime`` instance. Defaults to | ||||
``datetime.datetime.now()``. | ||||
:param branch: branch name, as string. If none given, default backend's | ||||
branch would be used. | ||||
:raises ``CommitError``: if any error occurs while committing | ||||
""" | ||||
raise NotImplementedError | ||||
class BaseInMemoryChangesetClass(type): | ||||
def __instancecheck__(self, instance): | ||||
return isinstance(instance, BaseInMemoryCommit) | ||||
r5074 | class BaseInMemoryChangeset(BaseInMemoryCommit, metaclass=BaseInMemoryChangesetClass): | |||
r1 | def __new__(cls, *args, **kwargs): | |||
r5647 | warnings.warn("Use BaseCommit instead of BaseInMemoryCommit", DeprecationWarning) | |||
r5091 | return super().__new__(cls, *args, **kwargs) | |||
r1 | ||||
class EmptyCommit(BaseCommit): | ||||
""" | ||||
An dummy empty commit. It's possible to pass hash when creating | ||||
an EmptyCommit | ||||
""" | ||||
r5647 | def __init__(self, commit_id=EMPTY_COMMIT_ID, repo=None, alias=None, idx=-1, message="", author="", date=None): | |||
r1 | self._empty_commit_id = commit_id | |||
# TODO: johbo: Solve idx parameter, default value does not make | ||||
# too much sense | ||||
self.idx = idx | ||||
self.message = message | ||||
self.author = author | ||||
self.date = date or datetime.datetime.fromtimestamp(0) | ||||
self.repository = repo | ||||
self.alias = alias | ||||
@LazyProperty | ||||
def raw_id(self): | ||||
""" | ||||
Returns raw string identifying this commit, useful for web | ||||
representation. | ||||
""" | ||||
return self._empty_commit_id | ||||
@LazyProperty | ||||
def branch(self): | ||||
if self.alias: | ||||
from rhodecode.lib.vcs.backends import get_backend | ||||
r5647 | ||||
r1 | return get_backend(self.alias).DEFAULT_BRANCH_NAME | |||
@LazyProperty | ||||
def short_id(self): | ||||
return self.raw_id[:12] | ||||
r621 | @LazyProperty | |||
def id(self): | ||||
return self.raw_id | ||||
r5074 | def get_path_commit(self, path, pre_load=None): | |||
r1 | return self | |||
r5074 | def get_file_content(self, path) -> bytes: | |||
r5647 | return b"" | |||
r1 | ||||
r3895 | def get_file_content_streamed(self, path): | |||
r5074 | yield self.get_file_content(path) | |||
r3895 | ||||
r1 | def get_file_size(self, path): | |||
return 0 | ||||
class EmptyChangesetClass(type): | ||||
def __instancecheck__(self, instance): | ||||
return isinstance(instance, EmptyCommit) | ||||
r5074 | class EmptyChangeset(EmptyCommit, metaclass=EmptyChangesetClass): | |||
r1 | def __new__(cls, *args, **kwargs): | |||
r5647 | warnings.warn("Use EmptyCommit instead of EmptyChangeset", DeprecationWarning) | |||
r1 | return super(EmptyCommit, cls).__new__(cls, *args, **kwargs) | |||
r5647 | def __init__( | |||
self, | ||||
cs=EMPTY_COMMIT_ID, | ||||
repo=None, | ||||
requested_revision=None, | ||||
alias=None, | ||||
revision=-1, | ||||
message="", | ||||
author="", | ||||
date=None, | ||||
): | ||||
r1 | if requested_revision is not None: | |||
r5647 | warnings.warn("Parameter requested_revision not supported anymore", DeprecationWarning) | |||
super().__init__(commit_id=cs, repo=repo, alias=alias, idx=revision, message=message, author=author, date=date) | ||||
r1 | ||||
@property | ||||
def revision(self): | ||||
warnings.warn("Use idx instead", DeprecationWarning) | ||||
return self.idx | ||||
@revision.setter | ||||
def revision(self, value): | ||||
warnings.warn("Use idx instead", DeprecationWarning) | ||||
self.idx = value | ||||
r1255 | class EmptyRepository(BaseRepository): | |||
def __init__(self, repo_path=None, config=None, create=False, **kwargs): | ||||
pass | ||||
def get_diff(self, *args, **kwargs): | ||||
from rhodecode.lib.vcs.backends.git.diff import GitDiff | ||||
r5647 | ||||
return GitDiff(b"") | ||||
r1255 | ||||
r1 | class CollectionGenerator(object): | |||
r3468 | def __init__(self, repo, commit_ids, collection_size=None, pre_load=None, translate_tag=None): | |||
r1 | self.repo = repo | |||
self.commit_ids = commit_ids | ||||
r5074 | self.collection_size = collection_size | |||
r1 | self.pre_load = pre_load | |||
r3468 | self.translate_tag = translate_tag | |||
r1 | ||||
def __len__(self): | ||||
if self.collection_size is not None: | ||||
return self.collection_size | ||||
return self.commit_ids.__len__() | ||||
def __iter__(self): | ||||
for commit_id in self.commit_ids: | ||||
# TODO: johbo: Mercurial passes in commit indices or commit ids | ||||
yield self._commit_factory(commit_id) | ||||
def _commit_factory(self, commit_id): | ||||
""" | ||||
Allows backends to override the way commits are generated. | ||||
""" | ||||
r5647 | return self.repo.get_commit(commit_id=commit_id, pre_load=self.pre_load, translate_tag=self.translate_tag) | |||
r1 | ||||
r5074 | def __getitem__(self, key): | |||
"""Return either a single element by index, or a sliced collection.""" | ||||
if isinstance(key, slice): | ||||
r5647 | commit_ids = self.commit_ids[key.start : key.stop] | |||
r5074 | ||||
else: | ||||
# single item | ||||
commit_ids = self.commit_ids[key] | ||||
r5647 | return self.__class__(self.repo, commit_ids, pre_load=self.pre_load, translate_tag=self.translate_tag) | |||
r1 | ||||
def __repr__(self): | ||||
r5647 | return "<CollectionGenerator[len:%s]>" % (self.__len__()) | |||
r1 | ||||
class Config(object): | ||||
""" | ||||
Represents the configuration for a repository. | ||||
The API is inspired by :class:`ConfigParser.ConfigParser` from the | ||||
standard library. It implements only the needed subset. | ||||
""" | ||||
def __init__(self): | ||||
self._values = {} | ||||
def copy(self): | ||||
clone = Config() | ||||
for section, values in self._values.items(): | ||||
clone._values[section] = values.copy() | ||||
return clone | ||||
def __repr__(self): | ||||
r5647 | return "<Config({} sections) at {}>".format(len(self._values), hex(id(self))) | |||
r1 | ||||
def items(self, section): | ||||
r4932 | return self._values.get(section, {}).items() | |||
r1 | ||||
def get(self, section, option): | ||||
return self._values.get(section, {}).get(option) | ||||
def set(self, section, option, value): | ||||
section_values = self._values.setdefault(section, {}) | ||||
section_values[option] = value | ||||
def clear_section(self, section): | ||||
self._values[section] = {} | ||||
r5250 | def drop_option(self, section, option): | |||
if section not in self._values: | ||||
r5647 | raise ValueError(f"Section {section} does not exist") | |||
r5250 | del self._values[section][option] | |||
r1 | def serialize(self): | |||
""" | ||||
Creates a list of three tuples (section, key, value) representing | ||||
this config object. | ||||
""" | ||||
items = [] | ||||
for section in self._values: | ||||
for option, value in self._values[section].items(): | ||||
r5647 | items.append((safe_str(section), safe_str(option), safe_str(value))) | |||
r1 | return items | |||
class Diff(object): | ||||
""" | ||||
Represents a diff result from a repository backend. | ||||
r1259 | Subclasses have to provide a backend specific value for | |||
:attr:`_header_re` and :attr:`_meta_re`. | ||||
r1 | """ | |||
r5647 | ||||
r1259 | _meta_re = None | |||
r5647 | _header_re: bytes = re.compile(rb"") | |||
r1 | ||||
r5074 | def __init__(self, raw_diff: bytes): | |||
if not isinstance(raw_diff, bytes): | ||||
r5647 | raise Exception(f"raw_diff must be bytes - got {type(raw_diff)}") | |||
r5074 | ||||
self.raw = memoryview(raw_diff) | ||||
def get_header_re(self): | ||||
return self._header_re | ||||
r1 | ||||
def chunks(self): | ||||
""" | ||||
split the diff in chunks of separate --git a/file b/file chunks | ||||
to make diffs consistent we must prepend with \n, and make sure | ||||
we can detect last chunk as this was also has special rule | ||||
""" | ||||
r1259 | ||||
r5647 | diff_parts = (b"\n" + bytes(self.raw)).split(b"\ndiff --git") | |||
r1259 | ||||
chunks = diff_parts[1:] | ||||
r1 | total_chunks = len(chunks) | |||
r1259 | ||||
r5074 | def diff_iter(_chunks): | |||
for cur_chunk, chunk in enumerate(_chunks, start=1): | ||||
yield DiffChunk(chunk, self, cur_chunk == total_chunks) | ||||
r5647 | ||||
r5074 | return diff_iter(chunks) | |||
r1 | ||||
class DiffChunk(object): | ||||
r5074 | def __init__(self, chunk: bytes, diff_obj: Diff, is_last_chunk: bool): | |||
self.diff_obj = diff_obj | ||||
r1 | ||||
# since we split by \ndiff --git that part is lost from original diff | ||||
# we need to re-apply it at the end, EXCEPT ! if it's last chunk | ||||
r5074 | if not is_last_chunk: | |||
r5647 | chunk += b"\n" | |||
r5074 | header_re = self.diff_obj.get_header_re() | |||
r5647 | ||||
r5074 | match = header_re.match(chunk) | |||
r1 | self.header = match.groupdict() | |||
r5647 | self.diff = chunk[match.end() :] | |||
r1 | self.raw = chunk | |||
r2618 | ||||
r5074 | @property | |||
def header_as_str(self): | ||||
if self.header: | ||||
r5647 | ||||
r5074 | def safe_str_on_bytes(val): | |||
if isinstance(val, bytes): | ||||
return safe_str(val) | ||||
return val | ||||
r5647 | ||||
r5074 | return {safe_str(k): safe_str_on_bytes(v) for k, v in self.header.items()} | |||
def __repr__(self): | ||||
r5647 | return f"DiffChunk({self.header_as_str})" | |||
r5074 | ||||
r2618 | ||||
class BasePathPermissionChecker(object): | ||||
r2619 | @staticmethod | |||
def create_from_patterns(includes, excludes): | ||||
r5647 | if includes and "*" in includes and not excludes: | |||
r2619 | return AllPathPermissionChecker() | |||
r5647 | elif excludes and "*" in excludes: | |||
r2619 | return NonePathPermissionChecker() | |||
else: | ||||
return PatternPathPermissionChecker(includes, excludes) | ||||
@property | ||||
def has_full_access(self): | ||||
r5074 | raise NotImplementedError() | |||
r2618 | ||||
def has_access(self, path): | ||||
r5074 | raise NotImplementedError() | |||
r2619 | ||||
class AllPathPermissionChecker(BasePathPermissionChecker): | ||||
@property | ||||
def has_full_access(self): | ||||
return True | ||||
def has_access(self, path): | ||||
return True | ||||
class NonePathPermissionChecker(BasePathPermissionChecker): | ||||
@property | ||||
def has_full_access(self): | ||||
return False | ||||
def has_access(self, path): | ||||
return False | ||||
class PatternPathPermissionChecker(BasePathPermissionChecker): | ||||
def __init__(self, includes, excludes): | ||||
self.includes = includes | ||||
self.excludes = excludes | ||||
r5647 | self.includes_re = [] if not includes else [re.compile(fnmatch.translate(pattern)) for pattern in includes] | |||
self.excludes_re = [] if not excludes else [re.compile(fnmatch.translate(pattern)) for pattern in excludes] | ||||
r2619 | ||||
@property | ||||
def has_full_access(self): | ||||
r5647 | return "*" in self.includes and not self.excludes | |||
r2619 | ||||
def has_access(self, path): | ||||
r2623 | for regex in self.excludes_re: | |||
if regex.match(path): | ||||
r2619 | return False | |||
r2623 | for regex in self.includes_re: | |||
if regex.match(path): | ||||
r2619 | return True | |||
r2623 | return False | |||