# RhodeCode VCSServer provides access to different vcs backends via network. # Copyright (C) 2014-2023 RhodeCode GmbH # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software Foundation, # Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA import os import sys import tempfile import traceback import logging import urllib.parse from vcsserver.lib.rc_cache.archive_cache import get_archival_cache_store from vcsserver import exceptions from vcsserver.exceptions import NoContentException from vcsserver.hgcompat import archival from vcsserver.str_utils import safe_bytes log = logging.getLogger(__name__) class RepoFactory(object): """ Utility to create instances of repository It provides internal caching of the `repo` object based on the :term:`call context`. """ repo_type = None def __init__(self): pass def _create_config(self, path, config): config = {} return config def _create_repo(self, wire, create): raise NotImplementedError() def repo(self, wire, create=False): raise NotImplementedError() def obfuscate_qs(query_string): if query_string is None: return None parsed = [] for k, v in urllib.parse.parse_qsl(query_string, keep_blank_values=True): if k in ['auth_token', 'api_key']: v = "*****" parsed.append((k, v)) return '&'.join('{}{}'.format( k, f'={v}' if v else '') for k, v in parsed) def raise_from_original(new_type, org_exc: Exception): """ Raise a new exception type with original args and traceback. """ exc_type, exc_value, exc_traceback = sys.exc_info() new_exc = new_type(*exc_value.args) # store the original traceback into the new exc new_exc._org_exc_tb = traceback.format_tb(exc_traceback) try: raise new_exc.with_traceback(exc_traceback) finally: del exc_traceback class ArchiveNode(object): def __init__(self, path, mode, is_link, raw_bytes): self.path = path self.mode = mode self.is_link = is_link self.raw_bytes = raw_bytes def store_archive_in_cache(node_walker, archive_key, kind, mtime, archive_at_path, archive_dir_name, commit_id, write_metadata=True, extra_metadata=None, cache_config=None): """ Function that would store generate archive and send it to a dedicated backend store In here we use diskcache :param node_walker: a generator returning nodes to add to archive :param archive_key: key used to store the path :param kind: archive kind :param mtime: time of creation :param archive_at_path: default '/' the path at archive was started. If this is not '/' it means it's a partial archive :param archive_dir_name: inside dir name when creating an archive :param commit_id: commit sha of revision archive was created at :param write_metadata: :param extra_metadata: :param cache_config: walker should be a file walker, for example, def node_walker(): for file_info in files: yield ArchiveNode(fn, mode, is_link, ctx[fn].data) """ extra_metadata = extra_metadata or {} d_cache = get_archival_cache_store(config=cache_config) if archive_key in d_cache: with d_cache as d_cache_reader: reader, tag = d_cache_reader.get(archive_key, read=True, tag=True, retry=True) return reader.name archive_tmp_path = safe_bytes(tempfile.mkstemp()[1]) log.debug('Creating new temp archive in %s', archive_tmp_path) if kind == "tgz": archiver = archival.tarit(archive_tmp_path, mtime, b"gz") elif kind == "tbz2": archiver = archival.tarit(archive_tmp_path, mtime, b"bz2") elif kind == 'zip': archiver = archival.zipit(archive_tmp_path, mtime) else: raise exceptions.ArchiveException()( f'Remote does not support: "{kind}" archive type.') for f in node_walker(commit_id, archive_at_path): f_path = os.path.join(safe_bytes(archive_dir_name), safe_bytes(f.path).lstrip(b'/')) try: archiver.addfile(f_path, f.mode, f.is_link, f.raw_bytes()) except NoContentException: # NOTE(marcink): this is a special case for SVN so we can create "empty" # directories which are not supported by archiver archiver.addfile(os.path.join(f_path, b'.dir'), f.mode, f.is_link, b'') if write_metadata: metadata = dict([ ('commit_id', commit_id), ('mtime', mtime), ]) metadata.update(extra_metadata) meta = [safe_bytes(f"{f_name}:{value}") for f_name, value in metadata.items()] f_path = os.path.join(safe_bytes(archive_dir_name), b'.archival.txt') archiver.addfile(f_path, 0o644, False, b'\n'.join(meta)) archiver.done() # ensure set & get are atomic with d_cache.transact(): with open(archive_tmp_path, 'rb') as archive_file: add_result = d_cache.set(archive_key, archive_file, read=True, tag='db-name', retry=True) if not add_result: log.error('Failed to store cache for key=%s', archive_key) os.remove(archive_tmp_path) reader, tag = d_cache.get(archive_key, read=True, tag=True, retry=True) if not reader: raise AssertionError(f'empty reader on key={archive_key} added={add_result}') return reader.name class BinaryEnvelope(object): def __init__(self, val): self.val = val class BytesEnvelope(bytes): def __new__(cls, content): if isinstance(content, bytes): return super().__new__(cls, content) else: raise TypeError('Content must be bytes.') class BinaryBytesEnvelope(BytesEnvelope): pass