fixture_utils.py
1697 lines
| 54.2 KiB
| text/x-python
|
PythonLexer
r5607 | # Copyright (C) 2010-2024 RhodeCode GmbH | |||
# | ||||
# This program is free software: you can redistribute it and/or modify | ||||
# it under the terms of the GNU Affero General Public License, version 3 | ||||
# (only), as published by the Free Software Foundation. | ||||
# | ||||
# This program is distributed in the hope that it will be useful, | ||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of | ||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||||
# GNU General Public License for more details. | ||||
# | ||||
# You should have received a copy of the GNU Affero General Public License | ||||
# along with this program. If not, see <http://www.gnu.org/licenses/>. | ||||
# | ||||
# This program is dual-licensed. If you wish to learn more about the | ||||
# RhodeCode Enterprise Edition, including its added features, Support services, | ||||
# and proprietary license terms, please see https://rhodecode.com/licenses/ | ||||
import collections | ||||
import datetime | ||||
import os | ||||
import re | ||||
import pprint | ||||
import shutil | ||||
import socket | ||||
import subprocess | ||||
import time | ||||
import uuid | ||||
import dateutil.tz | ||||
import logging | ||||
import functools | ||||
import textwrap | ||||
import mock | ||||
import pyramid.testing | ||||
import pytest | ||||
import colander | ||||
import requests | ||||
import pyramid.paster | ||||
import rhodecode | ||||
import rhodecode.lib | ||||
from rhodecode.model.changeset_status import ChangesetStatusModel | ||||
from rhodecode.model.comment import CommentsModel | ||||
from rhodecode.model.db import ( | ||||
PullRequest, | ||||
PullRequestReviewers, | ||||
Repository, | ||||
RhodeCodeSetting, | ||||
ChangesetStatus, | ||||
RepoGroup, | ||||
UserGroup, | ||||
RepoRhodeCodeUi, | ||||
RepoRhodeCodeSetting, | ||||
RhodeCodeUi, | ||||
) | ||||
from rhodecode.model.meta import Session | ||||
from rhodecode.model.pull_request import PullRequestModel | ||||
from rhodecode.model.repo import RepoModel | ||||
from rhodecode.model.repo_group import RepoGroupModel | ||||
from rhodecode.model.user import UserModel | ||||
from rhodecode.model.settings import VcsSettingsModel | ||||
from rhodecode.model.user_group import UserGroupModel | ||||
from rhodecode.model.integration import IntegrationModel | ||||
from rhodecode.integrations import integration_type_registry | ||||
from rhodecode.integrations.types.base import IntegrationTypeBase | ||||
from rhodecode.lib.utils import repo2db_mapper | ||||
from rhodecode.lib.str_utils import safe_bytes | ||||
from rhodecode.lib.hash_utils import sha1_safe | ||||
from rhodecode.lib.vcs.backends import get_backend | ||||
from rhodecode.lib.vcs.nodes import FileNode | ||||
from rhodecode.lib.base import bootstrap_config | ||||
from rhodecode.tests import ( | ||||
login_user_session, | ||||
get_new_dir, | ||||
utils, | ||||
TESTS_TMP_PATH, | ||||
TEST_USER_ADMIN_LOGIN, | ||||
TEST_USER_REGULAR_LOGIN, | ||||
TEST_USER_REGULAR2_LOGIN, | ||||
TEST_USER_REGULAR_PASS, | ||||
console_printer, | ||||
) | ||||
from rhodecode.tests.utils import set_anonymous_access | ||||
from rhodecode.tests.fixtures.rc_fixture import Fixture | ||||
from rhodecode.config import utils as config_utils | ||||
log = logging.getLogger(__name__) | ||||
def cmp(a, b): | ||||
# backport cmp from python2 so we can still use it in the custom code in this module | ||||
return (a > b) - (a < b) | ||||
@pytest.fixture(scope="session") | ||||
def http_environ_session(): | ||||
""" | ||||
Allow to use "http_environ" in session scope. | ||||
""" | ||||
return plain_http_environ() | ||||
def plain_http_host_stub(): | ||||
""" | ||||
Value of HTTP_HOST in the test run. | ||||
""" | ||||
return "example.com:80" | ||||
def plain_config_stub(request, request_stub): | ||||
""" | ||||
Set up pyramid.testing and return the Configurator. | ||||
""" | ||||
config = bootstrap_config(request=request_stub) | ||||
@request.addfinalizer | ||||
def cleanup(): | ||||
pyramid.testing.tearDown() | ||||
return config | ||||
def plain_request_stub(): | ||||
""" | ||||
Stub request object. | ||||
""" | ||||
from rhodecode.lib.base import bootstrap_request | ||||
_request = bootstrap_request(scheme="https") | ||||
return _request | ||||
@pytest.fixture() | ||||
def http_host_stub(): | ||||
""" | ||||
Value of HTTP_HOST in the test run. | ||||
""" | ||||
return plain_http_host_stub() | ||||
def plain_http_host_only_stub(): | ||||
""" | ||||
Value of HTTP_HOST in the test run. | ||||
""" | ||||
return plain_http_host_stub().split(":")[0] | ||||
@pytest.fixture() | ||||
def http_host_only_stub(): | ||||
""" | ||||
Value of HTTP_HOST in the test run. | ||||
""" | ||||
return plain_http_host_only_stub() | ||||
def plain_http_environ(): | ||||
""" | ||||
HTTP extra environ keys. | ||||
Used by the test application and as well for setting up the pylons | ||||
environment. In the case of the fixture "app" it should be possible | ||||
to override this for a specific test case. | ||||
""" | ||||
return { | ||||
"SERVER_NAME": plain_http_host_only_stub(), | ||||
"SERVER_PORT": plain_http_host_stub().split(":")[1], | ||||
"HTTP_HOST": plain_http_host_stub(), | ||||
"HTTP_USER_AGENT": "rc-test-agent", | ||||
"REQUEST_METHOD": "GET", | ||||
} | ||||
@pytest.fixture(scope="session") | ||||
def baseapp(request, ini_config, http_environ_session, available_port_factory, vcsserver_factory, celery_factory): | ||||
from rhodecode.lib.config_utils import get_app_config | ||||
from rhodecode.config.middleware import make_pyramid_app | ||||
log.info("Using the RhodeCode configuration:%s", ini_config) | ||||
pyramid.paster.setup_logging(ini_config) | ||||
settings = get_app_config(ini_config) | ||||
store_dir = os.path.dirname(ini_config) | ||||
# start vcsserver | ||||
_vcsserver_port = available_port_factory() | ||||
vcsserver_instance = vcsserver_factory( | ||||
request, | ||||
store_dir=store_dir, | ||||
port=_vcsserver_port, | ||||
info_prefix="base-app-" | ||||
) | ||||
settings["vcs.server"] = vcsserver_instance.bind_addr | ||||
# we skip setting store_dir for baseapp, it's internally set via testing rhodecode.ini | ||||
# settings['repo_store.path'] = str(store_dir) | ||||
console_printer(f' :warning: [green]pytest-setup[/green] Starting base pyramid-app: {ini_config}') | ||||
pyramid_baseapp = make_pyramid_app({"__file__": ini_config}, **settings) | ||||
# start celery | ||||
celery_factory( | ||||
request, | ||||
store_dir=store_dir, | ||||
port=None, | ||||
info_prefix="base-app-", | ||||
overrides=( | ||||
{'handler_console': {'level': 'DEBUG'}}, | ||||
{'app:main': {'vcs.server': vcsserver_instance.bind_addr}}, | ||||
{'app:main': {'repo_store.path': store_dir}} | ||||
) | ||||
) | ||||
return pyramid_baseapp | ||||
@pytest.fixture(scope="session") | ||||
def app_settings(baseapp, ini_config): | ||||
""" | ||||
Settings dictionary used to create the app. | ||||
Parses the ini file and passes the result through the sanitize and apply | ||||
defaults mechanism in `rhodecode.config.middleware`. | ||||
""" | ||||
return baseapp.config.get_settings() | ||||
@pytest.fixture(scope="session") | ||||
def db_connection(ini_settings): | ||||
# Initialize the database connection. | ||||
config_utils.initialize_database(ini_settings) | ||||
LoginData = collections.namedtuple("LoginData", ("csrf_token", "user")) | ||||
def _autologin_user(app, *args): | ||||
session = login_user_session(app, *args) | ||||
csrf_token = rhodecode.lib.auth.get_csrf_token(session) | ||||
return LoginData(csrf_token, session["rhodecode_user"]) | ||||
@pytest.fixture() | ||||
def autologin_user(app): | ||||
""" | ||||
Utility fixture which makes sure that the admin user is logged in | ||||
""" | ||||
return _autologin_user(app) | ||||
@pytest.fixture() | ||||
def autologin_regular_user(app): | ||||
""" | ||||
Utility fixture which makes sure that the regular user is logged in | ||||
""" | ||||
return _autologin_user(app, TEST_USER_REGULAR_LOGIN, TEST_USER_REGULAR_PASS) | ||||
@pytest.fixture(scope="function") | ||||
def csrf_token(request, autologin_user): | ||||
return autologin_user.csrf_token | ||||
@pytest.fixture(scope="function") | ||||
def xhr_header(request): | ||||
return {"HTTP_X_REQUESTED_WITH": "XMLHttpRequest"} | ||||
@pytest.fixture() | ||||
def real_crypto_backend(monkeypatch): | ||||
""" | ||||
Switch the production crypto backend on for this test. | ||||
During the test run the crypto backend is replaced with a faster | ||||
implementation based on the MD5 algorithm. | ||||
""" | ||||
monkeypatch.setattr(rhodecode, "is_test", False) | ||||
@pytest.fixture(scope="class") | ||||
def index_location(request, baseapp): | ||||
index_location = baseapp.config.get_settings()["search.location"] | ||||
if request.cls: | ||||
request.cls.index_location = index_location | ||||
return index_location | ||||
@pytest.fixture(scope="session", autouse=True) | ||||
def tests_tmp_path(request): | ||||
""" | ||||
Create temporary directory to be used during the test session. | ||||
""" | ||||
if not os.path.exists(TESTS_TMP_PATH): | ||||
os.makedirs(TESTS_TMP_PATH) | ||||
if not request.config.getoption("--keep-tmp-path"): | ||||
@request.addfinalizer | ||||
def remove_tmp_path(): | ||||
shutil.rmtree(TESTS_TMP_PATH) | ||||
return TESTS_TMP_PATH | ||||
@pytest.fixture() | ||||
def test_repo_group(request): | ||||
""" | ||||
Create a temporary repository group, and destroy it after | ||||
usage automatically | ||||
""" | ||||
fixture = Fixture() | ||||
repogroupid = "test_repo_group_%s" % str(time.time()).replace(".", "") | ||||
repo_group = fixture.create_repo_group(repogroupid) | ||||
def _cleanup(): | ||||
fixture.destroy_repo_group(repogroupid) | ||||
request.addfinalizer(_cleanup) | ||||
return repo_group | ||||
@pytest.fixture() | ||||
def test_user_group(request): | ||||
""" | ||||
Create a temporary user group, and destroy it after | ||||
usage automatically | ||||
""" | ||||
fixture = Fixture() | ||||
usergroupid = "test_user_group_%s" % str(time.time()).replace(".", "") | ||||
user_group = fixture.create_user_group(usergroupid) | ||||
def _cleanup(): | ||||
fixture.destroy_user_group(user_group) | ||||
request.addfinalizer(_cleanup) | ||||
return user_group | ||||
@pytest.fixture(scope="session") | ||||
def test_repo(request): | ||||
container = TestRepoContainer() | ||||
request.addfinalizer(container._cleanup) | ||||
return container | ||||
class TestRepoContainer(object): | ||||
""" | ||||
Container for test repositories which are used read only. | ||||
Repositories will be created on demand and re-used during the lifetime | ||||
of this object. | ||||
Usage to get the svn test repository "minimal":: | ||||
test_repo = TestContainer() | ||||
repo = test_repo('minimal', 'svn') | ||||
""" | ||||
dump_extractors = { | ||||
"git": utils.extract_git_repo_from_dump, | ||||
"hg": utils.extract_hg_repo_from_dump, | ||||
"svn": utils.extract_svn_repo_from_dump, | ||||
} | ||||
def __init__(self): | ||||
self._cleanup_repos = [] | ||||
self._fixture = Fixture() | ||||
self._repos = {} | ||||
def __call__(self, dump_name, backend_alias, config=None): | ||||
key = (dump_name, backend_alias) | ||||
if key not in self._repos: | ||||
repo = self._create_repo(dump_name, backend_alias, config) | ||||
self._repos[key] = repo.repo_id | ||||
return Repository.get(self._repos[key]) | ||||
def _create_repo(self, dump_name, backend_alias, config): | ||||
repo_name = f"{backend_alias}-{dump_name}" | ||||
backend = get_backend(backend_alias) | ||||
dump_extractor = self.dump_extractors[backend_alias] | ||||
repo_path = dump_extractor(dump_name, repo_name) | ||||
vcs_repo = backend(repo_path, config=config) | ||||
repo2db_mapper({repo_name: vcs_repo}) | ||||
repo = RepoModel().get_by_repo_name(repo_name) | ||||
self._cleanup_repos.append(repo_name) | ||||
return repo | ||||
def _cleanup(self): | ||||
for repo_name in reversed(self._cleanup_repos): | ||||
self._fixture.destroy_repo(repo_name) | ||||
def backend_base(request, backend_alias, test_repo): | ||||
if backend_alias not in request.config.getoption("--backends"): | ||||
pytest.skip(f"Backend {backend_alias} not selected.") | ||||
utils.check_xfail_backends(request.node, backend_alias) | ||||
utils.check_skip_backends(request.node, backend_alias) | ||||
repo_name = "vcs_test_%s" % (backend_alias,) | ||||
backend = Backend( | ||||
alias=backend_alias, repo_name=repo_name, test_name=request.node.name, test_repo_container=test_repo | ||||
) | ||||
request.addfinalizer(backend.cleanup) | ||||
return backend | ||||
@pytest.fixture() | ||||
def backend(request, backend_alias, baseapp, test_repo): | ||||
""" | ||||
Parametrized fixture which represents a single backend implementation. | ||||
It respects the option `--backends` to focus the test run on specific | ||||
backend implementations. | ||||
It also supports `pytest.mark.xfail_backends` to mark tests as failing | ||||
for specific backends. This is intended as a utility for incremental | ||||
development of a new backend implementation. | ||||
""" | ||||
return backend_base(request, backend_alias, test_repo) | ||||
@pytest.fixture() | ||||
def backend_git(request, baseapp, test_repo): | ||||
return backend_base(request, "git", test_repo) | ||||
@pytest.fixture() | ||||
def backend_hg(request, baseapp, test_repo): | ||||
return backend_base(request, "hg", test_repo) | ||||
@pytest.fixture() | ||||
def backend_svn(request, baseapp, test_repo): | ||||
return backend_base(request, "svn", test_repo) | ||||
@pytest.fixture() | ||||
def backend_random(backend_git): | ||||
""" | ||||
Use this to express that your tests need "a backend. | ||||
A few of our tests need a backend, so that we can run the code. This | ||||
fixture is intended to be used for such cases. It will pick one of the | ||||
backends and run the tests. | ||||
The fixture `backend` would run the test multiple times for each | ||||
available backend which is a pure waste of time if the test is | ||||
independent of the backend type. | ||||
""" | ||||
# TODO: johbo: Change this to pick a random backend | ||||
return backend_git | ||||
@pytest.fixture() | ||||
def backend_stub(backend_git): | ||||
""" | ||||
Use this to express that your tests need a backend stub | ||||
TODO: mikhail: Implement a real stub logic instead of returning | ||||
a git backend | ||||
""" | ||||
return backend_git | ||||
@pytest.fixture() | ||||
def repo_stub(backend_stub): | ||||
""" | ||||
Use this to express that your tests need a repository stub | ||||
""" | ||||
return backend_stub.create_repo() | ||||
class Backend(object): | ||||
""" | ||||
Represents the test configuration for one supported backend | ||||
Provides easy access to different test repositories based on | ||||
`__getitem__`. Such repositories will only be created once per test | ||||
session. | ||||
""" | ||||
invalid_repo_name = re.compile(r"[^0-9a-zA-Z]+") | ||||
_master_repo = None | ||||
_master_repo_path = "" | ||||
_commit_ids = {} | ||||
def __init__(self, alias, repo_name, test_name, test_repo_container): | ||||
self.alias = alias | ||||
self.repo_name = repo_name | ||||
self._cleanup_repos = [] | ||||
self._test_name = test_name | ||||
self._test_repo_container = test_repo_container | ||||
# TODO: johbo: Used as a delegate interim. Not yet sure if Backend or | ||||
# Fixture will survive in the end. | ||||
self._fixture = Fixture() | ||||
def __getitem__(self, key): | ||||
return self._test_repo_container(key, self.alias) | ||||
def create_test_repo(self, key, config=None): | ||||
return self._test_repo_container(key, self.alias, config) | ||||
@property | ||||
def repo_id(self): | ||||
# just fake some repo_id | ||||
return self.repo.repo_id | ||||
@property | ||||
def repo(self): | ||||
""" | ||||
Returns the "current" repository. This is the vcs_test repo or the | ||||
last repo which has been created with `create_repo`. | ||||
""" | ||||
from rhodecode.model.db import Repository | ||||
return Repository.get_by_repo_name(self.repo_name) | ||||
@property | ||||
def default_branch_name(self): | ||||
VcsRepository = get_backend(self.alias) | ||||
return VcsRepository.DEFAULT_BRANCH_NAME | ||||
@property | ||||
def default_head_id(self): | ||||
""" | ||||
Returns the default head id of the underlying backend. | ||||
This will be the default branch name in case the backend does have a | ||||
default branch. In the other cases it will point to a valid head | ||||
which can serve as the base to create a new commit on top of it. | ||||
""" | ||||
vcsrepo = self.repo.scm_instance() | ||||
head_id = vcsrepo.DEFAULT_BRANCH_NAME or vcsrepo.commit_ids[-1] | ||||
return head_id | ||||
@property | ||||
def commit_ids(self): | ||||
""" | ||||
Returns the list of commits for the last created repository | ||||
""" | ||||
return self._commit_ids | ||||
def create_master_repo(self, commits): | ||||
""" | ||||
Create a repository and remember it as a template. | ||||
This allows to easily create derived repositories to construct | ||||
more complex scenarios for diff, compare and pull requests. | ||||
Returns a commit map which maps from commit message to raw_id. | ||||
""" | ||||
self._master_repo = self.create_repo(commits=commits) | ||||
self._master_repo_path = self._master_repo.repo_full_path | ||||
return self._commit_ids | ||||
def create_repo(self, commits=None, number_of_commits=0, heads=None, name_suffix="", bare=False, **kwargs): | ||||
""" | ||||
Create a repository and record it for later cleanup. | ||||
:param commits: Optional. A sequence of dict instances. | ||||
Will add a commit per entry to the new repository. | ||||
:param number_of_commits: Optional. If set to a number, this number of | ||||
commits will be added to the new repository. | ||||
:param heads: Optional. Can be set to a sequence of of commit | ||||
names which shall be pulled in from the master repository. | ||||
:param name_suffix: adds special suffix to generated repo name | ||||
:param bare: set a repo as bare (no checkout) | ||||
""" | ||||
self.repo_name = self._next_repo_name() + name_suffix | ||||
repo = self._fixture.create_repo(self.repo_name, repo_type=self.alias, bare=bare, **kwargs) | ||||
self._cleanup_repos.append(repo.repo_name) | ||||
commits = commits or [{"message": f"Commit {x} of {self.repo_name}"} for x in range(number_of_commits)] | ||||
vcs_repo = repo.scm_instance() | ||||
vcs_repo.count() | ||||
self._add_commits_to_repo(vcs_repo, commits) | ||||
if heads: | ||||
self.pull_heads(repo, heads) | ||||
return repo | ||||
def pull_heads(self, repo, heads, do_fetch=False): | ||||
""" | ||||
Make sure that repo contains all commits mentioned in `heads` | ||||
""" | ||||
vcsrepo = repo.scm_instance() | ||||
vcsrepo.config.clear_section("hooks") | ||||
commit_ids = [self._commit_ids[h] for h in heads] | ||||
if do_fetch: | ||||
vcsrepo.fetch(self._master_repo_path, commit_ids=commit_ids) | ||||
vcsrepo.pull(self._master_repo_path, commit_ids=commit_ids) | ||||
def create_fork(self): | ||||
repo_to_fork = self.repo_name | ||||
self.repo_name = self._next_repo_name() | ||||
repo = self._fixture.create_fork(repo_to_fork, self.repo_name) | ||||
self._cleanup_repos.append(self.repo_name) | ||||
return repo | ||||
def new_repo_name(self, suffix=""): | ||||
self.repo_name = self._next_repo_name() + suffix | ||||
self._cleanup_repos.append(self.repo_name) | ||||
return self.repo_name | ||||
def _next_repo_name(self): | ||||
return "%s_%s" % (self.invalid_repo_name.sub("_", self._test_name), len(self._cleanup_repos)) | ||||
def ensure_file(self, filename, content=b"Test content\n"): | ||||
assert self._cleanup_repos, "Avoid writing into vcs_test repos" | ||||
commits = [ | ||||
{ | ||||
"added": [ | ||||
FileNode(filename, content=content), | ||||
] | ||||
}, | ||||
] | ||||
self._add_commits_to_repo(self.repo.scm_instance(), commits) | ||||
def enable_downloads(self): | ||||
repo = self.repo | ||||
repo.enable_downloads = True | ||||
Session().add(repo) | ||||
Session().commit() | ||||
def cleanup(self): | ||||
for repo_name in reversed(self._cleanup_repos): | ||||
self._fixture.destroy_repo(repo_name) | ||||
def _add_commits_to_repo(self, repo, commits): | ||||
commit_ids = _add_commits_to_repo(repo, commits) | ||||
if not commit_ids: | ||||
return | ||||
self._commit_ids = commit_ids | ||||
# Creating refs for Git to allow fetching them from remote repository | ||||
if self.alias == "git": | ||||
refs = {} | ||||
for message in self._commit_ids: | ||||
cleanup_message = message.replace(" ", "") | ||||
ref_name = f"refs/test-refs/{cleanup_message}" | ||||
refs[ref_name] = self._commit_ids[message] | ||||
self._create_refs(repo, refs) | ||||
def _create_refs(self, repo, refs): | ||||
for ref_name, ref_val in refs.items(): | ||||
repo.set_refs(ref_name, ref_val) | ||||
class VcsBackend(object): | ||||
""" | ||||
Represents the test configuration for one supported vcs backend. | ||||
""" | ||||
invalid_repo_name = re.compile(r"[^0-9a-zA-Z]+") | ||||
def __init__(self, alias, repo_path, test_name, test_repo_container): | ||||
self.alias = alias | ||||
self._repo_path = repo_path | ||||
self._cleanup_repos = [] | ||||
self._test_name = test_name | ||||
self._test_repo_container = test_repo_container | ||||
def __getitem__(self, key): | ||||
return self._test_repo_container(key, self.alias).scm_instance() | ||||
def __repr__(self): | ||||
return f"{self.__class__.__name__}(alias={self.alias}, repo={self._repo_path})" | ||||
@property | ||||
def repo(self): | ||||
""" | ||||
Returns the "current" repository. This is the vcs_test repo of the last | ||||
repo which has been created. | ||||
""" | ||||
Repository = get_backend(self.alias) | ||||
return Repository(self._repo_path) | ||||
@property | ||||
def backend(self): | ||||
""" | ||||
Returns the backend implementation class. | ||||
""" | ||||
return get_backend(self.alias) | ||||
def create_repo(self, commits=None, number_of_commits=0, _clone_repo=None, bare=False): | ||||
repo_name = self._next_repo_name() | ||||
self._repo_path = get_new_dir(repo_name) | ||||
repo_class = get_backend(self.alias) | ||||
src_url = None | ||||
if _clone_repo: | ||||
src_url = _clone_repo.path | ||||
repo = repo_class(self._repo_path, create=True, src_url=src_url, bare=bare) | ||||
self._cleanup_repos.append(repo) | ||||
commits = commits or [{"message": "Commit %s of %s" % (x, repo_name)} for x in range(number_of_commits)] | ||||
_add_commits_to_repo(repo, commits) | ||||
return repo | ||||
def clone_repo(self, repo): | ||||
return self.create_repo(_clone_repo=repo) | ||||
def cleanup(self): | ||||
for repo in self._cleanup_repos: | ||||
shutil.rmtree(repo.path) | ||||
def new_repo_path(self): | ||||
repo_name = self._next_repo_name() | ||||
self._repo_path = get_new_dir(repo_name) | ||||
return self._repo_path | ||||
def _next_repo_name(self): | ||||
return "{}_{}".format(self.invalid_repo_name.sub("_", self._test_name), len(self._cleanup_repos)) | ||||
def add_file(self, repo, filename, content="Test content\n"): | ||||
imc = repo.in_memory_commit | ||||
imc.add(FileNode(safe_bytes(filename), content=safe_bytes(content))) | ||||
imc.commit(message="Automatic commit from vcsbackend fixture", author="Automatic <automatic@rhodecode.com>") | ||||
def ensure_file(self, filename, content="Test content\n"): | ||||
assert self._cleanup_repos, "Avoid writing into vcs_test repos" | ||||
self.add_file(self.repo, filename, content) | ||||
def vcsbackend_base(request, backend_alias, tests_tmp_path, baseapp, test_repo) -> VcsBackend: | ||||
if backend_alias not in request.config.getoption("--backends"): | ||||
pytest.skip("Backend %s not selected." % (backend_alias,)) | ||||
utils.check_xfail_backends(request.node, backend_alias) | ||||
utils.check_skip_backends(request.node, backend_alias) | ||||
repo_name = f"vcs_test_{backend_alias}" | ||||
repo_path = os.path.join(tests_tmp_path, repo_name) | ||||
backend = VcsBackend( | ||||
alias=backend_alias, repo_path=repo_path, test_name=request.node.name, test_repo_container=test_repo | ||||
) | ||||
request.addfinalizer(backend.cleanup) | ||||
return backend | ||||
@pytest.fixture() | ||||
def vcsbackend(request, backend_alias, tests_tmp_path, baseapp, test_repo): | ||||
""" | ||||
Parametrized fixture which represents a single vcs backend implementation. | ||||
See the fixture `backend` for more details. This one implements the same | ||||
concept, but on vcs level. So it does not provide model instances etc. | ||||
Parameters are generated dynamically, see :func:`pytest_generate_tests` | ||||
for how this works. | ||||
""" | ||||
return vcsbackend_base(request, backend_alias, tests_tmp_path, baseapp, test_repo) | ||||
@pytest.fixture() | ||||
def vcsbackend_git(request, tests_tmp_path, baseapp, test_repo): | ||||
return vcsbackend_base(request, "git", tests_tmp_path, baseapp, test_repo) | ||||
@pytest.fixture() | ||||
def vcsbackend_hg(request, tests_tmp_path, baseapp, test_repo): | ||||
return vcsbackend_base(request, "hg", tests_tmp_path, baseapp, test_repo) | ||||
@pytest.fixture() | ||||
def vcsbackend_svn(request, tests_tmp_path, baseapp, test_repo): | ||||
return vcsbackend_base(request, "svn", tests_tmp_path, baseapp, test_repo) | ||||
@pytest.fixture() | ||||
def vcsbackend_stub(vcsbackend_git): | ||||
""" | ||||
Use this to express that your test just needs a stub of a vcsbackend. | ||||
Plan is to eventually implement an in-memory stub to speed tests up. | ||||
""" | ||||
return vcsbackend_git | ||||
def _add_commits_to_repo(vcs_repo, commits): | ||||
commit_ids = {} | ||||
if not commits: | ||||
return commit_ids | ||||
imc = vcs_repo.in_memory_commit | ||||
for idx, commit in enumerate(commits): | ||||
message = str(commit.get("message", f"Commit {idx}")) | ||||
for node in commit.get("added", []): | ||||
imc.add(FileNode(safe_bytes(node.path), content=node.content)) | ||||
for node in commit.get("changed", []): | ||||
imc.change(FileNode(safe_bytes(node.path), content=node.content)) | ||||
for node in commit.get("removed", []): | ||||
imc.remove(FileNode(safe_bytes(node.path))) | ||||
parents = [vcs_repo.get_commit(commit_id=commit_ids[p]) for p in commit.get("parents", [])] | ||||
operations = ("added", "changed", "removed") | ||||
if not any((commit.get(o) for o in operations)): | ||||
imc.add(FileNode(b"file_%b" % safe_bytes(str(idx)), content=safe_bytes(message))) | ||||
commit = imc.commit( | ||||
message=message, | ||||
author=str(commit.get("author", "Automatic <automatic@rhodecode.com>")), | ||||
date=commit.get("date"), | ||||
branch=commit.get("branch"), | ||||
parents=parents, | ||||
) | ||||
commit_ids[commit.message] = commit.raw_id | ||||
return commit_ids | ||||
@pytest.fixture() | ||||
def reposerver(request): | ||||
""" | ||||
Allows to serve a backend repository | ||||
""" | ||||
repo_server = RepoServer() | ||||
request.addfinalizer(repo_server.cleanup) | ||||
return repo_server | ||||
class RepoServer(object): | ||||
""" | ||||
Utility to serve a local repository for the duration of a test case. | ||||
Supports only Subversion so far. | ||||
""" | ||||
url = None | ||||
def __init__(self): | ||||
self._cleanup_servers = [] | ||||
def serve(self, vcsrepo): | ||||
if vcsrepo.alias != "svn": | ||||
raise TypeError("Backend %s not supported" % vcsrepo.alias) | ||||
proc = subprocess.Popen( | ||||
["svnserve", "-d", "--foreground", "--listen-host", "localhost", "--root", vcsrepo.path] | ||||
) | ||||
self._cleanup_servers.append(proc) | ||||
self.url = "svn://localhost" | ||||
def cleanup(self): | ||||
for proc in self._cleanup_servers: | ||||
proc.terminate() | ||||
@pytest.fixture() | ||||
def pr_util(backend, request, config_stub): | ||||
""" | ||||
Utility for tests of models and for functional tests around pull requests. | ||||
It gives an instance of :class:`PRTestUtility` which provides various | ||||
utility methods around one pull request. | ||||
This fixture uses `backend` and inherits its parameterization. | ||||
""" | ||||
util = PRTestUtility(backend) | ||||
request.addfinalizer(util.cleanup) | ||||
return util | ||||
class PRTestUtility(object): | ||||
pull_request = None | ||||
pull_request_id = None | ||||
mergeable_patcher = None | ||||
mergeable_mock = None | ||||
notification_patcher = None | ||||
commit_ids: dict | ||||
def __init__(self, backend): | ||||
self.backend = backend | ||||
def create_pull_request( | ||||
self, | ||||
commits=None, | ||||
target_head=None, | ||||
source_head=None, | ||||
revisions=None, | ||||
approved=False, | ||||
author=None, | ||||
mergeable=False, | ||||
enable_notifications=True, | ||||
name_suffix="", | ||||
reviewers=None, | ||||
observers=None, | ||||
title="Test", | ||||
description="Description", | ||||
): | ||||
self.set_mergeable(mergeable) | ||||
if not enable_notifications: | ||||
# mock notification side effect | ||||
self.notification_patcher = mock.patch("rhodecode.model.notification.NotificationModel.create") | ||||
self.notification_patcher.start() | ||||
if not self.pull_request: | ||||
if not commits: | ||||
commits = [ | ||||
{"message": "c1"}, | ||||
{"message": "c2"}, | ||||
{"message": "c3"}, | ||||
] | ||||
target_head = "c1" | ||||
source_head = "c2" | ||||
revisions = ["c2"] | ||||
self.commit_ids = self.backend.create_master_repo(commits) | ||||
self.target_repository = self.backend.create_repo(heads=[target_head], name_suffix=name_suffix) | ||||
self.source_repository = self.backend.create_repo(heads=[source_head], name_suffix=name_suffix) | ||||
self.author = author or UserModel().get_by_username(TEST_USER_ADMIN_LOGIN) | ||||
model = PullRequestModel() | ||||
self.create_parameters = { | ||||
"created_by": self.author, | ||||
"source_repo": self.source_repository.repo_name, | ||||
"source_ref": self._default_branch_reference(source_head), | ||||
"target_repo": self.target_repository.repo_name, | ||||
"target_ref": self._default_branch_reference(target_head), | ||||
"revisions": [self.commit_ids[r] for r in revisions], | ||||
"reviewers": reviewers or self._get_reviewers(), | ||||
"observers": observers or self._get_observers(), | ||||
"title": title, | ||||
"description": description, | ||||
} | ||||
self.pull_request = model.create(**self.create_parameters) | ||||
assert model.get_versions(self.pull_request) == [] | ||||
self.pull_request_id = self.pull_request.pull_request_id | ||||
if approved: | ||||
self.approve() | ||||
Session().add(self.pull_request) | ||||
Session().commit() | ||||
return self.pull_request | ||||
def approve(self): | ||||
self.create_status_votes(ChangesetStatus.STATUS_APPROVED, *self.pull_request.reviewers) | ||||
def close(self): | ||||
PullRequestModel().close_pull_request(self.pull_request, self.author) | ||||
def _default_branch_reference(self, commit_message, branch: str = None) -> str: | ||||
default_branch = branch or self.backend.default_branch_name | ||||
message = self.commit_ids[commit_message] | ||||
reference = f"branch:{default_branch}:{message}" | ||||
return reference | ||||
def _get_reviewers(self): | ||||
role = PullRequestReviewers.ROLE_REVIEWER | ||||
return [ | ||||
(TEST_USER_REGULAR_LOGIN, ["default1"], False, role, []), | ||||
(TEST_USER_REGULAR2_LOGIN, ["default2"], False, role, []), | ||||
] | ||||
def _get_observers(self): | ||||
return [] | ||||
def update_source_repository(self, head=None, do_fetch=False): | ||||
heads = [head or "c3"] | ||||
self.backend.pull_heads(self.source_repository, heads=heads, do_fetch=do_fetch) | ||||
def update_target_repository(self, head=None, do_fetch=False): | ||||
heads = [head or "c3"] | ||||
self.backend.pull_heads(self.target_repository, heads=heads, do_fetch=do_fetch) | ||||
def set_pr_target_ref(self, ref_type: str = "branch", ref_name: str = "branch", ref_commit_id: str = "") -> str: | ||||
full_ref = f"{ref_type}:{ref_name}:{ref_commit_id}" | ||||
self.pull_request.target_ref = full_ref | ||||
return full_ref | ||||
def set_pr_source_ref(self, ref_type: str = "branch", ref_name: str = "branch", ref_commit_id: str = "") -> str: | ||||
full_ref = f"{ref_type}:{ref_name}:{ref_commit_id}" | ||||
self.pull_request.source_ref = full_ref | ||||
return full_ref | ||||
def add_one_commit(self, head=None): | ||||
self.update_source_repository(head=head) | ||||
old_commit_ids = set(self.pull_request.revisions) | ||||
PullRequestModel().update_commits(self.pull_request, self.pull_request.author) | ||||
commit_ids = set(self.pull_request.revisions) | ||||
new_commit_ids = commit_ids - old_commit_ids | ||||
assert len(new_commit_ids) == 1 | ||||
return new_commit_ids.pop() | ||||
def remove_one_commit(self): | ||||
assert len(self.pull_request.revisions) == 2 | ||||
source_vcs = self.source_repository.scm_instance() | ||||
removed_commit_id = source_vcs.commit_ids[-1] | ||||
# TODO: johbo: Git and Mercurial have an inconsistent vcs api here, | ||||
# remove the if once that's sorted out. | ||||
if self.backend.alias == "git": | ||||
kwargs = {"branch_name": self.backend.default_branch_name} | ||||
else: | ||||
kwargs = {} | ||||
source_vcs.strip(removed_commit_id, **kwargs) | ||||
PullRequestModel().update_commits(self.pull_request, self.pull_request.author) | ||||
assert len(self.pull_request.revisions) == 1 | ||||
return removed_commit_id | ||||
def create_comment(self, linked_to=None): | ||||
comment = CommentsModel().create( | ||||
text="Test comment", repo=self.target_repository.repo_name, user=self.author, pull_request=self.pull_request | ||||
) | ||||
assert comment.pull_request_version_id is None | ||||
if linked_to: | ||||
PullRequestModel()._link_comments_to_version(linked_to) | ||||
return comment | ||||
def create_inline_comment(self, linked_to=None, line_no="n1", file_path="file_1"): | ||||
comment = CommentsModel().create( | ||||
text="Test comment", | ||||
repo=self.target_repository.repo_name, | ||||
user=self.author, | ||||
line_no=line_no, | ||||
f_path=file_path, | ||||
pull_request=self.pull_request, | ||||
) | ||||
assert comment.pull_request_version_id is None | ||||
if linked_to: | ||||
PullRequestModel()._link_comments_to_version(linked_to) | ||||
return comment | ||||
def create_version_of_pull_request(self): | ||||
pull_request = self.create_pull_request() | ||||
version = PullRequestModel()._create_version_from_snapshot(pull_request) | ||||
return version | ||||
def create_status_votes(self, status, *reviewers): | ||||
for reviewer in reviewers: | ||||
ChangesetStatusModel().set_status( | ||||
repo=self.pull_request.target_repo, status=status, user=reviewer.user_id, pull_request=self.pull_request | ||||
) | ||||
def set_mergeable(self, value): | ||||
if not self.mergeable_patcher: | ||||
self.mergeable_patcher = mock.patch.object(VcsSettingsModel, "get_general_settings") | ||||
self.mergeable_mock = self.mergeable_patcher.start() | ||||
self.mergeable_mock.return_value = {"rhodecode_pr_merge_enabled": value} | ||||
def cleanup(self): | ||||
# In case the source repository is already cleaned up, the pull | ||||
# request will already be deleted. | ||||
pull_request = PullRequest().get(self.pull_request_id) | ||||
if pull_request: | ||||
PullRequestModel().delete(pull_request, pull_request.author) | ||||
Session().commit() | ||||
if self.notification_patcher: | ||||
self.notification_patcher.stop() | ||||
if self.mergeable_patcher: | ||||
self.mergeable_patcher.stop() | ||||
@pytest.fixture() | ||||
def user_admin(baseapp): | ||||
""" | ||||
Provides the default admin test user as an instance of `db.User`. | ||||
""" | ||||
user = UserModel().get_by_username(TEST_USER_ADMIN_LOGIN) | ||||
return user | ||||
@pytest.fixture() | ||||
def user_regular(baseapp): | ||||
""" | ||||
Provides the default regular test user as an instance of `db.User`. | ||||
""" | ||||
user = UserModel().get_by_username(TEST_USER_REGULAR_LOGIN) | ||||
return user | ||||
@pytest.fixture() | ||||
def user_util(request, db_connection): | ||||
""" | ||||
Provides a wired instance of `UserUtility` with integrated cleanup. | ||||
""" | ||||
utility = UserUtility(test_name=request.node.name) | ||||
request.addfinalizer(utility.cleanup) | ||||
return utility | ||||
# TODO: johbo: Split this up into utilities per domain or something similar | ||||
class UserUtility(object): | ||||
def __init__(self, test_name="test"): | ||||
self._test_name = self._sanitize_name(test_name) | ||||
self.fixture = Fixture() | ||||
self.repo_group_ids = [] | ||||
self.repos_ids = [] | ||||
self.user_ids = [] | ||||
self.user_group_ids = [] | ||||
self.user_repo_permission_ids = [] | ||||
self.user_group_repo_permission_ids = [] | ||||
self.user_repo_group_permission_ids = [] | ||||
self.user_group_repo_group_permission_ids = [] | ||||
self.user_user_group_permission_ids = [] | ||||
self.user_group_user_group_permission_ids = [] | ||||
self.user_permissions = [] | ||||
def _sanitize_name(self, name): | ||||
for char in ["[", "]"]: | ||||
name = name.replace(char, "_") | ||||
return name | ||||
def create_repo_group(self, owner=TEST_USER_ADMIN_LOGIN, auto_cleanup=True): | ||||
group_name = "{prefix}_repogroup_{count}".format(prefix=self._test_name, count=len(self.repo_group_ids)) | ||||
repo_group = self.fixture.create_repo_group(group_name, cur_user=owner) | ||||
if auto_cleanup: | ||||
self.repo_group_ids.append(repo_group.group_id) | ||||
return repo_group | ||||
def create_repo(self, owner=TEST_USER_ADMIN_LOGIN, parent=None, auto_cleanup=True, repo_type="hg", bare=False): | ||||
repo_name = "{prefix}_repository_{count}".format(prefix=self._test_name, count=len(self.repos_ids)) | ||||
repository = self.fixture.create_repo( | ||||
repo_name, cur_user=owner, repo_group=parent, repo_type=repo_type, bare=bare | ||||
) | ||||
if auto_cleanup: | ||||
self.repos_ids.append(repository.repo_id) | ||||
return repository | ||||
def create_user(self, auto_cleanup=True, **kwargs): | ||||
user_name = "{prefix}_user_{count}".format(prefix=self._test_name, count=len(self.user_ids)) | ||||
user = self.fixture.create_user(user_name, **kwargs) | ||||
if auto_cleanup: | ||||
self.user_ids.append(user.user_id) | ||||
return user | ||||
def create_additional_user_email(self, user, email): | ||||
uem = self.fixture.create_additional_user_email(user=user, email=email) | ||||
return uem | ||||
def create_user_with_group(self): | ||||
user = self.create_user() | ||||
user_group = self.create_user_group(members=[user]) | ||||
return user, user_group | ||||
def create_user_group(self, owner=TEST_USER_ADMIN_LOGIN, members=None, auto_cleanup=True, **kwargs): | ||||
group_name = "{prefix}_usergroup_{count}".format(prefix=self._test_name, count=len(self.user_group_ids)) | ||||
user_group = self.fixture.create_user_group(group_name, cur_user=owner, **kwargs) | ||||
if auto_cleanup: | ||||
self.user_group_ids.append(user_group.users_group_id) | ||||
if members: | ||||
for user in members: | ||||
UserGroupModel().add_user_to_group(user_group, user) | ||||
return user_group | ||||
def grant_user_permission(self, user_name, permission_name): | ||||
self.inherit_default_user_permissions(user_name, False) | ||||
self.user_permissions.append((user_name, permission_name)) | ||||
def grant_user_permission_to_repo_group(self, repo_group, user, permission_name): | ||||
permission = RepoGroupModel().grant_user_permission(repo_group, user, permission_name) | ||||
self.user_repo_group_permission_ids.append((repo_group.group_id, user.user_id)) | ||||
return permission | ||||
def grant_user_group_permission_to_repo_group(self, repo_group, user_group, permission_name): | ||||
permission = RepoGroupModel().grant_user_group_permission(repo_group, user_group, permission_name) | ||||
self.user_group_repo_group_permission_ids.append((repo_group.group_id, user_group.users_group_id)) | ||||
return permission | ||||
def grant_user_permission_to_repo(self, repo, user, permission_name): | ||||
permission = RepoModel().grant_user_permission(repo, user, permission_name) | ||||
self.user_repo_permission_ids.append((repo.repo_id, user.user_id)) | ||||
return permission | ||||
def grant_user_group_permission_to_repo(self, repo, user_group, permission_name): | ||||
permission = RepoModel().grant_user_group_permission(repo, user_group, permission_name) | ||||
self.user_group_repo_permission_ids.append((repo.repo_id, user_group.users_group_id)) | ||||
return permission | ||||
def grant_user_permission_to_user_group(self, target_user_group, user, permission_name): | ||||
permission = UserGroupModel().grant_user_permission(target_user_group, user, permission_name) | ||||
self.user_user_group_permission_ids.append((target_user_group.users_group_id, user.user_id)) | ||||
return permission | ||||
def grant_user_group_permission_to_user_group(self, target_user_group, user_group, permission_name): | ||||
permission = UserGroupModel().grant_user_group_permission(target_user_group, user_group, permission_name) | ||||
self.user_group_user_group_permission_ids.append((target_user_group.users_group_id, user_group.users_group_id)) | ||||
return permission | ||||
def revoke_user_permission(self, user_name, permission_name): | ||||
self.inherit_default_user_permissions(user_name, True) | ||||
UserModel().revoke_perm(user_name, permission_name) | ||||
def inherit_default_user_permissions(self, user_name, value): | ||||
user = UserModel().get_by_username(user_name) | ||||
user.inherit_default_permissions = value | ||||
Session().add(user) | ||||
Session().commit() | ||||
def cleanup(self): | ||||
self._cleanup_permissions() | ||||
self._cleanup_repos() | ||||
self._cleanup_repo_groups() | ||||
self._cleanup_user_groups() | ||||
self._cleanup_users() | ||||
def _cleanup_permissions(self): | ||||
if self.user_permissions: | ||||
for user_name, permission_name in self.user_permissions: | ||||
self.revoke_user_permission(user_name, permission_name) | ||||
for permission in self.user_repo_permission_ids: | ||||
RepoModel().revoke_user_permission(*permission) | ||||
for permission in self.user_group_repo_permission_ids: | ||||
RepoModel().revoke_user_group_permission(*permission) | ||||
for permission in self.user_repo_group_permission_ids: | ||||
RepoGroupModel().revoke_user_permission(*permission) | ||||
for permission in self.user_group_repo_group_permission_ids: | ||||
RepoGroupModel().revoke_user_group_permission(*permission) | ||||
for permission in self.user_user_group_permission_ids: | ||||
UserGroupModel().revoke_user_permission(*permission) | ||||
for permission in self.user_group_user_group_permission_ids: | ||||
UserGroupModel().revoke_user_group_permission(*permission) | ||||
def _cleanup_repo_groups(self): | ||||
def _repo_group_compare(first_group_id, second_group_id): | ||||
""" | ||||
Gives higher priority to the groups with the most complex paths | ||||
""" | ||||
first_group = RepoGroup.get(first_group_id) | ||||
second_group = RepoGroup.get(second_group_id) | ||||
first_group_parts = len(first_group.group_name.split("/")) if first_group else 0 | ||||
second_group_parts = len(second_group.group_name.split("/")) if second_group else 0 | ||||
return cmp(second_group_parts, first_group_parts) | ||||
sorted_repo_group_ids = sorted(self.repo_group_ids, key=functools.cmp_to_key(_repo_group_compare)) | ||||
for repo_group_id in sorted_repo_group_ids: | ||||
self.fixture.destroy_repo_group(repo_group_id) | ||||
def _cleanup_repos(self): | ||||
sorted_repos_ids = sorted(self.repos_ids) | ||||
for repo_id in sorted_repos_ids: | ||||
self.fixture.destroy_repo(repo_id) | ||||
def _cleanup_user_groups(self): | ||||
def _user_group_compare(first_group_id, second_group_id): | ||||
""" | ||||
Gives higher priority to the groups with the most complex paths | ||||
""" | ||||
first_group = UserGroup.get(first_group_id) | ||||
second_group = UserGroup.get(second_group_id) | ||||
first_group_parts = len(first_group.users_group_name.split("/")) if first_group else 0 | ||||
second_group_parts = len(second_group.users_group_name.split("/")) if second_group else 0 | ||||
return cmp(second_group_parts, first_group_parts) | ||||
sorted_user_group_ids = sorted(self.user_group_ids, key=functools.cmp_to_key(_user_group_compare)) | ||||
for user_group_id in sorted_user_group_ids: | ||||
self.fixture.destroy_user_group(user_group_id) | ||||
def _cleanup_users(self): | ||||
for user_id in self.user_ids: | ||||
self.fixture.destroy_user(user_id) | ||||
@pytest.fixture(scope="session") | ||||
def testrun(): | ||||
return { | ||||
"uuid": uuid.uuid4(), | ||||
"start": datetime.datetime.utcnow().isoformat(), | ||||
"timestamp": int(time.time()), | ||||
} | ||||
class AppenlightClient(object): | ||||
url_template = "{url}?protocol_version=0.5" | ||||
def __init__(self, url, api_key, add_server=True, add_timestamp=True, namespace=None, request=None, testrun=None): | ||||
self.url = self.url_template.format(url=url) | ||||
self.api_key = api_key | ||||
self.add_server = add_server | ||||
self.add_timestamp = add_timestamp | ||||
self.namespace = namespace | ||||
self.request = request | ||||
self.server = socket.getfqdn(socket.gethostname()) | ||||
self.tags_before = {} | ||||
self.tags_after = {} | ||||
self.stats = [] | ||||
self.testrun = testrun or {} | ||||
def tag_before(self, tag, value): | ||||
self.tags_before[tag] = value | ||||
def tag_after(self, tag, value): | ||||
self.tags_after[tag] = value | ||||
def collect(self, data): | ||||
if self.add_server: | ||||
data.setdefault("server", self.server) | ||||
if self.add_timestamp: | ||||
data.setdefault("date", datetime.datetime.utcnow().isoformat()) | ||||
if self.namespace: | ||||
data.setdefault("namespace", self.namespace) | ||||
if self.request: | ||||
data.setdefault("request", self.request) | ||||
self.stats.append(data) | ||||
def send_stats(self): | ||||
tags = [ | ||||
("testrun", self.request), | ||||
("testrun.start", self.testrun["start"]), | ||||
("testrun.timestamp", self.testrun["timestamp"]), | ||||
("test", self.namespace), | ||||
] | ||||
for key, value in self.tags_before.items(): | ||||
tags.append((key + ".before", value)) | ||||
try: | ||||
delta = self.tags_after[key] - value | ||||
tags.append((key + ".delta", delta)) | ||||
except Exception: | ||||
pass | ||||
for key, value in self.tags_after.items(): | ||||
tags.append((key + ".after", value)) | ||||
self.collect( | ||||
{ | ||||
"message": "Collected tags", | ||||
"tags": tags, | ||||
} | ||||
) | ||||
response = requests.post( | ||||
self.url, | ||||
headers={"X-appenlight-api-key": self.api_key}, | ||||
json=self.stats, | ||||
) | ||||
if not response.status_code == 200: | ||||
pprint.pprint(self.stats) | ||||
print(response.headers) | ||||
print(response.text) | ||||
raise Exception("Sending to appenlight failed") | ||||
@pytest.fixture() | ||||
def gist_util(request, db_connection): | ||||
""" | ||||
Provides a wired instance of `GistUtility` with integrated cleanup. | ||||
""" | ||||
utility = GistUtility() | ||||
request.addfinalizer(utility.cleanup) | ||||
return utility | ||||
class GistUtility(object): | ||||
def __init__(self): | ||||
self.fixture = Fixture() | ||||
self.gist_ids = [] | ||||
def create_gist(self, **kwargs): | ||||
gist = self.fixture.create_gist(**kwargs) | ||||
self.gist_ids.append(gist.gist_id) | ||||
return gist | ||||
def cleanup(self): | ||||
for id_ in self.gist_ids: | ||||
self.fixture.destroy_gists(str(id_)) | ||||
@pytest.fixture() | ||||
def enabled_backends(request): | ||||
backends = request.config.option.backends | ||||
return backends[:] | ||||
@pytest.fixture() | ||||
def settings_util(request, db_connection): | ||||
""" | ||||
Provides a wired instance of `SettingsUtility` with integrated cleanup. | ||||
""" | ||||
utility = SettingsUtility() | ||||
request.addfinalizer(utility.cleanup) | ||||
return utility | ||||
class SettingsUtility(object): | ||||
def __init__(self): | ||||
self.rhodecode_ui_ids = [] | ||||
self.rhodecode_setting_ids = [] | ||||
self.repo_rhodecode_ui_ids = [] | ||||
self.repo_rhodecode_setting_ids = [] | ||||
def create_repo_rhodecode_ui(self, repo, section, value, key=None, active=True, cleanup=True): | ||||
key = key or sha1_safe(f"{section}{value}{repo.repo_id}") | ||||
setting = RepoRhodeCodeUi() | ||||
setting.repository_id = repo.repo_id | ||||
setting.ui_section = section | ||||
setting.ui_value = value | ||||
setting.ui_key = key | ||||
setting.ui_active = active | ||||
Session().add(setting) | ||||
Session().commit() | ||||
if cleanup: | ||||
self.repo_rhodecode_ui_ids.append(setting.ui_id) | ||||
return setting | ||||
def create_rhodecode_ui(self, section, value, key=None, active=True, cleanup=True): | ||||
key = key or sha1_safe(f"{section}{value}") | ||||
setting = RhodeCodeUi() | ||||
setting.ui_section = section | ||||
setting.ui_value = value | ||||
setting.ui_key = key | ||||
setting.ui_active = active | ||||
Session().add(setting) | ||||
Session().commit() | ||||
if cleanup: | ||||
self.rhodecode_ui_ids.append(setting.ui_id) | ||||
return setting | ||||
def create_repo_rhodecode_setting(self, repo, name, value, type_, cleanup=True): | ||||
setting = RepoRhodeCodeSetting(repo.repo_id, key=name, val=value, type=type_) | ||||
Session().add(setting) | ||||
Session().commit() | ||||
if cleanup: | ||||
self.repo_rhodecode_setting_ids.append(setting.app_settings_id) | ||||
return setting | ||||
def create_rhodecode_setting(self, name, value, type_, cleanup=True): | ||||
setting = RhodeCodeSetting(key=name, val=value, type=type_) | ||||
Session().add(setting) | ||||
Session().commit() | ||||
if cleanup: | ||||
self.rhodecode_setting_ids.append(setting.app_settings_id) | ||||
return setting | ||||
def cleanup(self): | ||||
for id_ in self.rhodecode_ui_ids: | ||||
setting = RhodeCodeUi.get(id_) | ||||
Session().delete(setting) | ||||
for id_ in self.rhodecode_setting_ids: | ||||
setting = RhodeCodeSetting.get(id_) | ||||
Session().delete(setting) | ||||
for id_ in self.repo_rhodecode_ui_ids: | ||||
setting = RepoRhodeCodeUi.get(id_) | ||||
Session().delete(setting) | ||||
for id_ in self.repo_rhodecode_setting_ids: | ||||
setting = RepoRhodeCodeSetting.get(id_) | ||||
Session().delete(setting) | ||||
Session().commit() | ||||
@pytest.fixture() | ||||
def no_notifications(request): | ||||
notification_patcher = mock.patch("rhodecode.model.notification.NotificationModel.create") | ||||
notification_patcher.start() | ||||
request.addfinalizer(notification_patcher.stop) | ||||
@pytest.fixture(scope="session") | ||||
def repeat(request): | ||||
""" | ||||
The number of repetitions is based on this fixture. | ||||
Slower calls may divide it by 10 or 100. It is chosen in a way so that the | ||||
tests are not too slow in our default test suite. | ||||
""" | ||||
return request.config.getoption("--repeat") | ||||
@pytest.fixture() | ||||
def rhodecode_fixtures(): | ||||
return Fixture() | ||||
@pytest.fixture() | ||||
def context_stub(): | ||||
""" | ||||
Stub context object. | ||||
""" | ||||
context = pyramid.testing.DummyResource() | ||||
return context | ||||
@pytest.fixture() | ||||
def StubIntegrationType(): | ||||
class _StubIntegrationType(IntegrationTypeBase): | ||||
"""Test integration type class""" | ||||
key = "test" | ||||
display_name = "Test integration type" | ||||
description = "A test integration type for testing" | ||||
@classmethod | ||||
def icon(cls): | ||||
return "test_icon_html_image" | ||||
def __init__(self, settings): | ||||
super(_StubIntegrationType, self).__init__(settings) | ||||
self.sent_events = [] # for testing | ||||
def send_event(self, event): | ||||
self.sent_events.append(event) | ||||
def settings_schema(self): | ||||
class SettingsSchema(colander.Schema): | ||||
test_string_field = colander.SchemaNode( | ||||
colander.String(), | ||||
missing=colander.required, | ||||
title="test string field", | ||||
) | ||||
test_int_field = colander.SchemaNode( | ||||
colander.Int(), | ||||
title="some integer setting", | ||||
) | ||||
return SettingsSchema() | ||||
integration_type_registry.register_integration_type(_StubIntegrationType) | ||||
return _StubIntegrationType | ||||
@pytest.fixture() | ||||
def stub_integration_settings(): | ||||
return { | ||||
"test_string_field": "some data", | ||||
"test_int_field": 100, | ||||
} | ||||
@pytest.fixture() | ||||
def repo_integration_stub(request, repo_stub, StubIntegrationType, stub_integration_settings): | ||||
r5619 | repo_id = repo_stub.repo_id | |||
r5607 | integration = IntegrationModel().create( | |||
StubIntegrationType, | ||||
settings=stub_integration_settings, | ||||
enabled=True, | ||||
name="test repo integration", | ||||
repo=repo_stub, | ||||
repo_group=None, | ||||
child_repos_only=None, | ||||
) | ||||
@request.addfinalizer | ||||
def cleanup(): | ||||
IntegrationModel().delete(integration) | ||||
r5619 | RepoModel().delete(repo_id) | |||
r5607 | ||||
return integration | ||||
@pytest.fixture() | ||||
def repogroup_integration_stub(request, test_repo_group, StubIntegrationType, stub_integration_settings): | ||||
integration = IntegrationModel().create( | ||||
StubIntegrationType, | ||||
settings=stub_integration_settings, | ||||
enabled=True, | ||||
name="test repogroup integration", | ||||
repo=None, | ||||
repo_group=test_repo_group, | ||||
child_repos_only=True, | ||||
) | ||||
@request.addfinalizer | ||||
def cleanup(): | ||||
IntegrationModel().delete(integration) | ||||
return integration | ||||
@pytest.fixture() | ||||
def repogroup_recursive_integration_stub(request, test_repo_group, StubIntegrationType, stub_integration_settings): | ||||
integration = IntegrationModel().create( | ||||
StubIntegrationType, | ||||
settings=stub_integration_settings, | ||||
enabled=True, | ||||
name="test recursive repogroup integration", | ||||
repo=None, | ||||
repo_group=test_repo_group, | ||||
child_repos_only=False, | ||||
) | ||||
@request.addfinalizer | ||||
def cleanup(): | ||||
IntegrationModel().delete(integration) | ||||
return integration | ||||
@pytest.fixture() | ||||
def global_integration_stub(request, StubIntegrationType, stub_integration_settings): | ||||
integration = IntegrationModel().create( | ||||
StubIntegrationType, | ||||
settings=stub_integration_settings, | ||||
enabled=True, | ||||
name="test global integration", | ||||
repo=None, | ||||
repo_group=None, | ||||
child_repos_only=None, | ||||
) | ||||
@request.addfinalizer | ||||
def cleanup(): | ||||
IntegrationModel().delete(integration) | ||||
return integration | ||||
@pytest.fixture() | ||||
def root_repos_integration_stub(request, StubIntegrationType, stub_integration_settings): | ||||
integration = IntegrationModel().create( | ||||
StubIntegrationType, | ||||
settings=stub_integration_settings, | ||||
enabled=True, | ||||
name="test global integration", | ||||
repo=None, | ||||
repo_group=None, | ||||
child_repos_only=True, | ||||
) | ||||
@request.addfinalizer | ||||
def cleanup(): | ||||
IntegrationModel().delete(integration) | ||||
return integration | ||||
@pytest.fixture() | ||||
def local_dt_to_utc(): | ||||
def _factory(dt): | ||||
return dt.replace(tzinfo=dateutil.tz.tzlocal()).astimezone(dateutil.tz.tzutc()).replace(tzinfo=None) | ||||
return _factory | ||||
@pytest.fixture() | ||||
def disable_anonymous_user(request, baseapp): | ||||
set_anonymous_access(False) | ||||
@request.addfinalizer | ||||
def cleanup(): | ||||
set_anonymous_access(True) | ||||
@pytest.fixture(scope="module") | ||||
def rc_fixture(request): | ||||
return Fixture() | ||||
@pytest.fixture() | ||||
def repo_groups(request): | ||||
fixture = Fixture() | ||||
session = Session() | ||||
zombie_group = fixture.create_repo_group("zombie") | ||||
parent_group = fixture.create_repo_group("parent") | ||||
child_group = fixture.create_repo_group("parent/child") | ||||
groups_in_db = session.query(RepoGroup).all() | ||||
assert len(groups_in_db) == 3 | ||||
assert child_group.group_parent_id == parent_group.group_id | ||||
@request.addfinalizer | ||||
def cleanup(): | ||||
fixture.destroy_repo_group(zombie_group) | ||||
fixture.destroy_repo_group(child_group) | ||||
fixture.destroy_repo_group(parent_group) | ||||
return zombie_group, parent_group, child_group | ||||