scm.py
396 lines
| 13.5 KiB
| text/x-python
|
PythonLexer
r757 | # -*- coding: utf-8 -*- | |||
""" | ||||
r811 | rhodecode.model.scm | |||
~~~~~~~~~~~~~~~~~~~ | ||||
r757 | ||||
r811 | Scm model for RhodeCode | |||
r757 | :created_on: Apr 9, 2010 | |||
:author: marcink | ||||
r902 | :copyright: (C) 2009-2011 Marcin Kuzminski <marcin@python-works.com> | |||
r757 | :license: GPLv3, see COPYING for more details. | |||
""" | ||||
r691 | # This program is free software; you can redistribute it and/or | |||
# modify it under the terms of the GNU General Public License | ||||
# as published by the Free Software Foundation; version 2 | ||||
# of the License or (at your opinion) any later version of the license. | ||||
# | ||||
# This program is distributed in the hope that it will be useful, | ||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of | ||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||||
# GNU General Public License for more details. | ||||
# | ||||
# You should have received a copy of the GNU General Public License | ||||
# along with this program; if not, write to the Free Software | ||||
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, | ||||
# MA 02110-1301, USA. | ||||
r757 | import os | |||
import time | ||||
import traceback | ||||
import logging | ||||
r1022 | from mercurial import ui | |||
from sqlalchemy.orm import joinedload | ||||
from sqlalchemy.orm.session import make_transient | ||||
from sqlalchemy.exc import DatabaseError | ||||
from beaker.cache import cache_region, region_invalidate | ||||
r757 | from vcs import get_backend | |||
from vcs.utils.helpers import get_scm | ||||
from vcs.exceptions import RepositoryError, VCSError | ||||
from vcs.utils.lazy import LazyProperty | ||||
r710 | from rhodecode import BACKENDS | |||
r691 | from rhodecode.lib import helpers as h | |||
from rhodecode.lib.auth import HasRepoPermissionAny | ||||
r877 | from rhodecode.lib.utils import get_repos as get_filesystem_repos, make_ui, action_logger | |||
r752 | from rhodecode.model import BaseModel | |||
r758 | from rhodecode.model.user import UserModel | |||
from rhodecode.model.db import Repository, RhodeCodeUi, CacheInvalidation, \ | ||||
r784 | UserFollowing, UserLog | |||
r692 | from rhodecode.model.caching_query import FromCache | |||
r757 | ||||
r691 | log = logging.getLogger(__name__) | |||
r757 | ||||
r735 | class UserTemp(object): | |||
def __init__(self, user_id): | ||||
self.user_id = user_id | ||||
r901 | ||||
def __repr__(self): | ||||
return "<%s('id:%s')>" % (self.__class__.__name__, self.user_id) | ||||
r735 | class RepoTemp(object): | |||
def __init__(self, repo_id): | ||||
self.repo_id = repo_id | ||||
r747 | ||||
r901 | def __repr__(self): | |||
return "<%s('id:%s')>" % (self.__class__.__name__, self.repo_id) | ||||
r752 | class ScmModel(BaseModel): | |||
r811 | """Generic Scm Model | |||
r691 | """ | |||
@LazyProperty | ||||
def repos_path(self): | ||||
r811 | """Get's the repositories root path from database | |||
r691 | """ | |||
r811 | ||||
r691 | q = self.sa.query(RhodeCodeUi).filter(RhodeCodeUi.ui_key == '/').one() | |||
return q.ui_value | ||||
r757 | def repo_scan(self, repos_path, baseui): | |||
r811 | """Listing of repositories in given path. This path should not be a | |||
r691 | repository itself. Return a dictionary of repository objects | |||
:param repos_path: path to directory containing repositories | ||||
r851 | :param baseui: baseui instance to instantiate MercurialRepostitory with | |||
r691 | """ | |||
r811 | ||||
r691 | log.info('scanning for repositories in %s', repos_path) | |||
if not isinstance(baseui, ui.ui): | ||||
r724 | baseui = make_ui('db') | |||
r691 | repos_list = {} | |||
r877 | for name, path in get_filesystem_repos(repos_path, recursive=True): | |||
r691 | try: | |||
if repos_list.has_key(name): | ||||
raise RepositoryError('Duplicate repository name %s ' | ||||
'found in %s' % (name, path)) | ||||
else: | ||||
klass = get_backend(path[0]) | ||||
r710 | if path[0] == 'hg' and path[0] in BACKENDS.keys(): | |||
r691 | repos_list[name] = klass(path[1], baseui=baseui) | |||
r710 | if path[0] == 'git' and path[0] in BACKENDS.keys(): | |||
r691 | repos_list[name] = klass(path[1]) | |||
except OSError: | ||||
continue | ||||
return repos_list | ||||
def get_repos(self, all_repos=None): | ||||
r811 | """Get all repos from db and for each repo create it's backend instance. | |||
r691 | and fill that backed with information from database | |||
:param all_repos: give specific repositories list, good for filtering | ||||
""" | ||||
r811 | ||||
r767 | if all_repos is None: | |||
r693 | all_repos = self.sa.query(Repository)\ | |||
.order_by(Repository.repo_name).all() | ||||
r691 | ||||
r791 | #get the repositories that should be invalidated | |||
r726 | invalidation_list = [str(x.cache_key) for x in \ | |||
self.sa.query(CacheInvalidation.cache_key)\ | ||||
.filter(CacheInvalidation.cache_active == False)\ | ||||
.all()] | ||||
r691 | for r in all_repos: | |||
r726 | repo = self.get(r.repo_name, invalidation_list) | |||
r691 | ||||
if repo is not None: | ||||
last_change = repo.last_change | ||||
tip = h.get_changeset_safe(repo, 'tip') | ||||
tmp_d = {} | ||||
r878 | tmp_d['name'] = r.repo_name | |||
r691 | tmp_d['name_sort'] = tmp_d['name'].lower() | |||
tmp_d['description'] = repo.dbrepo.description | ||||
tmp_d['description_sort'] = tmp_d['description'] | ||||
tmp_d['last_change'] = last_change | ||||
tmp_d['last_change_sort'] = time.mktime(last_change.timetuple()) | ||||
tmp_d['tip'] = tip.raw_id | ||||
tmp_d['tip_sort'] = tip.revision | ||||
tmp_d['rev'] = tip.revision | ||||
r828 | tmp_d['contact'] = repo.dbrepo.user.full_contact | |||
r691 | tmp_d['contact_sort'] = tmp_d['contact'] | |||
r1005 | tmp_d['owner_sort'] = tmp_d['contact'] | |||
r691 | tmp_d['repo_archives'] = list(repo._get_archives()) | |||
tmp_d['last_msg'] = tip.message | ||||
tmp_d['repo'] = repo | ||||
yield tmp_d | ||||
def get_repo(self, repo_name): | ||||
return self.get(repo_name) | ||||
r726 | def get(self, repo_name, invalidation_list=None): | |||
r811 | """Get's repository from given name, creates BackendInstance and | |||
r691 | propagates it's data from database with all additional information | |||
r791 | ||||
r691 | :param repo_name: | |||
r791 | :param invalidation_list: if a invalidation list is given the get | |||
method should not manually check if this repository needs | ||||
invalidation and just invalidate the repositories in list | ||||
r691 | """ | |||
if not HasRepoPermissionAny('repository.read', 'repository.write', | ||||
'repository.admin')(repo_name, 'get repo check'): | ||||
return | ||||
r791 | #====================================================================== | |||
# CACHE FUNCTION | ||||
#====================================================================== | ||||
r692 | @cache_region('long_term') | |||
r691 | def _get_repo(repo_name): | |||
repo_path = os.path.join(self.repos_path, repo_name) | ||||
r757 | ||||
try: | ||||
alias = get_scm(repo_path)[0] | ||||
log.debug('Creating instance of %s repository', alias) | ||||
backend = get_backend(alias) | ||||
except VCSError: | ||||
log.error(traceback.format_exc()) | ||||
r1028 | log.error('Perhaps this repository is in db and not in filesystem' | |||
'run rescan repositories with "destroy old data "' | ||||
'option from admin panel') | ||||
r757 | return | |||
r691 | ||||
if alias == 'hg': | ||||
r715 | from pylons import app_globals as g | |||
repo = backend(repo_path, create=False, baseui=g.baseui) | ||||
r691 | #skip hidden web repository | |||
if repo._get_hidden(): | ||||
return | ||||
else: | ||||
repo = backend(repo_path, create=False) | ||||
dbrepo = self.sa.query(Repository)\ | ||||
.options(joinedload(Repository.fork))\ | ||||
.options(joinedload(Repository.user))\ | ||||
.filter(Repository.repo_name == repo_name)\ | ||||
.scalar() | ||||
r753 | ||||
r1033 | self.sa.expunge_all() | |||
log.debug('making transient %s', dbrepo) | ||||
r757 | make_transient(dbrepo) | |||
r1033 | ||||
for attr in ['user', 'forks', 'followers', 'group', 'repo_to_perm', | ||||
'users_group_to_perm', 'stats', 'logs']: | ||||
attr = getattr(dbrepo, attr, False) | ||||
if attr: | ||||
if isinstance(attr, list): | ||||
for a in attr: | ||||
log.debug('making transient %s', a) | ||||
make_transient(a) | ||||
else: | ||||
log.debug('making transient %s', attr) | ||||
make_transient(attr) | ||||
r753 | ||||
r691 | repo.dbrepo = dbrepo | |||
return repo | ||||
r792 | pre_invalidate = True | |||
if invalidation_list is not None: | ||||
pre_invalidate = repo_name in invalidation_list | ||||
if pre_invalidate: | ||||
invalidate = self._should_invalidate(repo_name) | ||||
if invalidate: | ||||
log.info('invalidating cache for repository %s', repo_name) | ||||
region_invalidate(_get_repo, None, repo_name) | ||||
self._mark_invalidated(invalidate) | ||||
r691 | return _get_repo(repo_name) | |||
r692 | ||||
def mark_for_invalidation(self, repo_name): | ||||
r811 | """Puts cache invalidation task into db for | |||
r692 | further global cache invalidation | |||
:param repo_name: this repo that should invalidation take place | ||||
""" | ||||
r811 | ||||
r692 | log.debug('marking %s for invalidation', repo_name) | |||
cache = self.sa.query(CacheInvalidation)\ | ||||
.filter(CacheInvalidation.cache_key == repo_name).scalar() | ||||
if cache: | ||||
#mark this cache as inactive | ||||
cache.cache_active = False | ||||
else: | ||||
log.debug('cache key not found in invalidation db -> creating one') | ||||
cache = CacheInvalidation(repo_name) | ||||
try: | ||||
self.sa.add(cache) | ||||
self.sa.commit() | ||||
r758 | except (DatabaseError,): | |||
r692 | log.error(traceback.format_exc()) | |||
self.sa.rollback() | ||||
r734 | def toggle_following_repo(self, follow_repo_id, user_id): | |||
r692 | ||||
r734 | f = self.sa.query(UserFollowing)\ | |||
.filter(UserFollowing.follows_repo_id == follow_repo_id)\ | ||||
.filter(UserFollowing.user_id == user_id).scalar() | ||||
if f is not None: | ||||
r747 | ||||
r734 | try: | |||
self.sa.delete(f) | ||||
self.sa.commit() | ||||
r735 | action_logger(UserTemp(user_id), | |||
'stopped_following_repo', | ||||
r747 | RepoTemp(follow_repo_id)) | |||
r734 | return | |||
except: | ||||
log.error(traceback.format_exc()) | ||||
self.sa.rollback() | ||||
raise | ||||
try: | ||||
f = UserFollowing() | ||||
f.user_id = user_id | ||||
f.follows_repo_id = follow_repo_id | ||||
self.sa.add(f) | ||||
self.sa.commit() | ||||
r735 | action_logger(UserTemp(user_id), | |||
'started_following_repo', | ||||
r747 | RepoTemp(follow_repo_id)) | |||
r734 | except: | |||
log.error(traceback.format_exc()) | ||||
self.sa.rollback() | ||||
raise | ||||
def toggle_following_user(self, follow_user_id , user_id): | ||||
f = self.sa.query(UserFollowing)\ | ||||
.filter(UserFollowing.follows_user_id == follow_user_id)\ | ||||
.filter(UserFollowing.user_id == user_id).scalar() | ||||
if f is not None: | ||||
try: | ||||
self.sa.delete(f) | ||||
self.sa.commit() | ||||
return | ||||
except: | ||||
log.error(traceback.format_exc()) | ||||
self.sa.rollback() | ||||
raise | ||||
try: | ||||
f = UserFollowing() | ||||
f.user_id = user_id | ||||
f.follows_user_id = follow_user_id | ||||
self.sa.add(f) | ||||
self.sa.commit() | ||||
except: | ||||
log.error(traceback.format_exc()) | ||||
self.sa.rollback() | ||||
raise | ||||
r999 | def is_following_repo(self, repo_name, user_id, cache=False): | |||
r734 | r = self.sa.query(Repository)\ | |||
.filter(Repository.repo_name == repo_name).scalar() | ||||
f = self.sa.query(UserFollowing)\ | ||||
.filter(UserFollowing.follows_repository == r)\ | ||||
.filter(UserFollowing.user_id == user_id).scalar() | ||||
return f is not None | ||||
r999 | def is_following_user(self, username, user_id, cache=False): | |||
r758 | u = UserModel(self.sa).get_by_username(username) | |||
r734 | ||||
f = self.sa.query(UserFollowing)\ | ||||
.filter(UserFollowing.follows_user == u)\ | ||||
.filter(UserFollowing.user_id == user_id).scalar() | ||||
return f is not None | ||||
r692 | ||||
r747 | def get_followers(self, repo_id): | |||
return self.sa.query(UserFollowing)\ | ||||
.filter(UserFollowing.follows_repo_id == repo_id).count() | ||||
def get_forks(self, repo_id): | ||||
return self.sa.query(Repository)\ | ||||
.filter(Repository.fork_id == repo_id).count() | ||||
r692 | ||||
r784 | ||||
def get_unread_journal(self): | ||||
return self.sa.query(UserLog).count() | ||||
r692 | def _should_invalidate(self, repo_name): | |||
r811 | """Looks up database for invalidation signals for this repo_name | |||
r692 | :param repo_name: | |||
""" | ||||
ret = self.sa.query(CacheInvalidation)\ | ||||
.options(FromCache('sql_cache_short', | ||||
'get_invalidation_%s' % repo_name))\ | ||||
.filter(CacheInvalidation.cache_key == repo_name)\ | ||||
.filter(CacheInvalidation.cache_active == False)\ | ||||
.scalar() | ||||
return ret | ||||
def _mark_invalidated(self, cache_key): | ||||
r811 | """ Marks all occurences of cache to invaldation as already invalidated | |||
:param cache_key: | ||||
r692 | """ | |||
r811 | ||||
r692 | if cache_key: | |||
log.debug('marking %s as already invalidated', cache_key) | ||||
try: | ||||
cache_key.cache_active = True | ||||
self.sa.add(cache_key) | ||||
self.sa.commit() | ||||
r758 | except (DatabaseError,): | |||
r692 | log.error(traceback.format_exc()) | |||
self.sa.rollback() | ||||