# Copyright (C) 2010-2024 RhodeCode GmbH # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License, version 3 # (only), as published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see . # # This program is dual-licensed. If you wish to learn more about the # RhodeCode Enterprise Edition, including its added features, Support services, # and proprietary license terms, please see https://rhodecode.com/licenses/ """ Helper functions Consists of functions to typically be used within templates, but also available to Controllers. This module is available to both as 'h'. """ import base64 import collections import os import random import hashlib import io import textwrap import urllib.request import urllib.parse import urllib.error import math import logging import re import time import string import regex from collections import OrderedDict import pygments import itertools import fnmatch from datetime import datetime from functools import partial from pygments.formatters.html import HtmlFormatter from pygments.lexers import ( get_lexer_by_name, get_lexer_for_filename, get_lexer_for_mimetype) from pyramid.threadlocal import get_current_request from tempita import looper from webhelpers2.html import literal, HTML, escape from webhelpers2.html._autolink import _auto_link_urls from webhelpers2.html.tools import ( button_to, highlight, js_obfuscate, strip_links, strip_tags) from webhelpers2.text import ( chop_at, collapse, convert_accented_entities, convert_misc_entities, lchop, plural, rchop, remove_formatting, replace_whitespace, urlify, truncate, wrap_paragraphs) from webhelpers2.date import time_ago_in_words from webhelpers2.html.tags import ( _input, NotGiven, _make_safe_id_component as safeid, form as insecure_form, auto_discovery_link, checkbox, end_form, file, hidden, image, javascript_link, link_to, link_to_if, link_to_unless, ol, stylesheet_link, submit, text, password, textarea, ul, radio, Options) from webhelpers2.number import format_byte_size # python3.11 backport fixes for webhelpers2 from rhodecode import ConfigGet from rhodecode.lib._vendor.webhelpers_backports import raw_select from rhodecode.lib.action_parser import action_parser from rhodecode.lib.html_filters import sanitize_html from rhodecode.lib.pagination import Page, RepoPage, SqlPage from rhodecode.lib import ext_json from rhodecode.lib.ext_json import json, formatted_str_json from rhodecode.lib.str_utils import safe_bytes, convert_special_chars, base64_to_str from rhodecode.lib.utils import repo_name_slug, get_custom_lexer from rhodecode.lib.str_utils import safe_str from rhodecode.lib.utils2 import ( str2bool, get_commit_safe, datetime_to_time, time_to_datetime, time_to_utcdatetime, AttributeDict, safe_int, md5, md5_safe, get_host_info) from rhodecode.lib.markup_renderer import MarkupRenderer, relative_links from rhodecode.lib.vcs.exceptions import CommitDoesNotExistError from rhodecode.lib.vcs.backends.base import BaseChangeset, EmptyCommit from rhodecode.lib.vcs.conf.settings import ARCHIVE_SPECS from rhodecode.lib.index.search_utils import get_matching_line_offsets from rhodecode.config.conf import DATE_FORMAT, DATETIME_FORMAT from rhodecode.model.changeset_status import ChangesetStatusModel from rhodecode.model.db import Permission, User, Repository, UserApiKeys, FileStore from rhodecode.model.repo_group import RepoGroupModel from rhodecode.model.settings import IssueTrackerSettingsModel log = logging.getLogger(__name__) DEFAULT_USER = User.DEFAULT_USER DEFAULT_USER_EMAIL = User.DEFAULT_USER_EMAIL def asset(path, ver=None, **kwargs): """ Helper to generate a static asset file path for rhodecode assets eg. h.asset('images/image.png', ver='3923') :param path: path of asset :param ver: optional version query param to append as ?ver= """ request = get_current_request() query = {} query.update(kwargs) if ver: query = {'ver': ver} return request.static_path( f'rhodecode:public/{path}', _query=query) default_html_escape_table = { ord('&'): '&', ord('<'): '<', ord('>'): '>', ord('"'): '"', ord("'"): ''', } def html_escape(text, html_escape_table=default_html_escape_table): """Produce entities within text.""" return text.translate(html_escape_table) def str_json(*args, **kwargs): return ext_json.str_json(*args, **kwargs) def formatted_str_json(*args, **kwargs): return ext_json.formatted_str_json(*args, **kwargs) def chop_at_smart(s, sub, inclusive=False, suffix_if_chopped=None): """ Truncate string ``s`` at the first occurrence of ``sub``. If ``inclusive`` is true, truncate just after ``sub`` rather than at it. """ suffix_if_chopped = suffix_if_chopped or '' pos = s.find(sub) if pos == -1: return s if inclusive: pos += len(sub) chopped = s[:pos] left = s[pos:].strip() if left and suffix_if_chopped: chopped += suffix_if_chopped return chopped def shorter(text, size=20, prefix=False): postfix = '...' if len(text) > size: if prefix: # shorten in front return postfix + text[-(size - len(postfix)):] else: return text[:size - len(postfix)] + postfix return text def reset(name, value=None, id=NotGiven, type="reset", **attrs): """ Reset button """ return _input(type, name, value, id, attrs) def select(name, selected_values, options, id=NotGiven, **attrs): if isinstance(options, (list, tuple)): options_iter = options # Handle old value,label lists ... where value also can be value,label lists options = Options() for opt in options_iter: if isinstance(opt, tuple) and len(opt) == 2: value, label = opt elif isinstance(opt, str): value = label = opt else: raise ValueError('invalid select option type %r' % type(opt)) if isinstance(value, (list, tuple)): option_group = options.add_optgroup(label) for opt2 in value: if isinstance(opt2, tuple) and len(opt2) == 2: group_value, group_label = opt2 elif isinstance(opt2, str): group_value = group_label = opt2 else: raise ValueError('invalid select option type %r' % type(opt2)) option_group.add_option(group_label, group_value) else: options.add_option(label, value) return raw_select(name, selected_values, options, id=id, **attrs) def branding(name, length=40): return truncate(name, length, indicator="") def FID(raw_id, path): """ Creates a unique ID for filenode based on it's hash of path and commit it's safe to use in urls :param raw_id: :param path: """ return 'c-%s-%s' % (short_id(raw_id), md5_safe(path)[:12]) class _GetError(object): """Get error from form_errors, and represent it as span wrapped error message :param field_name: field to fetch errors for :param form_errors: form errors dict """ def __call__(self, field_name, form_errors): tmpl = """%s""" if form_errors and field_name in form_errors: return literal(tmpl % form_errors.get(field_name)) get_error = _GetError() class _ToolTip(object): def __call__(self, tooltip_title, trim_at=50): """ Special function just to wrap our text into nice formatted autowrapped text :param tooltip_title: """ tooltip_title = escape(tooltip_title) tooltip_title = tooltip_title.replace('<', '<').replace('>', '>') return tooltip_title tooltip = _ToolTip() files_icon = '' def files_breadcrumbs(repo_name, repo_type, commit_id, file_path, landing_ref_name=None, at_ref=None, limit_items=False, linkify_last_item=False, hide_last_item=False, copy_path_icon=True): if at_ref: route_qry = {'at': at_ref} default_landing_ref = at_ref or landing_ref_name or commit_id else: route_qry = None default_landing_ref = commit_id # first segment is a `HOME` link to repo files root location root_name = literal('') url_segments = [ link_to( root_name, repo_files_by_ref_url( repo_name, repo_type, f_path=None, # None here is a special case for SVN repos, # that won't prefix with a ref ref_name=default_landing_ref, commit_id=commit_id, query=route_qry ) )] path_segments = file_path.split('/') last_cnt = len(path_segments) - 1 for cnt, segment in enumerate(path_segments): if not segment: continue segment_html = escape(segment) last_item = cnt == last_cnt if last_item and hide_last_item: # iterate over and hide last element continue if last_item and linkify_last_item is False: # plain version url_segments.append(segment_html) else: url_segments.append( link_to( segment_html, repo_files_by_ref_url( repo_name, repo_type, f_path='/'.join(path_segments[:cnt + 1]), ref_name=default_landing_ref, commit_id=commit_id, query=route_qry ), )) limited_url_segments = url_segments[:1] + ['...'] + url_segments[-5:] if limit_items and len(limited_url_segments) < len(url_segments): url_segments = limited_url_segments full_path = file_path if copy_path_icon: icon = files_icon.format(escape(full_path)) else: icon = '' if file_path == '': return root_name else: return literal(' / '.join(url_segments) + icon) def files_url_data(request): matchdict = request.matchdict if 'f_path' not in matchdict: matchdict['f_path'] = '' else: matchdict['f_path'] = urllib.parse.quote(safe_str(matchdict['f_path'])) if 'commit_id' not in matchdict: matchdict['commit_id'] = 'tip' return ext_json.str_json(matchdict) def repo_files_by_ref_url(db_repo_name, db_repo_type, f_path, ref_name, commit_id, query=None, ): _is_svn = is_svn(db_repo_type) final_f_path = f_path if _is_svn: """ For SVN the ref_name cannot be used as a commit_id, it needs to be prefixed with actually commit_id followed by the ref_name. This should be done only in case This is a initial landing url, without additional paths. like: /1000/tags/1.0.0/?at=tags/1.0.0 """ if ref_name and ref_name != 'tip': # NOTE(marcink): for svn the ref_name is actually the stored path, so we prefix it # for SVN we only do this magic prefix if it's root, .eg landing revision # of files link. If we are in the tree we don't need this since we traverse the url # that has everything stored if f_path in ['', '/']: final_f_path = '/'.join([ref_name, f_path]) # SVN always needs a commit_id explicitly, without a named REF default_commit_id = commit_id else: """ For git and mercurial we construct a new URL using the names instead of commit_id like: /master/some_path?at=master """ # We currently do not support branches with slashes if '/' in ref_name: default_commit_id = commit_id else: default_commit_id = ref_name # sometimes we pass f_path as None, to indicate explicit no prefix, # we translate it to string to not have None final_f_path = final_f_path or '' files_url = route_path( 'repo_files', repo_name=db_repo_name, commit_id=default_commit_id, f_path=final_f_path, _query=query ) return files_url def code_highlight(code, lexer, formatter, use_hl_filter=False): """ Lex ``code`` with ``lexer`` and format it with the formatter ``formatter``. If ``outfile`` is given and a valid file object (an object with a ``write`` method), the result will be written to it, otherwise it is returned as a string. """ if use_hl_filter: # add HL filter from rhodecode.lib.index import search_utils lexer.add_filter(search_utils.ElasticSearchHLFilter()) return pygments.format(pygments.lex(code, lexer), formatter) class CodeHtmlFormatter(HtmlFormatter): """ My code Html Formatter for source codes """ def wrap(self, source): return self._wrap_div(self._wrap_pre(self._wrap_code(source))) def _wrap_code(self, source): for cnt, it in enumerate(source): i, t = it t = f'
{t}
' yield i, t def _wrap_tablelinenos(self, inner): dummyoutfile = io.StringIO() lncount = 0 for t, line in inner: if t: lncount += 1 dummyoutfile.write(line) fl = self.linenostart mw = len(str(lncount + fl - 1)) sp = self.linenospecial st = self.linenostep la = self.lineanchors aln = self.anchorlinenos nocls = self.noclasses if sp: lines = [] for i in range(fl, fl + lncount): if i % st == 0: if i % sp == 0: if aln: lines.append('%*d' % (la, i, mw, i)) else: lines.append('%*d' % (mw, i)) else: if aln: lines.append('%*d' % (la, i, mw, i)) else: lines.append('%*d' % (mw, i)) else: lines.append('') ls = '\n'.join(lines) else: lines = [] for i in range(fl, fl + lncount): if i % st == 0: if aln: lines.append('%*d' % (la, i, mw, i)) else: lines.append('%*d' % (mw, i)) else: lines.append('') ls = '\n'.join(lines) # in case you wonder about the seemingly redundant
here: since the # content in the other cell also is wrapped in a div, some browsers in # some configurations seem to mess up the formatting... if nocls: yield 0, ('' % self.cssclass + '
' '
' +
                      ls + '
') else: yield 0, ('' % self.cssclass + '
' +
                      ls + '
') yield 0, dummyoutfile.getvalue() yield 0, '
' class SearchContentCodeHtmlFormatter(CodeHtmlFormatter): def __init__(self, **kw): # only show these line numbers if set self.only_lines = kw.pop('only_line_numbers', []) self.query_terms = kw.pop('query_terms', []) self.max_lines = kw.pop('max_lines', 5) self.line_context = kw.pop('line_context', 3) self.url = kw.pop('url', None) super(CodeHtmlFormatter, self).__init__(**kw) def _wrap_code(self, source): for cnt, it in enumerate(source): i, t = it t = '
%s
' % t yield i, t def _wrap_tablelinenos(self, inner): yield 0, '' % self.cssclass last_shown_line_number = 0 current_line_number = 1 for t, line in inner: if not t: yield t, line continue if current_line_number in self.only_lines: if last_shown_line_number + 1 != current_line_number: yield 0, '' yield 0, '' yield 0, '' yield 0, '' yield 0, '' if self.url: yield 0, '' % ( self.url, current_line_number, current_line_number) else: yield 0, '' % ( current_line_number) yield 0, '' yield 0, '' last_shown_line_number = current_line_number current_line_number += 1 yield 0, '
...
%i%i' + line + '
' def hsv_to_rgb(h, s, v): """ Convert hsv color values to rgb """ if s == 0.0: return v, v, v i = int(h * 6.0) # XXX assume int() truncates! f = (h * 6.0) - i p = v * (1.0 - s) q = v * (1.0 - s * f) t = v * (1.0 - s * (1.0 - f)) i = i % 6 if i == 0: return v, t, p if i == 1: return q, v, p if i == 2: return p, v, t if i == 3: return p, q, v if i == 4: return t, p, v if i == 5: return v, p, q def unique_color_generator(n=10000, saturation=0.10, lightness=0.95): """ Generator for getting n of evenly distributed colors using hsv color and golden ratio. It always return same order of colors :param n: number of colors to generate :param saturation: saturation of returned colors :param lightness: lightness of returned colors :returns: RGB tuple """ golden_ratio = 0.618033988749895 h = 0.22717784590367374 for _ in range(n): h += golden_ratio h %= 1 HSV_tuple = [h, saturation, lightness] RGB_tuple = hsv_to_rgb(*HSV_tuple) yield [str(int(x * 256)) for x in RGB_tuple] def color_hasher(n=10000, saturation=0.10, lightness=0.95): """ Returns a function which when called with an argument returns a unique color for that argument, eg. :param n: number of colors to generate :param saturation: saturation of returned colors :param lightness: lightness of returned colors :returns: css RGB string >>> color_hash = color_hasher() >>> color_hash('hello') 'rgb(34, 12, 59)' >>> color_hash('hello') 'rgb(34, 12, 59)' >>> color_hash('other') 'rgb(90, 224, 159)' """ color_dict = {} cgenerator = unique_color_generator( saturation=saturation, lightness=lightness) def get_color_string(thing): if thing in color_dict: col = color_dict[thing] else: col = color_dict[thing] = next(cgenerator) return "rgb(%s)" % (', '.join(col)) return get_color_string def get_lexer_safe(mimetype=None, filepath=None): """ Tries to return a relevant pygments lexer using mimetype/filepath name, defaulting to plain text if none could be found """ lexer = None try: if mimetype: lexer = get_lexer_for_mimetype(mimetype) if not lexer: lexer = get_lexer_for_filename(filepath) except pygments.util.ClassNotFound: pass if not lexer: lexer = get_lexer_by_name('text') return lexer def get_lexer_for_filenode(filenode): lexer = get_custom_lexer(filenode.extension) or filenode.lexer return lexer def pygmentize(filenode, **kwargs): """ pygmentize function using pygments :param filenode: """ lexer = get_lexer_for_filenode(filenode) return literal(code_highlight(filenode.content, lexer, CodeHtmlFormatter(**kwargs))) def is_following_repo(repo_name, user_id): from rhodecode.model.scm import ScmModel return ScmModel().is_following_repo(repo_name, user_id) class _Message(object): """A message returned by ``Flash.pop_messages()``. Converting the message to a string returns the message text. Instances also have the following attributes: * ``message``: the message text. * ``category``: the category specified when the message was created. """ def __init__(self, category, message, sub_data=None): self.category = category self.message = message self.sub_data = sub_data or {} def __str__(self): return self.message __unicode__ = __str__ def __html__(self): return escape(safe_str(self.message)) class Flash(object): # List of allowed categories. If None, allow any category. categories = ["warning", "notice", "error", "success"] # Default category if none is specified. default_category = "notice" def __init__(self, session_key="flash", categories=None, default_category=None): """ Instantiate a ``Flash`` object. ``session_key`` is the key to save the messages under in the user's session. ``categories`` is an optional list which overrides the default list of categories. ``default_category`` overrides the default category used for messages when none is specified. """ self.session_key = session_key if categories is not None: self.categories = categories if default_category is not None: self.default_category = default_category if self.categories and self.default_category not in self.categories: raise ValueError( "unrecognized default category %r" % (self.default_category,)) def pop_messages(self, session=None, request=None): """ Return all accumulated messages and delete them from the session. The return value is a list of ``Message`` objects. """ messages = [] if not session: if not request: request = get_current_request() session = request.session # Pop the 'old' pylons flash messages. They are tuples of the form # (category, message) for cat, msg in session.pop(self.session_key, []): messages.append(_Message(cat, msg)) # Pop the 'new' pyramid flash messages for each category as list # of strings. for cat in self.categories: for msg in session.pop_flash(queue=cat): sub_data = {} if hasattr(msg, 'rsplit'): flash_data = msg.rsplit('|DELIM|', 1) org_message = flash_data[0] if len(flash_data) > 1: sub_data = json.loads(flash_data[1]) else: org_message = msg messages.append(_Message(cat, org_message, sub_data=sub_data)) # Map messages from the default queue to the 'notice' category. for msg in session.pop_flash(): messages.append(_Message('notice', msg)) session.save() return messages def json_alerts(self, session=None, request=None): payloads = [] messages = flash.pop_messages(session=session, request=request) or [] for message in messages: payloads.append({ 'message': { 'message': '{}'.format(message.message), 'level': message.category, 'force': True, 'subdata': message.sub_data } }) return safe_str(json.dumps(payloads)) def __call__(self, message, category=None, ignore_duplicate=True, session=None, request=None): if not session: if not request: request = get_current_request() session = request.session session.flash( message, queue=category, allow_duplicate=not ignore_duplicate) flash = Flash() #============================================================================== # SCM FILTERS available via h. #============================================================================== from rhodecode.lib.vcs.utils import author_name, author_email from rhodecode.lib.utils2 import age, age_from_seconds from rhodecode.model.db import User, ChangesetStatus email = author_email def capitalize(raw_text): return raw_text.capitalize() def short_id(long_id): return long_id[:12] def hide_credentials(url): from rhodecode.lib.utils2 import credentials_filter return credentials_filter(url) import zoneinfo import tzlocal local_timezone = tzlocal.get_localzone() def get_timezone(datetime_iso, time_is_local=False): tzinfo = '+00:00' # detect if we have a timezone info, otherwise, add it if time_is_local and isinstance(datetime_iso, datetime) and not datetime_iso.tzinfo: force_timezone = os.environ.get('RC_TIMEZONE', '') if force_timezone: force_timezone = zoneinfo.ZoneInfo(force_timezone) timezone = force_timezone or local_timezone offset = datetime_iso.replace(tzinfo=timezone).strftime('%z') tzinfo = '{}:{}'.format(offset[:-2], offset[-2:]) return tzinfo def age_component(datetime_iso, value=None, time_is_local=False, tooltip=True): title = value or format_date(datetime_iso) tzinfo = get_timezone(datetime_iso, time_is_local=time_is_local) return literal( ''.format( cls='tooltip' if tooltip else '', tt_title=('{title}{tzinfo}'.format(title=title, tzinfo=tzinfo)) if tooltip else '', title=title, dt=datetime_iso, tzinfo=tzinfo )) def _shorten_commit_id(commit_id, commit_len=None): if commit_len is None: request = get_current_request() commit_len = request.call_context.visual.show_sha_length return commit_id[:commit_len] def show_id(commit, show_idx=None, commit_len=None): """ Configurable function that shows ID by default it's r123:fffeeefffeee :param commit: commit instance """ if show_idx is None: request = get_current_request() show_idx = request.call_context.visual.show_revision_number raw_id = _shorten_commit_id(commit.raw_id, commit_len=commit_len) if show_idx: return 'r%s:%s' % (commit.idx, raw_id) else: return '%s' % (raw_id, ) def format_date(date): """ use a standardized formatting for dates used in RhodeCode :param date: date/datetime object :return: formatted date """ if date: _fmt = "%a, %d %b %Y %H:%M:%S" return safe_str(date.strftime(_fmt)) return "" class _RepoChecker(object): def __init__(self, backend_alias): self._backend_alias = backend_alias def __call__(self, repository): if hasattr(repository, 'alias'): _type = repository.alias elif hasattr(repository, 'repo_type'): _type = repository.repo_type else: _type = repository return _type == self._backend_alias is_git = _RepoChecker('git') is_hg = _RepoChecker('hg') is_svn = _RepoChecker('svn') def get_repo_type_by_name(repo_name): repo = Repository.get_by_repo_name(repo_name) if repo: return repo.repo_type def is_svn_without_proxy(repository): if is_svn(repository): return not ConfigGet().get_bool('vcs.svn.proxy.enabled') return False def discover_user(author): """ Tries to discover RhodeCode User based on the author string. Author string is typically `FirstName LastName ` """ # if author is already an instance use it for extraction if isinstance(author, User): return author # Valid email in the attribute passed, see if they're in the system _email = author_email(author) if _email != '': user = User.get_by_email(_email, case_insensitive=True, cache=True) if user is not None: return user # Maybe it's a username, we try to extract it and fetch by username ? _author = author_name(author) user = User.get_by_username(_author, case_insensitive=True, cache=True) if user is not None: return user return None def email_or_none(author): # extract email from the commit string _email = author_email(author) # If we have an email, use it, otherwise # see if it contains a username we can get an email from if _email != '': return _email else: user = User.get_by_username( author_name(author), case_insensitive=True, cache=True) if user is not None: return user.email # No valid email, not a valid user in the system, none! return None def link_to_user(author, length=0, **kwargs): user = discover_user(author) # user can be None, but if we have it already it means we can re-use it # in the person() function, so we save 1 intensive-query if user: author = user display_person = person(author, 'username_or_name_or_email') if length: display_person = shorter(display_person, length) if user and user.username != user.DEFAULT_USER: return link_to( escape(display_person), route_path('user_profile', username=user.username), **kwargs) else: return escape(display_person) def link_to_group(users_group_name, **kwargs): return link_to( escape(users_group_name), route_path('user_group_profile', user_group_name=users_group_name), **kwargs) def person(author, show_attr="username_and_name"): user = discover_user(author) if user: return getattr(user, show_attr) else: _author = author_name(author) _email = email(author) return _author or _email def author_string(email): if email: user = User.get_by_email(email, case_insensitive=True, cache=True) if user: if user.first_name or user.last_name: return '%s %s <%s>' % ( user.first_name, user.last_name, email) else: return email else: return email else: return None def person_by_id(id_, show_attr="username_and_name"): # attr to return from fetched user def person_getter(usr): return getattr(usr, show_attr) #maybe it's an ID ? if str(id_).isdigit() or isinstance(id_, int): id_ = int(id_) user = User.get(id_) if user is not None: return person_getter(user) return id_ def gravatar_with_user(request, author, show_disabled=False, tooltip=False): _render = request.get_partial_renderer('rhodecode:templates/base/base.mako') return _render('gravatar_with_user', author, show_disabled=show_disabled, tooltip=tooltip) tags_patterns = OrderedDict( ( ( "lang", ( re.compile(r"\[(lang|language)\ \=\>\ *([a-zA-Z\-\/\#\+\.]*)\]"), '
\\2
', ), ), ( "see", ( re.compile(r"\[see\ \=\>\ *([a-zA-Z0-9\/\=\?\&\ \:\/\.\-]*)\]"), '
see: \\1
', ), ), ( "url", ( re.compile( r"\[url\ \=\>\ \[([a-zA-Z0-9\ \.\-\_]+)\]\((http://|https://|/)(.*?)\)\]" ), '
\\1
', ), ), ( "license", ( re.compile( r"\[license\ \=\>\ *([a-zA-Z0-9\/\=\?\&\ \:\/\.\-]*)\]" ), # don't make it a raw string here... '', ), ), ( "ref", ( re.compile( r"\[(requires|recommends|conflicts|base)\ \=\>\ *([a-zA-Z0-9\-\/]*)\]" ), '
\\1: \\2
', ), ), ( "state", ( re.compile(r"\[(stable|featured|stale|dead|dev|deprecated)\]"), '
\\1
', ), ), # label in grey ( "label", (re.compile(r"\[([a-z]+)\]"), '
\\1
'), ), # generic catch all in grey ( "generic", ( re.compile(r"\[([a-zA-Z0-9\.\-\_]+)\]"), '
\\1
', ), ), ) ) def extract_metatags(value): """ Extract supported meta-tags from given text value """ tags = [] if not value: return tags, '' for key, val in list(tags_patterns.items()): pat, replace_html = val tags.extend([(key, x.group()) for x in pat.finditer(value)]) value = pat.sub('', value) return tags, value def style_metatag(tag_type, value): """ converts tags from value into html equivalent """ if not value: return '' html_value = value tag_data = tags_patterns.get(tag_type) if tag_data: pat, replace_html = tag_data # convert to plain `str` instead of a markup tag to be used in # regex expressions. safe_str doesn't work here html_value = pat.sub(replace_html, value) return html_value def bool2icon(value, show_at_false=True): """ Returns boolean value of a given value, represented as html element with classes that will represent icons :param value: given value to convert to html node """ if value: # does bool conversion return HTML.tag('i', class_="icon-true", title='True') else: # not true as bool if show_at_false: return HTML.tag('i', class_="icon-false", title='False') return HTML.tag('i') def b64(inp): return base64.b64encode(safe_bytes(inp)) #============================================================================== # PERMS #============================================================================== from rhodecode.lib.auth import ( HasPermissionAny, HasPermissionAll, HasRepoPermissionAny, HasRepoPermissionAll, HasRepoGroupPermissionAll, HasRepoGroupPermissionAny, HasRepoPermissionAnyApi, get_csrf_token, csrf_token_key, AuthUser) #============================================================================== # GRAVATAR URL #============================================================================== class InitialsGravatar(object): def __init__(self, email_address, first_name, last_name, size=30, background=None, text_color='#fff'): self.size = size self.first_name = first_name self.last_name = last_name self.email_address = email_address self.background = background or self.str2color(email_address) self.text_color = text_color def get_color_bank(self): """ returns a predefined list of colors that gravatars can use. Those are randomized distinct colors that guarantee readability and uniqueness. generated with: http://phrogz.net/css/distinct-colors.html """ return [ '#bf3030', '#a67f53', '#00ff00', '#5989b3', '#392040', '#d90000', '#402910', '#204020', '#79baf2', '#a700b3', '#bf6060', '#7f5320', '#008000', '#003059', '#ee00ff', '#ff0000', '#8c4b00', '#007300', '#005fb3', '#de73e6', '#ff4040', '#ffaa00', '#3df255', '#203140', '#47004d', '#591616', '#664400', '#59b365', '#0d2133', '#83008c', '#592d2d', '#bf9f60', '#73e682', '#1d3f73', '#73006b', '#402020', '#b2862d', '#397341', '#597db3', '#e600d6', '#a60000', '#736039', '#00b318', '#79aaf2', '#330d30', '#ff8080', '#403010', '#16591f', '#002459', '#8c4688', '#e50000', '#ffbf40', '#00732e', '#102340', '#bf60ac', '#8c4646', '#cc8800', '#00a642', '#1d3473', '#b32d98', '#660e00', '#ffd580', '#80ffb2', '#7391e6', '#733967', '#d97b6c', '#8c5e00', '#59b389', '#3967e6', '#590047', '#73281d', '#665200', '#00e67a', '#2d50b3', '#8c2377', '#734139', '#b2982d', '#16593a', '#001859', '#ff00aa', '#a65e53', '#ffcc00', '#0d3321', '#2d3959', '#731d56', '#401610', '#4c3d00', '#468c6c', '#002ca6', '#d936a3', '#d94c36', '#403920', '#36d9a3', '#0d1733', '#592d4a', '#993626', '#cca300', '#00734d', '#46598c', '#8c005e', '#7f1100', '#8c7000', '#00a66f', '#7382e6', '#b32d74', '#d9896c', '#ffe680', '#1d7362', '#364cd9', '#73003d', '#d93a00', '#998a4d', '#59b3a1', '#5965b3', '#e5007a', '#73341d', '#665f00', '#00b38f', '#0018b3', '#59163a', '#b2502d', '#bfb960', '#00ffcc', '#23318c', '#a6537f', '#734939', '#b2a700', '#104036', '#3d3df2', '#402031', '#e56739', '#736f39', '#79f2ea', '#000059', '#401029', '#4c1400', '#ffee00', '#005953', '#101040', '#990052', '#402820', '#403d10', '#00ffee', '#0000d9', '#ff80c4', '#a66953', '#eeff00', '#00ccbe', '#8080ff', '#e673a1', '#a62c00', '#474d00', '#1a3331', '#46468c', '#733950', '#662900', '#858c23', '#238c85', '#0f0073', '#b20047', '#d9986c', '#becc00', '#396f73', '#281d73', '#ff0066', '#ff6600', '#dee673', '#59adb3', '#6559b3', '#590024', '#b2622d', '#98b32d', '#36ced9', '#332d59', '#40001a', '#733f1d', '#526600', '#005359', '#242040', '#bf6079', '#735039', '#cef23d', '#007780', '#5630bf', '#66001b', '#b24700', '#acbf60', '#1d6273', '#25008c', '#731d34', '#a67453', '#50592d', '#00ccff', '#6600ff', '#ff0044', '#4c1f00', '#8a994d', '#79daf2', '#a173e6', '#d93662', '#402310', '#aaff00', '#2d98b3', '#8c40ff', '#592d39', '#ff8c40', '#354020', '#103640', '#1a0040', '#331a20', '#331400', '#334d00', '#1d5673', '#583973', '#7f0022', '#4c3626', '#88cc00', '#36a3d9', '#3d0073', '#d9364c', '#33241a', '#698c23', '#5995b3', '#300059', '#e57382', '#7f3300', '#366600', '#00aaff', '#3a1659', '#733941', '#663600', '#74b32d', '#003c59', '#7f53a6', '#73000f', '#ff8800', '#baf279', '#79caf2', '#291040', '#a6293a', '#b2742d', '#587339', '#0077b3', '#632699', '#400009', '#d9a66c', '#294010', '#2d4a59', '#aa00ff', '#4c131b', '#b25f00', '#5ce600', '#267399', '#a336d9', '#990014', '#664e33', '#86bf60', '#0088ff', '#7700b3', '#593a16', '#073300', '#1d4b73', '#ac60bf', '#e59539', '#4f8c46', '#368dd9', '#5c0073' ] def rgb_to_hex_color(self, rgb_tuple): """ Converts an rgb_tuple passed to an hex color. :param rgb_tuple: tuple with 3 ints represents rgb color space """ return '#' + ("".join(map(chr, rgb_tuple)).encode('hex')) def email_to_int_list(self, email_str): """ Get every byte of the hex digest value of email and turn it to integer. It's going to be always between 0-255 """ digest = md5_safe(email_str.lower()) return [int(digest[i * 2:i * 2 + 2], 16) for i in range(16)] def pick_color_bank_index(self, email_str, color_bank): return self.email_to_int_list(email_str)[0] % len(color_bank) def str2color(self, email_str): """ Tries to map in a stable algorithm an email to color :param email_str: """ color_bank = self.get_color_bank() # pick position (module it's length so we always find it in the # bank even if it's smaller than 256 values pos = self.pick_color_bank_index(email_str, color_bank) return color_bank[pos] def normalize_email(self, email_address): # default host used to fill in the fake/missing email default_host = 'localhost' if not email_address: email_address = f'{User.DEFAULT_USER}@{default_host}' email_address = safe_str(email_address) if '@' not in email_address: email_address = f'{email_address}@{default_host}' if email_address.endswith('@'): email_address = f'{email_address}{default_host}' email_address = convert_special_chars(email_address) return email_address def get_initials(self): """ Returns 2 letter initials calculated based on the input. The algorithm picks first given email address, and takes first letter of part before @, and then the first letter of server name. In case the part before @ is in a format of `somestring.somestring2` it replaces the server letter with first letter of somestring2 In case function was initialized with both first and lastname, this overrides the extraction from email by first letter of the first and last name. We add special logic to that functionality, In case Full name is compound, like Guido Von Rossum, we use last part of the last name (Von Rossum) picking `R`. Function also normalizes the non-ascii characters to they ascii representation, eg Ą => A """ # replace non-ascii to ascii first_name = convert_special_chars(self.first_name) last_name = convert_special_chars(self.last_name) # multi word last names, Guido Von Rossum, we take the last part only last_name = last_name.split(' ', 1)[-1] # do NFKD encoding, and also make sure email has proper format email_address = self.normalize_email(self.email_address) # first push the email initials prefix, server = email_address.split('@', 1) # check if prefix is maybe a 'first_name.last_name' syntax _dot_split = prefix.rsplit('.', 1) if len(_dot_split) == 2 and _dot_split[1]: initials = [_dot_split[0][0], _dot_split[1][0]] else: initials = [prefix[0], server[0]] # get first letter of first and last names to create initials fn_letter = (first_name or " ")[0].strip() ln_letter = (last_name or " ")[0].strip() if fn_letter: initials[0] = fn_letter if ln_letter: initials[1] = ln_letter return ''.join(initials).upper() def get_img_data_by_type(self, font_family, img_type): default_user = """ """.format( size=self.size, background='#979797', # @grey4 text_color=self.text_color, font_family=font_family) return { "default_user": default_user }[img_type] def get_img_data(self, svg_type=None): """ generates the svg metadata for image """ fonts = [ '-apple-system', 'BlinkMacSystemFont', 'Segoe UI', 'Roboto', 'Oxygen-Sans', 'Ubuntu', 'Cantarell', 'Helvetica Neue', 'sans-serif' ] font_family = ','.join(fonts) if svg_type: return self.get_img_data_by_type(font_family, svg_type) initials = self.get_initials() img_data = """ {text} """.format( size=self.size, f_size=self.size/2.05, # scale the text inside the box nicely background=self.background, text_color=self.text_color, text=initials.upper(), font_family=font_family) return img_data def generate_svg(self, svg_type=None): img_data = base64_to_str(self.get_img_data(svg_type)) return "data:image/svg+xml;base64,{}".format(img_data) def initials_gravatar(request, email_address, first_name, last_name, size=30): svg_type = None if email_address == User.DEFAULT_USER_EMAIL: svg_type = 'default_user' klass = InitialsGravatar(email_address, first_name, last_name, size) return klass.generate_svg(svg_type=svg_type) def gravatar_external(request, gravatar_url_tmpl, email_address, size=30): return safe_str(gravatar_url_tmpl)\ .replace('{email}', email_address) \ .replace('{md5email}', md5_safe(email_address.lower())) \ .replace('{netloc}', request.host) \ .replace('{scheme}', request.scheme) \ .replace('{size}', safe_str(size)) def gravatar_url(email_address, size=30, request=None): request = request or get_current_request() _use_gravatar = request.call_context.visual.use_gravatar email_address = email_address or User.DEFAULT_USER_EMAIL if isinstance(email_address, str): # hashlib crashes on unicode items email_address = safe_str(email_address) # empty email or default user if not email_address or email_address == User.DEFAULT_USER_EMAIL: return initials_gravatar(request, User.DEFAULT_USER_EMAIL, '', '', size=size) if _use_gravatar: gravatar_url_tmpl = request.call_context.visual.gravatar_url \ or User.DEFAULT_GRAVATAR_URL return gravatar_external(request, gravatar_url_tmpl, email_address, size=size) else: return initials_gravatar(request, email_address, '', '', size=size) def breadcrumb_repo_link(repo): """ Makes a breadcrumbs path link to repo ex:: group >> subgroup >> repo :param repo: a Repository instance """ path = [ link_to(group.name, route_path('repo_group_home', repo_group_name=group.group_name), title='last change:{}'.format(format_date(group.last_commit_change))) for group in repo.groups_with_parents ] + [ link_to(repo.just_name, route_path('repo_summary', repo_name=repo.repo_name), title='last change:{}'.format(format_date(repo.last_commit_change))) ] return literal(' » '.join(path)) def breadcrumb_repo_group_link(repo_group): """ Makes a breadcrumbs path link to repo ex:: group >> subgroup :param repo_group: a Repository Group instance """ path = [ link_to(group.name, route_path('repo_group_home', repo_group_name=group.group_name), title='last change:{}'.format(format_date(group.last_commit_change))) for group in repo_group.parents ] + [ link_to(repo_group.name, route_path('repo_group_home', repo_group_name=repo_group.group_name), title='last change:{}'.format(format_date(repo_group.last_commit_change))) ] return literal(' » '.join(path)) def format_byte_size_binary(file_size): """ Formats file/folder sizes to standard. """ if file_size is None: file_size = 0 formatted_size = format_byte_size(file_size, binary=True) return formatted_size def urlify_text(text_, safe=True, **href_attrs): """ Extract urls from text and make html links out of them """ url_pat = re.compile(r'''(http[s]?://(?:[a-zA-Z]|[0-9]|[$-_@#.&+]''' r'''|[!*\(\),]|(?:%[0-9a-fA-F][0-9a-fA-F]))+)''') def url_func(match_obj): url_full = match_obj.groups()[0] a_options = dict(href_attrs) a_options['href'] = url_full a_text = url_full return HTML.tag("a", a_text, **a_options) _new_text = url_pat.sub(url_func, text_) if safe: return literal(_new_text) return _new_text def urlify_commits(text_, repo_name): """ Extract commit ids from text and make link from them :param text_: :param repo_name: repo name to build the URL with """ url_pat = re.compile(r'(^|\s)([0-9a-fA-F]{12,40})($|\s)') def url_func(match_obj): commit_id = match_obj.groups()[1] pref = match_obj.groups()[0] suf = match_obj.groups()[2] tmpl = ( '%(pref)s' '%(commit_id)s%(suf)s' ) return tmpl % { 'pref': pref, 'cls': 'revision-link', 'url': route_url( 'repo_commit', repo_name=repo_name, commit_id=commit_id), 'commit_id': commit_id, 'suf': suf, 'hovercard_alt': 'Commit: {}'.format(commit_id), 'hovercard_url': route_url( 'hovercard_repo_commit', repo_name=repo_name, commit_id=commit_id) } new_text = url_pat.sub(url_func, text_) return new_text def _process_url_func(match_obj, repo_name, uid, entry, return_raw_data=False, link_format='html'): pref = '' if match_obj.group().startswith(' '): pref = ' ' issue_id = ''.join(match_obj.groups()) if link_format == 'html': tmpl = ( '%(pref)s' '%(issue-prefix)s%(id-repr)s' '') elif link_format == 'html+hovercard': tmpl = ( '%(pref)s' '%(issue-prefix)s%(id-repr)s' '') elif link_format in ['rst', 'rst+hovercard']: tmpl = '`%(issue-prefix)s%(id-repr)s <%(url)s>`_' elif link_format in ['markdown', 'markdown+hovercard']: tmpl = '[%(pref)s%(issue-prefix)s%(id-repr)s](%(url)s)' else: raise ValueError('Bad link_format:{}'.format(link_format)) (repo_name_cleaned, parent_group_name) = RepoGroupModel()._get_group_name_and_parent(repo_name) # variables replacement named_vars = { 'id': issue_id, 'repo': repo_name, 'repo_name': repo_name_cleaned, 'group_name': parent_group_name, # set dummy keys so we always have them 'hostname': '', 'netloc': '', 'scheme': '' } request = get_current_request() if request: # exposes, hostname, netloc, scheme host_data = get_host_info(request) named_vars.update(host_data) # named regex variables named_vars.update(match_obj.groupdict()) _url = string.Template(entry['url']).safe_substitute(**named_vars) desc = string.Template(escape(entry['desc'])).safe_substitute(**named_vars) hovercard_url = string.Template(entry.get('hovercard_url', '')).safe_substitute(**named_vars) def quote_cleaner(input_str): """Remove quotes as it's HTML""" return input_str.replace('"', '') data = { 'pref': pref, 'cls': quote_cleaner('issue-tracker-link'), 'url': quote_cleaner(_url), 'id-repr': issue_id, 'issue-prefix': entry['pref'], 'serv': entry['url'], 'title': sanitize_html(desc, strip=True), 'hovercard_url': hovercard_url } if return_raw_data: return { 'id': issue_id, 'url': _url } return tmpl % data def get_active_pattern_entries(repo_name): repo = None if repo_name: # Retrieving repo_name to avoid invalid repo_name to explode on # IssueTrackerSettingsModel but still passing invalid name further down repo = Repository.get_by_repo_name(repo_name, cache=True) settings_model = IssueTrackerSettingsModel(repo=repo) active_entries = settings_model.get_settings(cache=True) return active_entries pr_pattern_re = regex.compile(r'(?:(?:^!)|(?: !))(\d+)') allowed_link_formats = [ 'html', 'rst', 'markdown', 'html+hovercard', 'rst+hovercard', 'markdown+hovercard'] compile_cache = { } def process_patterns(text_string, repo_name, link_format='html', active_entries=None): if link_format not in allowed_link_formats: raise ValueError('Link format can be only one of:{} got {}'.format( allowed_link_formats, link_format)) issues_data = [] errors = [] new_text = text_string if active_entries is None: log.debug('Fetch active issue tracker patterns for repo: %s', repo_name) active_entries = get_active_pattern_entries(repo_name) log.debug('Got %s pattern entries to process', len(active_entries)) for uid, entry in list(active_entries.items()): if not (entry['pat'] and entry['url']): log.debug('skipping due to missing data') continue log.debug('issue tracker entry: uid: `%s` PAT:%s URL:%s PREFIX:%s', uid, entry['pat'], entry['url'], entry['pref']) if entry.get('pat_compiled'): pattern = entry['pat_compiled'] elif entry['pat'] in compile_cache: pattern = compile_cache[entry['pat']] else: try: pattern = regex.compile(r'%s' % entry['pat']) except regex.error as e: regex_err = ValueError('{}:{}'.format(entry['pat'], e)) log.exception('issue tracker pattern: `%s` failed to compile', regex_err) errors.append(regex_err) continue compile_cache[entry['pat']] = pattern data_func = partial( _process_url_func, repo_name=repo_name, entry=entry, uid=uid, return_raw_data=True) for match_obj in pattern.finditer(text_string): issues_data.append(data_func(match_obj)) url_func = partial( _process_url_func, repo_name=repo_name, entry=entry, uid=uid, link_format=link_format) new_text = pattern.sub(url_func, new_text) log.debug('processed prefix:uid `%s`', uid) # finally use global replace, eg !123 -> pr-link, those will not catch # if already similar pattern exists server_url = '${scheme}://${netloc}' pr_entry = { 'pref': '!', 'url': server_url + '/_admin/pull-requests/${id}', 'desc': 'Pull Request !${id}', 'hovercard_url': server_url + '/_hovercard/pull_request/${id}' } pr_url_func = partial( _process_url_func, repo_name=repo_name, entry=pr_entry, uid=None, link_format=link_format+'+hovercard') new_text = pr_pattern_re.sub(pr_url_func, new_text) log.debug('processed !pr pattern') return new_text, issues_data, errors def urlify_commit_message(commit_text, repository=None, active_pattern_entries=None, issues_container_callback=None, error_container=None): """ Parses given text message and makes proper links. issues are linked to given issue-server, and rest is a commit link """ def escaper(_text): return _text.replace('<', '<').replace('>', '>') new_text = escaper(commit_text) # extract http/https links and make them real urls new_text = urlify_text(new_text, safe=False) # urlify commits - extract commit ids and make link out of them, if we have # the scope of repository present. if repository: new_text = urlify_commits(new_text, repository) # process issue tracker patterns new_text, issues, errors = process_patterns( new_text, repository or '', active_entries=active_pattern_entries) if issues_container_callback is not None: for issue in issues: issues_container_callback(issue) if error_container is not None: error_container.extend(errors) return literal(new_text) def render_binary(repo_name, file_obj): """ Choose how to render a binary file """ # unicode filename = file_obj.name # images for ext in ['*.png', '*.jpeg', '*.jpg', '*.ico', '*.gif']: if fnmatch.fnmatch(filename, pat=ext): src = route_path( 'repo_file_raw', repo_name=repo_name, commit_id=file_obj.commit.raw_id, f_path=file_obj.path) return literal( 'rendered-image'.format(src)) def renderer_from_filename(filename, exclude=None): """ choose a renderer based on filename, this works only for text based files """ # ipython for ext in ['*.ipynb']: if fnmatch.fnmatch(filename, pat=ext): return 'jupyter' is_markup = MarkupRenderer.renderer_from_filename(filename, exclude=exclude) if is_markup: return is_markup return None def render(source, renderer='rst', mentions=False, relative_urls=None, repo_name=None, active_pattern_entries=None, issues_container_callback=None): def maybe_convert_relative_links(html_source): if relative_urls: return relative_links(html_source, relative_urls) return html_source if renderer == 'plain': return literal( MarkupRenderer.plain(source, leading_newline=False)) elif renderer == 'rst': if repo_name: # process patterns on comments if we pass in repo name source, issues, errors = process_patterns( source, repo_name, link_format='rst', active_entries=active_pattern_entries) if issues_container_callback is not None: for issue in issues: issues_container_callback(issue) rendered_block = maybe_convert_relative_links( MarkupRenderer.rst(source, mentions=mentions)) return literal(f'
{rendered_block}
') elif renderer == 'markdown': if repo_name: # process patterns on comments if we pass in repo name source, issues, errors = process_patterns( source, repo_name, link_format='markdown', active_entries=active_pattern_entries) if issues_container_callback is not None: for issue in issues: issues_container_callback(issue) rendered_block = maybe_convert_relative_links( MarkupRenderer.markdown(source, flavored=True, mentions=mentions)) return literal(f'
{rendered_block}
') elif renderer == 'jupyter': rendered_block = maybe_convert_relative_links( MarkupRenderer.jupyter(source)) return literal(f'
{rendered_block}
') # None means just show the file-source return None def commit_status(repo, commit_id): return ChangesetStatusModel().get_status(repo, commit_id) def commit_status_lbl(commit_status): return dict(ChangesetStatus.STATUSES).get(commit_status) def commit_time(repo_name, commit_id): repo = Repository.get_by_repo_name(repo_name) commit = repo.get_commit(commit_id=commit_id) return commit.date def get_permission_name(key): return dict(Permission.PERMS).get(key) def journal_filter_help(request): _ = request.translate from rhodecode.lib.audit_logger import ACTIONS actions = '\n'.join(textwrap.wrap(', '.join(sorted(ACTIONS.keys())), 80)) return _( 'Example filter terms:\n' + ' repository:vcs\n' + ' username:marcin\n' + ' username:(NOT marcin)\n' + ' action:*push*\n' + ' ip:127.0.0.1\n' + ' date:20120101\n' + ' date:[20120101100000 TO 20120102]\n' + '\n' + 'Actions: {actions}\n' + '\n' + 'Generate wildcards using \'*\' character:\n' + ' "repository:vcs*" - search everything starting with \'vcs\'\n' + ' "repository:*vcs*" - search for repository containing \'vcs\'\n' + '\n' + 'Optional AND / OR operators in queries\n' + ' "repository:vcs OR repository:test"\n' + ' "username:test AND repository:test*"\n' ).format(actions=actions) def not_mapped_error(repo_name): from rhodecode.translation import _ flash(_('%s repository is not mapped to db perhaps' ' it was created or renamed from the filesystem' ' please run the application again' ' in order to rescan repositories') % repo_name, category='error') def ip_range(ip_addr): from rhodecode.model.db import UserIpMap s, e = UserIpMap._get_ip_range(ip_addr) return '%s - %s' % (s, e) def form(url, method='post', needs_csrf_token=True, **attrs): """Wrapper around webhelpers.tags.form to prevent CSRF attacks.""" if method.lower() != 'get' and needs_csrf_token: raise Exception( 'Forms to POST/PUT/DELETE endpoints should have (in general) a ' + 'CSRF token. If the endpoint does not require such token you can ' + 'explicitly set the parameter needs_csrf_token to false.') return insecure_form(url, method=method, **attrs) def secure_form(form_url, method="POST", multipart=False, **attrs): """Start a form tag that points the action to an url. This form tag will also include the hidden field containing the auth token. The url options should be given either as a string, or as a ``url()`` function. The method for the form defaults to POST. Options: ``multipart`` If set to True, the enctype is set to "multipart/form-data". ``method`` The method to use when submitting the form, usually either "GET" or "POST". If "PUT", "DELETE", or another verb is used, a hidden input with name _method is added to simulate the verb over POST. """ if 'request' in attrs: session = attrs['request'].session del attrs['request'] else: raise ValueError( 'Calling this form requires request= to be passed as argument') _form = insecure_form(form_url, method, multipart, **attrs) token = literal( ''.format( csrf_token_key, get_csrf_token(session))) return literal("%s\n%s" % (_form, token)) def dropdownmenu(name, selected, options, enable_filter=False, **attrs): select_html = select(name, selected, options, **attrs) select2 = """ """ filter_option = """, minimumResultsForSearch: -1 """ input_id = attrs.get('id') or name extra_classes = ' '.join(attrs.pop('extra_classes', [])) filter_enabled = "" if enable_filter else filter_option select_script = literal(select2 % (input_id, extra_classes, filter_enabled)) return literal(select_html+select_script) def get_visual_attr(tmpl_context_var, attr_name): """ A safe way to get a variable from visual variable of template context :param tmpl_context_var: instance of tmpl_context, usually present as `c` :param attr_name: name of the attribute we fetch from the c.visual """ visual = getattr(tmpl_context_var, 'visual', None) if not visual: return else: return getattr(visual, attr_name, None) def get_last_path_part(file_node): if not file_node.path: return '/' path = safe_str(file_node.path.split('/')[-1]) return '../' + path def route_url(*args, **kwargs): """ Wrapper around pyramids `route_url` (fully qualified url) function. """ req = get_current_request() return req.route_url(*args, **kwargs) def route_path(*args, **kwargs): """ Wrapper around pyramids `route_path` function. """ req = get_current_request() return req.route_path(*args, **kwargs) def route_path_or_none(*args, **kwargs): try: return route_path(*args, **kwargs) except KeyError: return None def current_route_path(request, **kw): new_args = request.GET.mixed() new_args.update(kw) return request.current_route_path(_query=new_args) def curl_api_example(method, args): args_json = json.dumps(OrderedDict([ ('id', 1), ('auth_token', 'SECRET'), ('method', method), ('args', args) ])) return "curl {api_url} -X POST -H 'content-type:text/plain' --data-binary '{args_json}'".format( api_url=route_url('apiv2'), args_json=args_json ) def api_call_example(method, args): """ Generates an API call example via CURL """ curl_call = curl_api_example(method, args) return literal( curl_call + "

SECRET can be found in auth-tokens page, " "and needs to be of `api calls` role." .format(token_url=route_url('my_account_auth_tokens'))) def notification_description(notification, request): """ Generate notification human readable description based on notification type """ from rhodecode.model.notification import NotificationModel return NotificationModel().make_description( notification, translate=request.translate) def go_import_header(request, db_repo=None): """ Creates a header for go-import functionality in Go Lang """ if not db_repo: return if 'go-get' not in request.GET: return clone_url = db_repo.clone_url() prefix = re.split(r'^https?:\/\/', clone_url)[-1] # we have a repo and go-get flag, return literal(''.format( prefix, db_repo.repo_type, clone_url)) def reviewer_as_json(*args, **kwargs): from rhodecode.apps.repository.utils import reviewer_as_json as _reviewer_as_json return _reviewer_as_json(*args, **kwargs) def get_repo_view_type(request): route_name = request.matched_route.name route_to_view_type = { 'repo_changelog': 'commits', 'repo_commits': 'commits', 'repo_files': 'files', 'repo_summary': 'summary', 'repo_commit': 'commit' } return route_to_view_type.get(route_name) def is_active(menu_entry, selected): """ Returns active class for selecting menus in templates
  • """ if not isinstance(menu_entry, list): menu_entry = [menu_entry] if selected in menu_entry: return "active" class IssuesRegistry(object): """ issue_registry = IssuesRegistry() some_func(issues_callback=issues_registry(...)) """ def __init__(self): self.issues = [] self.unique_issues = collections.defaultdict(lambda: []) def __call__(self, commit_dict=None): def callback(issue): if commit_dict and issue: issue['commit'] = commit_dict self.issues.append(issue) self.unique_issues[issue['id']].append(issue) return callback def get_issues(self): return self.issues @property def issues_unique_count(self): return len(set(i['id'] for i in self.issues)) def get_directory_statistics(start_path): """ total_files, total_size, directory_stats = get_directory_statistics(start_path) print(f"Directory statistics for: {start_path}\n") print(f"Total files: {total_files}") print(f"Total size: {format_size(total_size)}\n") :param start_path: :return: """ total_files = 0 total_size = 0 directory_stats = {} for dir_path, dir_names, file_names in os.walk(start_path): dir_size = 0 file_count = len(file_names) for fname in file_names: filepath = os.path.join(dir_path, fname) file_size = os.path.getsize(filepath) dir_size += file_size directory_stats[dir_path] = {'file_count': file_count, 'size': dir_size} total_files += file_count total_size += dir_size return total_files, total_size, directory_stats