##// END OF EJS Templates
initial-gravatars: fix case of dot beeing present before @domain.
dan -
r2062:623ff10e default
parent child Browse files
Show More
@@ -1,2045 +1,2045 b''
1 # -*- coding: utf-8 -*-
1 # -*- coding: utf-8 -*-
2
2
3 # Copyright (C) 2010-2017 RhodeCode GmbH
3 # Copyright (C) 2010-2017 RhodeCode GmbH
4 #
4 #
5 # This program is free software: you can redistribute it and/or modify
5 # This program is free software: you can redistribute it and/or modify
6 # it under the terms of the GNU Affero General Public License, version 3
6 # it under the terms of the GNU Affero General Public License, version 3
7 # (only), as published by the Free Software Foundation.
7 # (only), as published by the Free Software Foundation.
8 #
8 #
9 # This program is distributed in the hope that it will be useful,
9 # This program is distributed in the hope that it will be useful,
10 # but WITHOUT ANY WARRANTY; without even the implied warranty of
10 # but WITHOUT ANY WARRANTY; without even the implied warranty of
11 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
11 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 # GNU General Public License for more details.
12 # GNU General Public License for more details.
13 #
13 #
14 # You should have received a copy of the GNU Affero General Public License
14 # You should have received a copy of the GNU Affero General Public License
15 # along with this program. If not, see <http://www.gnu.org/licenses/>.
15 # along with this program. If not, see <http://www.gnu.org/licenses/>.
16 #
16 #
17 # This program is dual-licensed. If you wish to learn more about the
17 # This program is dual-licensed. If you wish to learn more about the
18 # RhodeCode Enterprise Edition, including its added features, Support services,
18 # RhodeCode Enterprise Edition, including its added features, Support services,
19 # and proprietary license terms, please see https://rhodecode.com/licenses/
19 # and proprietary license terms, please see https://rhodecode.com/licenses/
20
20
21 """
21 """
22 Helper functions
22 Helper functions
23
23
24 Consists of functions to typically be used within templates, but also
24 Consists of functions to typically be used within templates, but also
25 available to Controllers. This module is available to both as 'h'.
25 available to Controllers. This module is available to both as 'h'.
26 """
26 """
27
27
28 import random
28 import random
29 import hashlib
29 import hashlib
30 import StringIO
30 import StringIO
31 import urllib
31 import urllib
32 import math
32 import math
33 import logging
33 import logging
34 import re
34 import re
35 import urlparse
35 import urlparse
36 import time
36 import time
37 import string
37 import string
38 import hashlib
38 import hashlib
39 from collections import OrderedDict
39 from collections import OrderedDict
40
40
41 import pygments
41 import pygments
42 import itertools
42 import itertools
43 import fnmatch
43 import fnmatch
44
44
45 from datetime import datetime
45 from datetime import datetime
46 from functools import partial
46 from functools import partial
47 from pygments.formatters.html import HtmlFormatter
47 from pygments.formatters.html import HtmlFormatter
48 from pygments import highlight as code_highlight
48 from pygments import highlight as code_highlight
49 from pygments.lexers import (
49 from pygments.lexers import (
50 get_lexer_by_name, get_lexer_for_filename, get_lexer_for_mimetype)
50 get_lexer_by_name, get_lexer_for_filename, get_lexer_for_mimetype)
51
51
52 from pyramid.threadlocal import get_current_request
52 from pyramid.threadlocal import get_current_request
53
53
54 from webhelpers.html import literal, HTML, escape
54 from webhelpers.html import literal, HTML, escape
55 from webhelpers.html.tools import *
55 from webhelpers.html.tools import *
56 from webhelpers.html.builder import make_tag
56 from webhelpers.html.builder import make_tag
57 from webhelpers.html.tags import auto_discovery_link, checkbox, css_classes, \
57 from webhelpers.html.tags import auto_discovery_link, checkbox, css_classes, \
58 end_form, file, form as wh_form, hidden, image, javascript_link, link_to, \
58 end_form, file, form as wh_form, hidden, image, javascript_link, link_to, \
59 link_to_if, link_to_unless, ol, required_legend, select, stylesheet_link, \
59 link_to_if, link_to_unless, ol, required_legend, select, stylesheet_link, \
60 submit, text, password, textarea, title, ul, xml_declaration, radio
60 submit, text, password, textarea, title, ul, xml_declaration, radio
61 from webhelpers.html.tools import auto_link, button_to, highlight, \
61 from webhelpers.html.tools import auto_link, button_to, highlight, \
62 js_obfuscate, mail_to, strip_links, strip_tags, tag_re
62 js_obfuscate, mail_to, strip_links, strip_tags, tag_re
63 from webhelpers.pylonslib import Flash as _Flash
63 from webhelpers.pylonslib import Flash as _Flash
64 from webhelpers.text import chop_at, collapse, convert_accented_entities, \
64 from webhelpers.text import chop_at, collapse, convert_accented_entities, \
65 convert_misc_entities, lchop, plural, rchop, remove_formatting, \
65 convert_misc_entities, lchop, plural, rchop, remove_formatting, \
66 replace_whitespace, urlify, truncate, wrap_paragraphs
66 replace_whitespace, urlify, truncate, wrap_paragraphs
67 from webhelpers.date import time_ago_in_words
67 from webhelpers.date import time_ago_in_words
68 from webhelpers.paginate import Page as _Page
68 from webhelpers.paginate import Page as _Page
69 from webhelpers.html.tags import _set_input_attrs, _set_id_attr, \
69 from webhelpers.html.tags import _set_input_attrs, _set_id_attr, \
70 convert_boolean_attrs, NotGiven, _make_safe_id_component
70 convert_boolean_attrs, NotGiven, _make_safe_id_component
71 from webhelpers2.number import format_byte_size
71 from webhelpers2.number import format_byte_size
72
72
73 from rhodecode.lib.action_parser import action_parser
73 from rhodecode.lib.action_parser import action_parser
74 from rhodecode.lib.ext_json import json
74 from rhodecode.lib.ext_json import json
75 from rhodecode.lib.utils import repo_name_slug, get_custom_lexer
75 from rhodecode.lib.utils import repo_name_slug, get_custom_lexer
76 from rhodecode.lib.utils2 import str2bool, safe_unicode, safe_str, \
76 from rhodecode.lib.utils2 import str2bool, safe_unicode, safe_str, \
77 get_commit_safe, datetime_to_time, time_to_datetime, time_to_utcdatetime, \
77 get_commit_safe, datetime_to_time, time_to_datetime, time_to_utcdatetime, \
78 AttributeDict, safe_int, md5, md5_safe
78 AttributeDict, safe_int, md5, md5_safe
79 from rhodecode.lib.markup_renderer import MarkupRenderer, relative_links
79 from rhodecode.lib.markup_renderer import MarkupRenderer, relative_links
80 from rhodecode.lib.vcs.exceptions import CommitDoesNotExistError
80 from rhodecode.lib.vcs.exceptions import CommitDoesNotExistError
81 from rhodecode.lib.vcs.backends.base import BaseChangeset, EmptyCommit
81 from rhodecode.lib.vcs.backends.base import BaseChangeset, EmptyCommit
82 from rhodecode.config.conf import DATE_FORMAT, DATETIME_FORMAT
82 from rhodecode.config.conf import DATE_FORMAT, DATETIME_FORMAT
83 from rhodecode.model.changeset_status import ChangesetStatusModel
83 from rhodecode.model.changeset_status import ChangesetStatusModel
84 from rhodecode.model.db import Permission, User, Repository
84 from rhodecode.model.db import Permission, User, Repository
85 from rhodecode.model.repo_group import RepoGroupModel
85 from rhodecode.model.repo_group import RepoGroupModel
86 from rhodecode.model.settings import IssueTrackerSettingsModel
86 from rhodecode.model.settings import IssueTrackerSettingsModel
87
87
88 log = logging.getLogger(__name__)
88 log = logging.getLogger(__name__)
89
89
90
90
91 DEFAULT_USER = User.DEFAULT_USER
91 DEFAULT_USER = User.DEFAULT_USER
92 DEFAULT_USER_EMAIL = User.DEFAULT_USER_EMAIL
92 DEFAULT_USER_EMAIL = User.DEFAULT_USER_EMAIL
93
93
94
94
95 def url(*args, **kw):
95 def url(*args, **kw):
96 from pylons import url as pylons_url
96 from pylons import url as pylons_url
97 return pylons_url(*args, **kw)
97 return pylons_url(*args, **kw)
98
98
99
99
100 def pylons_url_current(*args, **kw):
100 def pylons_url_current(*args, **kw):
101 """
101 """
102 This function overrides pylons.url.current() which returns the current
102 This function overrides pylons.url.current() which returns the current
103 path so that it will also work from a pyramid only context. This
103 path so that it will also work from a pyramid only context. This
104 should be removed once port to pyramid is complete.
104 should be removed once port to pyramid is complete.
105 """
105 """
106 from pylons import url as pylons_url
106 from pylons import url as pylons_url
107 if not args and not kw:
107 if not args and not kw:
108 request = get_current_request()
108 request = get_current_request()
109 return request.path
109 return request.path
110 return pylons_url.current(*args, **kw)
110 return pylons_url.current(*args, **kw)
111
111
112 url.current = pylons_url_current
112 url.current = pylons_url_current
113
113
114
114
115 def url_replace(**qargs):
115 def url_replace(**qargs):
116 """ Returns the current request url while replacing query string args """
116 """ Returns the current request url while replacing query string args """
117
117
118 request = get_current_request()
118 request = get_current_request()
119 new_args = request.GET.mixed()
119 new_args = request.GET.mixed()
120 new_args.update(qargs)
120 new_args.update(qargs)
121 return url('', **new_args)
121 return url('', **new_args)
122
122
123
123
124 def asset(path, ver=None, **kwargs):
124 def asset(path, ver=None, **kwargs):
125 """
125 """
126 Helper to generate a static asset file path for rhodecode assets
126 Helper to generate a static asset file path for rhodecode assets
127
127
128 eg. h.asset('images/image.png', ver='3923')
128 eg. h.asset('images/image.png', ver='3923')
129
129
130 :param path: path of asset
130 :param path: path of asset
131 :param ver: optional version query param to append as ?ver=
131 :param ver: optional version query param to append as ?ver=
132 """
132 """
133 request = get_current_request()
133 request = get_current_request()
134 query = {}
134 query = {}
135 query.update(kwargs)
135 query.update(kwargs)
136 if ver:
136 if ver:
137 query = {'ver': ver}
137 query = {'ver': ver}
138 return request.static_path(
138 return request.static_path(
139 'rhodecode:public/{}'.format(path), _query=query)
139 'rhodecode:public/{}'.format(path), _query=query)
140
140
141
141
142 default_html_escape_table = {
142 default_html_escape_table = {
143 ord('&'): u'&amp;',
143 ord('&'): u'&amp;',
144 ord('<'): u'&lt;',
144 ord('<'): u'&lt;',
145 ord('>'): u'&gt;',
145 ord('>'): u'&gt;',
146 ord('"'): u'&quot;',
146 ord('"'): u'&quot;',
147 ord("'"): u'&#39;',
147 ord("'"): u'&#39;',
148 }
148 }
149
149
150
150
151 def html_escape(text, html_escape_table=default_html_escape_table):
151 def html_escape(text, html_escape_table=default_html_escape_table):
152 """Produce entities within text."""
152 """Produce entities within text."""
153 return text.translate(html_escape_table)
153 return text.translate(html_escape_table)
154
154
155
155
156 def chop_at_smart(s, sub, inclusive=False, suffix_if_chopped=None):
156 def chop_at_smart(s, sub, inclusive=False, suffix_if_chopped=None):
157 """
157 """
158 Truncate string ``s`` at the first occurrence of ``sub``.
158 Truncate string ``s`` at the first occurrence of ``sub``.
159
159
160 If ``inclusive`` is true, truncate just after ``sub`` rather than at it.
160 If ``inclusive`` is true, truncate just after ``sub`` rather than at it.
161 """
161 """
162 suffix_if_chopped = suffix_if_chopped or ''
162 suffix_if_chopped = suffix_if_chopped or ''
163 pos = s.find(sub)
163 pos = s.find(sub)
164 if pos == -1:
164 if pos == -1:
165 return s
165 return s
166
166
167 if inclusive:
167 if inclusive:
168 pos += len(sub)
168 pos += len(sub)
169
169
170 chopped = s[:pos]
170 chopped = s[:pos]
171 left = s[pos:].strip()
171 left = s[pos:].strip()
172
172
173 if left and suffix_if_chopped:
173 if left and suffix_if_chopped:
174 chopped += suffix_if_chopped
174 chopped += suffix_if_chopped
175
175
176 return chopped
176 return chopped
177
177
178
178
179 def shorter(text, size=20):
179 def shorter(text, size=20):
180 postfix = '...'
180 postfix = '...'
181 if len(text) > size:
181 if len(text) > size:
182 return text[:size - len(postfix)] + postfix
182 return text[:size - len(postfix)] + postfix
183 return text
183 return text
184
184
185
185
186 def _reset(name, value=None, id=NotGiven, type="reset", **attrs):
186 def _reset(name, value=None, id=NotGiven, type="reset", **attrs):
187 """
187 """
188 Reset button
188 Reset button
189 """
189 """
190 _set_input_attrs(attrs, type, name, value)
190 _set_input_attrs(attrs, type, name, value)
191 _set_id_attr(attrs, id, name)
191 _set_id_attr(attrs, id, name)
192 convert_boolean_attrs(attrs, ["disabled"])
192 convert_boolean_attrs(attrs, ["disabled"])
193 return HTML.input(**attrs)
193 return HTML.input(**attrs)
194
194
195 reset = _reset
195 reset = _reset
196 safeid = _make_safe_id_component
196 safeid = _make_safe_id_component
197
197
198
198
199 def branding(name, length=40):
199 def branding(name, length=40):
200 return truncate(name, length, indicator="")
200 return truncate(name, length, indicator="")
201
201
202
202
203 def FID(raw_id, path):
203 def FID(raw_id, path):
204 """
204 """
205 Creates a unique ID for filenode based on it's hash of path and commit
205 Creates a unique ID for filenode based on it's hash of path and commit
206 it's safe to use in urls
206 it's safe to use in urls
207
207
208 :param raw_id:
208 :param raw_id:
209 :param path:
209 :param path:
210 """
210 """
211
211
212 return 'c-%s-%s' % (short_id(raw_id), md5_safe(path)[:12])
212 return 'c-%s-%s' % (short_id(raw_id), md5_safe(path)[:12])
213
213
214
214
215 class _GetError(object):
215 class _GetError(object):
216 """Get error from form_errors, and represent it as span wrapped error
216 """Get error from form_errors, and represent it as span wrapped error
217 message
217 message
218
218
219 :param field_name: field to fetch errors for
219 :param field_name: field to fetch errors for
220 :param form_errors: form errors dict
220 :param form_errors: form errors dict
221 """
221 """
222
222
223 def __call__(self, field_name, form_errors):
223 def __call__(self, field_name, form_errors):
224 tmpl = """<span class="error_msg">%s</span>"""
224 tmpl = """<span class="error_msg">%s</span>"""
225 if form_errors and field_name in form_errors:
225 if form_errors and field_name in form_errors:
226 return literal(tmpl % form_errors.get(field_name))
226 return literal(tmpl % form_errors.get(field_name))
227
227
228 get_error = _GetError()
228 get_error = _GetError()
229
229
230
230
231 class _ToolTip(object):
231 class _ToolTip(object):
232
232
233 def __call__(self, tooltip_title, trim_at=50):
233 def __call__(self, tooltip_title, trim_at=50):
234 """
234 """
235 Special function just to wrap our text into nice formatted
235 Special function just to wrap our text into nice formatted
236 autowrapped text
236 autowrapped text
237
237
238 :param tooltip_title:
238 :param tooltip_title:
239 """
239 """
240 tooltip_title = escape(tooltip_title)
240 tooltip_title = escape(tooltip_title)
241 tooltip_title = tooltip_title.replace('<', '&lt;').replace('>', '&gt;')
241 tooltip_title = tooltip_title.replace('<', '&lt;').replace('>', '&gt;')
242 return tooltip_title
242 return tooltip_title
243 tooltip = _ToolTip()
243 tooltip = _ToolTip()
244
244
245
245
246 def files_breadcrumbs(repo_name, commit_id, file_path):
246 def files_breadcrumbs(repo_name, commit_id, file_path):
247 if isinstance(file_path, str):
247 if isinstance(file_path, str):
248 file_path = safe_unicode(file_path)
248 file_path = safe_unicode(file_path)
249
249
250 # TODO: johbo: Is this always a url like path, or is this operating
250 # TODO: johbo: Is this always a url like path, or is this operating
251 # system dependent?
251 # system dependent?
252 path_segments = file_path.split('/')
252 path_segments = file_path.split('/')
253
253
254 repo_name_html = escape(repo_name)
254 repo_name_html = escape(repo_name)
255 if len(path_segments) == 1 and path_segments[0] == '':
255 if len(path_segments) == 1 and path_segments[0] == '':
256 url_segments = [repo_name_html]
256 url_segments = [repo_name_html]
257 else:
257 else:
258 url_segments = [
258 url_segments = [
259 link_to(
259 link_to(
260 repo_name_html,
260 repo_name_html,
261 route_path(
261 route_path(
262 'repo_files',
262 'repo_files',
263 repo_name=repo_name,
263 repo_name=repo_name,
264 commit_id=commit_id,
264 commit_id=commit_id,
265 f_path=''),
265 f_path=''),
266 class_='pjax-link')]
266 class_='pjax-link')]
267
267
268 last_cnt = len(path_segments) - 1
268 last_cnt = len(path_segments) - 1
269 for cnt, segment in enumerate(path_segments):
269 for cnt, segment in enumerate(path_segments):
270 if not segment:
270 if not segment:
271 continue
271 continue
272 segment_html = escape(segment)
272 segment_html = escape(segment)
273
273
274 if cnt != last_cnt:
274 if cnt != last_cnt:
275 url_segments.append(
275 url_segments.append(
276 link_to(
276 link_to(
277 segment_html,
277 segment_html,
278 route_path(
278 route_path(
279 'repo_files',
279 'repo_files',
280 repo_name=repo_name,
280 repo_name=repo_name,
281 commit_id=commit_id,
281 commit_id=commit_id,
282 f_path='/'.join(path_segments[:cnt + 1])),
282 f_path='/'.join(path_segments[:cnt + 1])),
283 class_='pjax-link'))
283 class_='pjax-link'))
284 else:
284 else:
285 url_segments.append(segment_html)
285 url_segments.append(segment_html)
286
286
287 return literal('/'.join(url_segments))
287 return literal('/'.join(url_segments))
288
288
289
289
290 class CodeHtmlFormatter(HtmlFormatter):
290 class CodeHtmlFormatter(HtmlFormatter):
291 """
291 """
292 My code Html Formatter for source codes
292 My code Html Formatter for source codes
293 """
293 """
294
294
295 def wrap(self, source, outfile):
295 def wrap(self, source, outfile):
296 return self._wrap_div(self._wrap_pre(self._wrap_code(source)))
296 return self._wrap_div(self._wrap_pre(self._wrap_code(source)))
297
297
298 def _wrap_code(self, source):
298 def _wrap_code(self, source):
299 for cnt, it in enumerate(source):
299 for cnt, it in enumerate(source):
300 i, t = it
300 i, t = it
301 t = '<div id="L%s">%s</div>' % (cnt + 1, t)
301 t = '<div id="L%s">%s</div>' % (cnt + 1, t)
302 yield i, t
302 yield i, t
303
303
304 def _wrap_tablelinenos(self, inner):
304 def _wrap_tablelinenos(self, inner):
305 dummyoutfile = StringIO.StringIO()
305 dummyoutfile = StringIO.StringIO()
306 lncount = 0
306 lncount = 0
307 for t, line in inner:
307 for t, line in inner:
308 if t:
308 if t:
309 lncount += 1
309 lncount += 1
310 dummyoutfile.write(line)
310 dummyoutfile.write(line)
311
311
312 fl = self.linenostart
312 fl = self.linenostart
313 mw = len(str(lncount + fl - 1))
313 mw = len(str(lncount + fl - 1))
314 sp = self.linenospecial
314 sp = self.linenospecial
315 st = self.linenostep
315 st = self.linenostep
316 la = self.lineanchors
316 la = self.lineanchors
317 aln = self.anchorlinenos
317 aln = self.anchorlinenos
318 nocls = self.noclasses
318 nocls = self.noclasses
319 if sp:
319 if sp:
320 lines = []
320 lines = []
321
321
322 for i in range(fl, fl + lncount):
322 for i in range(fl, fl + lncount):
323 if i % st == 0:
323 if i % st == 0:
324 if i % sp == 0:
324 if i % sp == 0:
325 if aln:
325 if aln:
326 lines.append('<a href="#%s%d" class="special">%*d</a>' %
326 lines.append('<a href="#%s%d" class="special">%*d</a>' %
327 (la, i, mw, i))
327 (la, i, mw, i))
328 else:
328 else:
329 lines.append('<span class="special">%*d</span>' % (mw, i))
329 lines.append('<span class="special">%*d</span>' % (mw, i))
330 else:
330 else:
331 if aln:
331 if aln:
332 lines.append('<a href="#%s%d">%*d</a>' % (la, i, mw, i))
332 lines.append('<a href="#%s%d">%*d</a>' % (la, i, mw, i))
333 else:
333 else:
334 lines.append('%*d' % (mw, i))
334 lines.append('%*d' % (mw, i))
335 else:
335 else:
336 lines.append('')
336 lines.append('')
337 ls = '\n'.join(lines)
337 ls = '\n'.join(lines)
338 else:
338 else:
339 lines = []
339 lines = []
340 for i in range(fl, fl + lncount):
340 for i in range(fl, fl + lncount):
341 if i % st == 0:
341 if i % st == 0:
342 if aln:
342 if aln:
343 lines.append('<a href="#%s%d">%*d</a>' % (la, i, mw, i))
343 lines.append('<a href="#%s%d">%*d</a>' % (la, i, mw, i))
344 else:
344 else:
345 lines.append('%*d' % (mw, i))
345 lines.append('%*d' % (mw, i))
346 else:
346 else:
347 lines.append('')
347 lines.append('')
348 ls = '\n'.join(lines)
348 ls = '\n'.join(lines)
349
349
350 # in case you wonder about the seemingly redundant <div> here: since the
350 # in case you wonder about the seemingly redundant <div> here: since the
351 # content in the other cell also is wrapped in a div, some browsers in
351 # content in the other cell also is wrapped in a div, some browsers in
352 # some configurations seem to mess up the formatting...
352 # some configurations seem to mess up the formatting...
353 if nocls:
353 if nocls:
354 yield 0, ('<table class="%stable">' % self.cssclass +
354 yield 0, ('<table class="%stable">' % self.cssclass +
355 '<tr><td><div class="linenodiv" '
355 '<tr><td><div class="linenodiv" '
356 'style="background-color: #f0f0f0; padding-right: 10px">'
356 'style="background-color: #f0f0f0; padding-right: 10px">'
357 '<pre style="line-height: 125%">' +
357 '<pre style="line-height: 125%">' +
358 ls + '</pre></div></td><td id="hlcode" class="code">')
358 ls + '</pre></div></td><td id="hlcode" class="code">')
359 else:
359 else:
360 yield 0, ('<table class="%stable">' % self.cssclass +
360 yield 0, ('<table class="%stable">' % self.cssclass +
361 '<tr><td class="linenos"><div class="linenodiv"><pre>' +
361 '<tr><td class="linenos"><div class="linenodiv"><pre>' +
362 ls + '</pre></div></td><td id="hlcode" class="code">')
362 ls + '</pre></div></td><td id="hlcode" class="code">')
363 yield 0, dummyoutfile.getvalue()
363 yield 0, dummyoutfile.getvalue()
364 yield 0, '</td></tr></table>'
364 yield 0, '</td></tr></table>'
365
365
366
366
367 class SearchContentCodeHtmlFormatter(CodeHtmlFormatter):
367 class SearchContentCodeHtmlFormatter(CodeHtmlFormatter):
368 def __init__(self, **kw):
368 def __init__(self, **kw):
369 # only show these line numbers if set
369 # only show these line numbers if set
370 self.only_lines = kw.pop('only_line_numbers', [])
370 self.only_lines = kw.pop('only_line_numbers', [])
371 self.query_terms = kw.pop('query_terms', [])
371 self.query_terms = kw.pop('query_terms', [])
372 self.max_lines = kw.pop('max_lines', 5)
372 self.max_lines = kw.pop('max_lines', 5)
373 self.line_context = kw.pop('line_context', 3)
373 self.line_context = kw.pop('line_context', 3)
374 self.url = kw.pop('url', None)
374 self.url = kw.pop('url', None)
375
375
376 super(CodeHtmlFormatter, self).__init__(**kw)
376 super(CodeHtmlFormatter, self).__init__(**kw)
377
377
378 def _wrap_code(self, source):
378 def _wrap_code(self, source):
379 for cnt, it in enumerate(source):
379 for cnt, it in enumerate(source):
380 i, t = it
380 i, t = it
381 t = '<pre>%s</pre>' % t
381 t = '<pre>%s</pre>' % t
382 yield i, t
382 yield i, t
383
383
384 def _wrap_tablelinenos(self, inner):
384 def _wrap_tablelinenos(self, inner):
385 yield 0, '<table class="code-highlight %stable">' % self.cssclass
385 yield 0, '<table class="code-highlight %stable">' % self.cssclass
386
386
387 last_shown_line_number = 0
387 last_shown_line_number = 0
388 current_line_number = 1
388 current_line_number = 1
389
389
390 for t, line in inner:
390 for t, line in inner:
391 if not t:
391 if not t:
392 yield t, line
392 yield t, line
393 continue
393 continue
394
394
395 if current_line_number in self.only_lines:
395 if current_line_number in self.only_lines:
396 if last_shown_line_number + 1 != current_line_number:
396 if last_shown_line_number + 1 != current_line_number:
397 yield 0, '<tr>'
397 yield 0, '<tr>'
398 yield 0, '<td class="line">...</td>'
398 yield 0, '<td class="line">...</td>'
399 yield 0, '<td id="hlcode" class="code"></td>'
399 yield 0, '<td id="hlcode" class="code"></td>'
400 yield 0, '</tr>'
400 yield 0, '</tr>'
401
401
402 yield 0, '<tr>'
402 yield 0, '<tr>'
403 if self.url:
403 if self.url:
404 yield 0, '<td class="line"><a href="%s#L%i">%i</a></td>' % (
404 yield 0, '<td class="line"><a href="%s#L%i">%i</a></td>' % (
405 self.url, current_line_number, current_line_number)
405 self.url, current_line_number, current_line_number)
406 else:
406 else:
407 yield 0, '<td class="line"><a href="">%i</a></td>' % (
407 yield 0, '<td class="line"><a href="">%i</a></td>' % (
408 current_line_number)
408 current_line_number)
409 yield 0, '<td id="hlcode" class="code">' + line + '</td>'
409 yield 0, '<td id="hlcode" class="code">' + line + '</td>'
410 yield 0, '</tr>'
410 yield 0, '</tr>'
411
411
412 last_shown_line_number = current_line_number
412 last_shown_line_number = current_line_number
413
413
414 current_line_number += 1
414 current_line_number += 1
415
415
416
416
417 yield 0, '</table>'
417 yield 0, '</table>'
418
418
419
419
420 def extract_phrases(text_query):
420 def extract_phrases(text_query):
421 """
421 """
422 Extracts phrases from search term string making sure phrases
422 Extracts phrases from search term string making sure phrases
423 contained in double quotes are kept together - and discarding empty values
423 contained in double quotes are kept together - and discarding empty values
424 or fully whitespace values eg.
424 or fully whitespace values eg.
425
425
426 'some text "a phrase" more' => ['some', 'text', 'a phrase', 'more']
426 'some text "a phrase" more' => ['some', 'text', 'a phrase', 'more']
427
427
428 """
428 """
429
429
430 in_phrase = False
430 in_phrase = False
431 buf = ''
431 buf = ''
432 phrases = []
432 phrases = []
433 for char in text_query:
433 for char in text_query:
434 if in_phrase:
434 if in_phrase:
435 if char == '"': # end phrase
435 if char == '"': # end phrase
436 phrases.append(buf)
436 phrases.append(buf)
437 buf = ''
437 buf = ''
438 in_phrase = False
438 in_phrase = False
439 continue
439 continue
440 else:
440 else:
441 buf += char
441 buf += char
442 continue
442 continue
443 else:
443 else:
444 if char == '"': # start phrase
444 if char == '"': # start phrase
445 in_phrase = True
445 in_phrase = True
446 phrases.append(buf)
446 phrases.append(buf)
447 buf = ''
447 buf = ''
448 continue
448 continue
449 elif char == ' ':
449 elif char == ' ':
450 phrases.append(buf)
450 phrases.append(buf)
451 buf = ''
451 buf = ''
452 continue
452 continue
453 else:
453 else:
454 buf += char
454 buf += char
455
455
456 phrases.append(buf)
456 phrases.append(buf)
457 phrases = [phrase.strip() for phrase in phrases if phrase.strip()]
457 phrases = [phrase.strip() for phrase in phrases if phrase.strip()]
458 return phrases
458 return phrases
459
459
460
460
461 def get_matching_offsets(text, phrases):
461 def get_matching_offsets(text, phrases):
462 """
462 """
463 Returns a list of string offsets in `text` that the list of `terms` match
463 Returns a list of string offsets in `text` that the list of `terms` match
464
464
465 >>> get_matching_offsets('some text here', ['some', 'here'])
465 >>> get_matching_offsets('some text here', ['some', 'here'])
466 [(0, 4), (10, 14)]
466 [(0, 4), (10, 14)]
467
467
468 """
468 """
469 offsets = []
469 offsets = []
470 for phrase in phrases:
470 for phrase in phrases:
471 for match in re.finditer(phrase, text):
471 for match in re.finditer(phrase, text):
472 offsets.append((match.start(), match.end()))
472 offsets.append((match.start(), match.end()))
473
473
474 return offsets
474 return offsets
475
475
476
476
477 def normalize_text_for_matching(x):
477 def normalize_text_for_matching(x):
478 """
478 """
479 Replaces all non alnum characters to spaces and lower cases the string,
479 Replaces all non alnum characters to spaces and lower cases the string,
480 useful for comparing two text strings without punctuation
480 useful for comparing two text strings without punctuation
481 """
481 """
482 return re.sub(r'[^\w]', ' ', x.lower())
482 return re.sub(r'[^\w]', ' ', x.lower())
483
483
484
484
485 def get_matching_line_offsets(lines, terms):
485 def get_matching_line_offsets(lines, terms):
486 """ Return a set of `lines` indices (starting from 1) matching a
486 """ Return a set of `lines` indices (starting from 1) matching a
487 text search query, along with `context` lines above/below matching lines
487 text search query, along with `context` lines above/below matching lines
488
488
489 :param lines: list of strings representing lines
489 :param lines: list of strings representing lines
490 :param terms: search term string to match in lines eg. 'some text'
490 :param terms: search term string to match in lines eg. 'some text'
491 :param context: number of lines above/below a matching line to add to result
491 :param context: number of lines above/below a matching line to add to result
492 :param max_lines: cut off for lines of interest
492 :param max_lines: cut off for lines of interest
493 eg.
493 eg.
494
494
495 text = '''
495 text = '''
496 words words words
496 words words words
497 words words words
497 words words words
498 some text some
498 some text some
499 words words words
499 words words words
500 words words words
500 words words words
501 text here what
501 text here what
502 '''
502 '''
503 get_matching_line_offsets(text, 'text', context=1)
503 get_matching_line_offsets(text, 'text', context=1)
504 {3: [(5, 9)], 6: [(0, 4)]]
504 {3: [(5, 9)], 6: [(0, 4)]]
505
505
506 """
506 """
507 matching_lines = {}
507 matching_lines = {}
508 phrases = [normalize_text_for_matching(phrase)
508 phrases = [normalize_text_for_matching(phrase)
509 for phrase in extract_phrases(terms)]
509 for phrase in extract_phrases(terms)]
510
510
511 for line_index, line in enumerate(lines, start=1):
511 for line_index, line in enumerate(lines, start=1):
512 match_offsets = get_matching_offsets(
512 match_offsets = get_matching_offsets(
513 normalize_text_for_matching(line), phrases)
513 normalize_text_for_matching(line), phrases)
514 if match_offsets:
514 if match_offsets:
515 matching_lines[line_index] = match_offsets
515 matching_lines[line_index] = match_offsets
516
516
517 return matching_lines
517 return matching_lines
518
518
519
519
520 def hsv_to_rgb(h, s, v):
520 def hsv_to_rgb(h, s, v):
521 """ Convert hsv color values to rgb """
521 """ Convert hsv color values to rgb """
522
522
523 if s == 0.0:
523 if s == 0.0:
524 return v, v, v
524 return v, v, v
525 i = int(h * 6.0) # XXX assume int() truncates!
525 i = int(h * 6.0) # XXX assume int() truncates!
526 f = (h * 6.0) - i
526 f = (h * 6.0) - i
527 p = v * (1.0 - s)
527 p = v * (1.0 - s)
528 q = v * (1.0 - s * f)
528 q = v * (1.0 - s * f)
529 t = v * (1.0 - s * (1.0 - f))
529 t = v * (1.0 - s * (1.0 - f))
530 i = i % 6
530 i = i % 6
531 if i == 0:
531 if i == 0:
532 return v, t, p
532 return v, t, p
533 if i == 1:
533 if i == 1:
534 return q, v, p
534 return q, v, p
535 if i == 2:
535 if i == 2:
536 return p, v, t
536 return p, v, t
537 if i == 3:
537 if i == 3:
538 return p, q, v
538 return p, q, v
539 if i == 4:
539 if i == 4:
540 return t, p, v
540 return t, p, v
541 if i == 5:
541 if i == 5:
542 return v, p, q
542 return v, p, q
543
543
544
544
545 def unique_color_generator(n=10000, saturation=0.10, lightness=0.95):
545 def unique_color_generator(n=10000, saturation=0.10, lightness=0.95):
546 """
546 """
547 Generator for getting n of evenly distributed colors using
547 Generator for getting n of evenly distributed colors using
548 hsv color and golden ratio. It always return same order of colors
548 hsv color and golden ratio. It always return same order of colors
549
549
550 :param n: number of colors to generate
550 :param n: number of colors to generate
551 :param saturation: saturation of returned colors
551 :param saturation: saturation of returned colors
552 :param lightness: lightness of returned colors
552 :param lightness: lightness of returned colors
553 :returns: RGB tuple
553 :returns: RGB tuple
554 """
554 """
555
555
556 golden_ratio = 0.618033988749895
556 golden_ratio = 0.618033988749895
557 h = 0.22717784590367374
557 h = 0.22717784590367374
558
558
559 for _ in xrange(n):
559 for _ in xrange(n):
560 h += golden_ratio
560 h += golden_ratio
561 h %= 1
561 h %= 1
562 HSV_tuple = [h, saturation, lightness]
562 HSV_tuple = [h, saturation, lightness]
563 RGB_tuple = hsv_to_rgb(*HSV_tuple)
563 RGB_tuple = hsv_to_rgb(*HSV_tuple)
564 yield map(lambda x: str(int(x * 256)), RGB_tuple)
564 yield map(lambda x: str(int(x * 256)), RGB_tuple)
565
565
566
566
567 def color_hasher(n=10000, saturation=0.10, lightness=0.95):
567 def color_hasher(n=10000, saturation=0.10, lightness=0.95):
568 """
568 """
569 Returns a function which when called with an argument returns a unique
569 Returns a function which when called with an argument returns a unique
570 color for that argument, eg.
570 color for that argument, eg.
571
571
572 :param n: number of colors to generate
572 :param n: number of colors to generate
573 :param saturation: saturation of returned colors
573 :param saturation: saturation of returned colors
574 :param lightness: lightness of returned colors
574 :param lightness: lightness of returned colors
575 :returns: css RGB string
575 :returns: css RGB string
576
576
577 >>> color_hash = color_hasher()
577 >>> color_hash = color_hasher()
578 >>> color_hash('hello')
578 >>> color_hash('hello')
579 'rgb(34, 12, 59)'
579 'rgb(34, 12, 59)'
580 >>> color_hash('hello')
580 >>> color_hash('hello')
581 'rgb(34, 12, 59)'
581 'rgb(34, 12, 59)'
582 >>> color_hash('other')
582 >>> color_hash('other')
583 'rgb(90, 224, 159)'
583 'rgb(90, 224, 159)'
584 """
584 """
585
585
586 color_dict = {}
586 color_dict = {}
587 cgenerator = unique_color_generator(
587 cgenerator = unique_color_generator(
588 saturation=saturation, lightness=lightness)
588 saturation=saturation, lightness=lightness)
589
589
590 def get_color_string(thing):
590 def get_color_string(thing):
591 if thing in color_dict:
591 if thing in color_dict:
592 col = color_dict[thing]
592 col = color_dict[thing]
593 else:
593 else:
594 col = color_dict[thing] = cgenerator.next()
594 col = color_dict[thing] = cgenerator.next()
595 return "rgb(%s)" % (', '.join(col))
595 return "rgb(%s)" % (', '.join(col))
596
596
597 return get_color_string
597 return get_color_string
598
598
599
599
600 def get_lexer_safe(mimetype=None, filepath=None):
600 def get_lexer_safe(mimetype=None, filepath=None):
601 """
601 """
602 Tries to return a relevant pygments lexer using mimetype/filepath name,
602 Tries to return a relevant pygments lexer using mimetype/filepath name,
603 defaulting to plain text if none could be found
603 defaulting to plain text if none could be found
604 """
604 """
605 lexer = None
605 lexer = None
606 try:
606 try:
607 if mimetype:
607 if mimetype:
608 lexer = get_lexer_for_mimetype(mimetype)
608 lexer = get_lexer_for_mimetype(mimetype)
609 if not lexer:
609 if not lexer:
610 lexer = get_lexer_for_filename(filepath)
610 lexer = get_lexer_for_filename(filepath)
611 except pygments.util.ClassNotFound:
611 except pygments.util.ClassNotFound:
612 pass
612 pass
613
613
614 if not lexer:
614 if not lexer:
615 lexer = get_lexer_by_name('text')
615 lexer = get_lexer_by_name('text')
616
616
617 return lexer
617 return lexer
618
618
619
619
620 def get_lexer_for_filenode(filenode):
620 def get_lexer_for_filenode(filenode):
621 lexer = get_custom_lexer(filenode.extension) or filenode.lexer
621 lexer = get_custom_lexer(filenode.extension) or filenode.lexer
622 return lexer
622 return lexer
623
623
624
624
625 def pygmentize(filenode, **kwargs):
625 def pygmentize(filenode, **kwargs):
626 """
626 """
627 pygmentize function using pygments
627 pygmentize function using pygments
628
628
629 :param filenode:
629 :param filenode:
630 """
630 """
631 lexer = get_lexer_for_filenode(filenode)
631 lexer = get_lexer_for_filenode(filenode)
632 return literal(code_highlight(filenode.content, lexer,
632 return literal(code_highlight(filenode.content, lexer,
633 CodeHtmlFormatter(**kwargs)))
633 CodeHtmlFormatter(**kwargs)))
634
634
635
635
636 def is_following_repo(repo_name, user_id):
636 def is_following_repo(repo_name, user_id):
637 from rhodecode.model.scm import ScmModel
637 from rhodecode.model.scm import ScmModel
638 return ScmModel().is_following_repo(repo_name, user_id)
638 return ScmModel().is_following_repo(repo_name, user_id)
639
639
640
640
641 class _Message(object):
641 class _Message(object):
642 """A message returned by ``Flash.pop_messages()``.
642 """A message returned by ``Flash.pop_messages()``.
643
643
644 Converting the message to a string returns the message text. Instances
644 Converting the message to a string returns the message text. Instances
645 also have the following attributes:
645 also have the following attributes:
646
646
647 * ``message``: the message text.
647 * ``message``: the message text.
648 * ``category``: the category specified when the message was created.
648 * ``category``: the category specified when the message was created.
649 """
649 """
650
650
651 def __init__(self, category, message):
651 def __init__(self, category, message):
652 self.category = category
652 self.category = category
653 self.message = message
653 self.message = message
654
654
655 def __str__(self):
655 def __str__(self):
656 return self.message
656 return self.message
657
657
658 __unicode__ = __str__
658 __unicode__ = __str__
659
659
660 def __html__(self):
660 def __html__(self):
661 return escape(safe_unicode(self.message))
661 return escape(safe_unicode(self.message))
662
662
663
663
664 class Flash(_Flash):
664 class Flash(_Flash):
665
665
666 def pop_messages(self, request=None):
666 def pop_messages(self, request=None):
667 """Return all accumulated messages and delete them from the session.
667 """Return all accumulated messages and delete them from the session.
668
668
669 The return value is a list of ``Message`` objects.
669 The return value is a list of ``Message`` objects.
670 """
670 """
671 messages = []
671 messages = []
672
672
673 if request:
673 if request:
674 session = request.session
674 session = request.session
675 else:
675 else:
676 from pylons import session
676 from pylons import session
677
677
678 # Pop the 'old' pylons flash messages. They are tuples of the form
678 # Pop the 'old' pylons flash messages. They are tuples of the form
679 # (category, message)
679 # (category, message)
680 for cat, msg in session.pop(self.session_key, []):
680 for cat, msg in session.pop(self.session_key, []):
681 messages.append(_Message(cat, msg))
681 messages.append(_Message(cat, msg))
682
682
683 # Pop the 'new' pyramid flash messages for each category as list
683 # Pop the 'new' pyramid flash messages for each category as list
684 # of strings.
684 # of strings.
685 for cat in self.categories:
685 for cat in self.categories:
686 for msg in session.pop_flash(queue=cat):
686 for msg in session.pop_flash(queue=cat):
687 messages.append(_Message(cat, msg))
687 messages.append(_Message(cat, msg))
688 # Map messages from the default queue to the 'notice' category.
688 # Map messages from the default queue to the 'notice' category.
689 for msg in session.pop_flash():
689 for msg in session.pop_flash():
690 messages.append(_Message('notice', msg))
690 messages.append(_Message('notice', msg))
691
691
692 session.save()
692 session.save()
693 return messages
693 return messages
694
694
695 def json_alerts(self, request=None):
695 def json_alerts(self, request=None):
696 payloads = []
696 payloads = []
697 messages = flash.pop_messages(request=request)
697 messages = flash.pop_messages(request=request)
698 if messages:
698 if messages:
699 for message in messages:
699 for message in messages:
700 subdata = {}
700 subdata = {}
701 if hasattr(message.message, 'rsplit'):
701 if hasattr(message.message, 'rsplit'):
702 flash_data = message.message.rsplit('|DELIM|', 1)
702 flash_data = message.message.rsplit('|DELIM|', 1)
703 org_message = flash_data[0]
703 org_message = flash_data[0]
704 if len(flash_data) > 1:
704 if len(flash_data) > 1:
705 subdata = json.loads(flash_data[1])
705 subdata = json.loads(flash_data[1])
706 else:
706 else:
707 org_message = message.message
707 org_message = message.message
708 payloads.append({
708 payloads.append({
709 'message': {
709 'message': {
710 'message': u'{}'.format(org_message),
710 'message': u'{}'.format(org_message),
711 'level': message.category,
711 'level': message.category,
712 'force': True,
712 'force': True,
713 'subdata': subdata
713 'subdata': subdata
714 }
714 }
715 })
715 })
716 return json.dumps(payloads)
716 return json.dumps(payloads)
717
717
718 flash = Flash()
718 flash = Flash()
719
719
720 #==============================================================================
720 #==============================================================================
721 # SCM FILTERS available via h.
721 # SCM FILTERS available via h.
722 #==============================================================================
722 #==============================================================================
723 from rhodecode.lib.vcs.utils import author_name, author_email
723 from rhodecode.lib.vcs.utils import author_name, author_email
724 from rhodecode.lib.utils2 import credentials_filter, age as _age
724 from rhodecode.lib.utils2 import credentials_filter, age as _age
725 from rhodecode.model.db import User, ChangesetStatus
725 from rhodecode.model.db import User, ChangesetStatus
726
726
727 age = _age
727 age = _age
728 capitalize = lambda x: x.capitalize()
728 capitalize = lambda x: x.capitalize()
729 email = author_email
729 email = author_email
730 short_id = lambda x: x[:12]
730 short_id = lambda x: x[:12]
731 hide_credentials = lambda x: ''.join(credentials_filter(x))
731 hide_credentials = lambda x: ''.join(credentials_filter(x))
732
732
733
733
734 def age_component(datetime_iso, value=None, time_is_local=False):
734 def age_component(datetime_iso, value=None, time_is_local=False):
735 title = value or format_date(datetime_iso)
735 title = value or format_date(datetime_iso)
736 tzinfo = '+00:00'
736 tzinfo = '+00:00'
737
737
738 # detect if we have a timezone info, otherwise, add it
738 # detect if we have a timezone info, otherwise, add it
739 if isinstance(datetime_iso, datetime) and not datetime_iso.tzinfo:
739 if isinstance(datetime_iso, datetime) and not datetime_iso.tzinfo:
740 if time_is_local:
740 if time_is_local:
741 tzinfo = time.strftime("+%H:%M",
741 tzinfo = time.strftime("+%H:%M",
742 time.gmtime(
742 time.gmtime(
743 (datetime.now() - datetime.utcnow()).seconds + 1
743 (datetime.now() - datetime.utcnow()).seconds + 1
744 )
744 )
745 )
745 )
746
746
747 return literal(
747 return literal(
748 '<time class="timeago tooltip" '
748 '<time class="timeago tooltip" '
749 'title="{1}{2}" datetime="{0}{2}">{1}</time>'.format(
749 'title="{1}{2}" datetime="{0}{2}">{1}</time>'.format(
750 datetime_iso, title, tzinfo))
750 datetime_iso, title, tzinfo))
751
751
752
752
753 def _shorten_commit_id(commit_id):
753 def _shorten_commit_id(commit_id):
754 from rhodecode import CONFIG
754 from rhodecode import CONFIG
755 def_len = safe_int(CONFIG.get('rhodecode_show_sha_length', 12))
755 def_len = safe_int(CONFIG.get('rhodecode_show_sha_length', 12))
756 return commit_id[:def_len]
756 return commit_id[:def_len]
757
757
758
758
759 def show_id(commit):
759 def show_id(commit):
760 """
760 """
761 Configurable function that shows ID
761 Configurable function that shows ID
762 by default it's r123:fffeeefffeee
762 by default it's r123:fffeeefffeee
763
763
764 :param commit: commit instance
764 :param commit: commit instance
765 """
765 """
766 from rhodecode import CONFIG
766 from rhodecode import CONFIG
767 show_idx = str2bool(CONFIG.get('rhodecode_show_revision_number', True))
767 show_idx = str2bool(CONFIG.get('rhodecode_show_revision_number', True))
768
768
769 raw_id = _shorten_commit_id(commit.raw_id)
769 raw_id = _shorten_commit_id(commit.raw_id)
770 if show_idx:
770 if show_idx:
771 return 'r%s:%s' % (commit.idx, raw_id)
771 return 'r%s:%s' % (commit.idx, raw_id)
772 else:
772 else:
773 return '%s' % (raw_id, )
773 return '%s' % (raw_id, )
774
774
775
775
776 def format_date(date):
776 def format_date(date):
777 """
777 """
778 use a standardized formatting for dates used in RhodeCode
778 use a standardized formatting for dates used in RhodeCode
779
779
780 :param date: date/datetime object
780 :param date: date/datetime object
781 :return: formatted date
781 :return: formatted date
782 """
782 """
783
783
784 if date:
784 if date:
785 _fmt = "%a, %d %b %Y %H:%M:%S"
785 _fmt = "%a, %d %b %Y %H:%M:%S"
786 return safe_unicode(date.strftime(_fmt))
786 return safe_unicode(date.strftime(_fmt))
787
787
788 return u""
788 return u""
789
789
790
790
791 class _RepoChecker(object):
791 class _RepoChecker(object):
792
792
793 def __init__(self, backend_alias):
793 def __init__(self, backend_alias):
794 self._backend_alias = backend_alias
794 self._backend_alias = backend_alias
795
795
796 def __call__(self, repository):
796 def __call__(self, repository):
797 if hasattr(repository, 'alias'):
797 if hasattr(repository, 'alias'):
798 _type = repository.alias
798 _type = repository.alias
799 elif hasattr(repository, 'repo_type'):
799 elif hasattr(repository, 'repo_type'):
800 _type = repository.repo_type
800 _type = repository.repo_type
801 else:
801 else:
802 _type = repository
802 _type = repository
803 return _type == self._backend_alias
803 return _type == self._backend_alias
804
804
805 is_git = _RepoChecker('git')
805 is_git = _RepoChecker('git')
806 is_hg = _RepoChecker('hg')
806 is_hg = _RepoChecker('hg')
807 is_svn = _RepoChecker('svn')
807 is_svn = _RepoChecker('svn')
808
808
809
809
810 def get_repo_type_by_name(repo_name):
810 def get_repo_type_by_name(repo_name):
811 repo = Repository.get_by_repo_name(repo_name)
811 repo = Repository.get_by_repo_name(repo_name)
812 return repo.repo_type
812 return repo.repo_type
813
813
814
814
815 def is_svn_without_proxy(repository):
815 def is_svn_without_proxy(repository):
816 if is_svn(repository):
816 if is_svn(repository):
817 from rhodecode.model.settings import VcsSettingsModel
817 from rhodecode.model.settings import VcsSettingsModel
818 conf = VcsSettingsModel().get_ui_settings_as_config_obj()
818 conf = VcsSettingsModel().get_ui_settings_as_config_obj()
819 return not str2bool(conf.get('vcs_svn_proxy', 'http_requests_enabled'))
819 return not str2bool(conf.get('vcs_svn_proxy', 'http_requests_enabled'))
820 return False
820 return False
821
821
822
822
823 def discover_user(author):
823 def discover_user(author):
824 """
824 """
825 Tries to discover RhodeCode User based on the autho string. Author string
825 Tries to discover RhodeCode User based on the autho string. Author string
826 is typically `FirstName LastName <email@address.com>`
826 is typically `FirstName LastName <email@address.com>`
827 """
827 """
828
828
829 # if author is already an instance use it for extraction
829 # if author is already an instance use it for extraction
830 if isinstance(author, User):
830 if isinstance(author, User):
831 return author
831 return author
832
832
833 # Valid email in the attribute passed, see if they're in the system
833 # Valid email in the attribute passed, see if they're in the system
834 _email = author_email(author)
834 _email = author_email(author)
835 if _email != '':
835 if _email != '':
836 user = User.get_by_email(_email, case_insensitive=True, cache=True)
836 user = User.get_by_email(_email, case_insensitive=True, cache=True)
837 if user is not None:
837 if user is not None:
838 return user
838 return user
839
839
840 # Maybe it's a username, we try to extract it and fetch by username ?
840 # Maybe it's a username, we try to extract it and fetch by username ?
841 _author = author_name(author)
841 _author = author_name(author)
842 user = User.get_by_username(_author, case_insensitive=True, cache=True)
842 user = User.get_by_username(_author, case_insensitive=True, cache=True)
843 if user is not None:
843 if user is not None:
844 return user
844 return user
845
845
846 return None
846 return None
847
847
848
848
849 def email_or_none(author):
849 def email_or_none(author):
850 # extract email from the commit string
850 # extract email from the commit string
851 _email = author_email(author)
851 _email = author_email(author)
852
852
853 # If we have an email, use it, otherwise
853 # If we have an email, use it, otherwise
854 # see if it contains a username we can get an email from
854 # see if it contains a username we can get an email from
855 if _email != '':
855 if _email != '':
856 return _email
856 return _email
857 else:
857 else:
858 user = User.get_by_username(
858 user = User.get_by_username(
859 author_name(author), case_insensitive=True, cache=True)
859 author_name(author), case_insensitive=True, cache=True)
860
860
861 if user is not None:
861 if user is not None:
862 return user.email
862 return user.email
863
863
864 # No valid email, not a valid user in the system, none!
864 # No valid email, not a valid user in the system, none!
865 return None
865 return None
866
866
867
867
868 def link_to_user(author, length=0, **kwargs):
868 def link_to_user(author, length=0, **kwargs):
869 user = discover_user(author)
869 user = discover_user(author)
870 # user can be None, but if we have it already it means we can re-use it
870 # user can be None, but if we have it already it means we can re-use it
871 # in the person() function, so we save 1 intensive-query
871 # in the person() function, so we save 1 intensive-query
872 if user:
872 if user:
873 author = user
873 author = user
874
874
875 display_person = person(author, 'username_or_name_or_email')
875 display_person = person(author, 'username_or_name_or_email')
876 if length:
876 if length:
877 display_person = shorter(display_person, length)
877 display_person = shorter(display_person, length)
878
878
879 if user:
879 if user:
880 return link_to(
880 return link_to(
881 escape(display_person),
881 escape(display_person),
882 route_path('user_profile', username=user.username),
882 route_path('user_profile', username=user.username),
883 **kwargs)
883 **kwargs)
884 else:
884 else:
885 return escape(display_person)
885 return escape(display_person)
886
886
887
887
888 def person(author, show_attr="username_and_name"):
888 def person(author, show_attr="username_and_name"):
889 user = discover_user(author)
889 user = discover_user(author)
890 if user:
890 if user:
891 return getattr(user, show_attr)
891 return getattr(user, show_attr)
892 else:
892 else:
893 _author = author_name(author)
893 _author = author_name(author)
894 _email = email(author)
894 _email = email(author)
895 return _author or _email
895 return _author or _email
896
896
897
897
898 def author_string(email):
898 def author_string(email):
899 if email:
899 if email:
900 user = User.get_by_email(email, case_insensitive=True, cache=True)
900 user = User.get_by_email(email, case_insensitive=True, cache=True)
901 if user:
901 if user:
902 if user.first_name or user.last_name:
902 if user.first_name or user.last_name:
903 return '%s %s &lt;%s&gt;' % (
903 return '%s %s &lt;%s&gt;' % (
904 user.first_name, user.last_name, email)
904 user.first_name, user.last_name, email)
905 else:
905 else:
906 return email
906 return email
907 else:
907 else:
908 return email
908 return email
909 else:
909 else:
910 return None
910 return None
911
911
912
912
913 def person_by_id(id_, show_attr="username_and_name"):
913 def person_by_id(id_, show_attr="username_and_name"):
914 # attr to return from fetched user
914 # attr to return from fetched user
915 person_getter = lambda usr: getattr(usr, show_attr)
915 person_getter = lambda usr: getattr(usr, show_attr)
916
916
917 #maybe it's an ID ?
917 #maybe it's an ID ?
918 if str(id_).isdigit() or isinstance(id_, int):
918 if str(id_).isdigit() or isinstance(id_, int):
919 id_ = int(id_)
919 id_ = int(id_)
920 user = User.get(id_)
920 user = User.get(id_)
921 if user is not None:
921 if user is not None:
922 return person_getter(user)
922 return person_getter(user)
923 return id_
923 return id_
924
924
925
925
926 def gravatar_with_user(request, author, show_disabled=False):
926 def gravatar_with_user(request, author, show_disabled=False):
927 _render = request.get_partial_renderer('base/base.mako')
927 _render = request.get_partial_renderer('base/base.mako')
928 return _render('gravatar_with_user', author, show_disabled=show_disabled)
928 return _render('gravatar_with_user', author, show_disabled=show_disabled)
929
929
930
930
931 def desc_stylize(value):
931 def desc_stylize(value):
932 """
932 """
933 converts tags from value into html equivalent
933 converts tags from value into html equivalent
934
934
935 :param value:
935 :param value:
936 """
936 """
937 if not value:
937 if not value:
938 return ''
938 return ''
939
939
940 value = re.sub(r'\[see\ \=\>\ *([a-zA-Z0-9\/\=\?\&\ \:\/\.\-]*)\]',
940 value = re.sub(r'\[see\ \=\>\ *([a-zA-Z0-9\/\=\?\&\ \:\/\.\-]*)\]',
941 '<div class="metatag" tag="see">see =&gt; \\1 </div>', value)
941 '<div class="metatag" tag="see">see =&gt; \\1 </div>', value)
942 value = re.sub(r'\[license\ \=\>\ *([a-zA-Z0-9\/\=\?\&\ \:\/\.\-]*)\]',
942 value = re.sub(r'\[license\ \=\>\ *([a-zA-Z0-9\/\=\?\&\ \:\/\.\-]*)\]',
943 '<div class="metatag" tag="license"><a href="http:\/\/www.opensource.org/licenses/\\1">\\1</a></div>', value)
943 '<div class="metatag" tag="license"><a href="http:\/\/www.opensource.org/licenses/\\1">\\1</a></div>', value)
944 value = re.sub(r'\[(requires|recommends|conflicts|base)\ \=\>\ *([a-zA-Z0-9\-\/]*)\]',
944 value = re.sub(r'\[(requires|recommends|conflicts|base)\ \=\>\ *([a-zA-Z0-9\-\/]*)\]',
945 '<div class="metatag" tag="\\1">\\1 =&gt; <a href="/\\2">\\2</a></div>', value)
945 '<div class="metatag" tag="\\1">\\1 =&gt; <a href="/\\2">\\2</a></div>', value)
946 value = re.sub(r'\[(lang|language)\ \=\>\ *([a-zA-Z\-\/\#\+]*)\]',
946 value = re.sub(r'\[(lang|language)\ \=\>\ *([a-zA-Z\-\/\#\+]*)\]',
947 '<div class="metatag" tag="lang">\\2</div>', value)
947 '<div class="metatag" tag="lang">\\2</div>', value)
948 value = re.sub(r'\[([a-z]+)\]',
948 value = re.sub(r'\[([a-z]+)\]',
949 '<div class="metatag" tag="\\1">\\1</div>', value)
949 '<div class="metatag" tag="\\1">\\1</div>', value)
950
950
951 return value
951 return value
952
952
953
953
954 def escaped_stylize(value):
954 def escaped_stylize(value):
955 """
955 """
956 converts tags from value into html equivalent, but escaping its value first
956 converts tags from value into html equivalent, but escaping its value first
957 """
957 """
958 if not value:
958 if not value:
959 return ''
959 return ''
960
960
961 # Using default webhelper escape method, but has to force it as a
961 # Using default webhelper escape method, but has to force it as a
962 # plain unicode instead of a markup tag to be used in regex expressions
962 # plain unicode instead of a markup tag to be used in regex expressions
963 value = unicode(escape(safe_unicode(value)))
963 value = unicode(escape(safe_unicode(value)))
964
964
965 value = re.sub(r'\[see\ \=\&gt;\ *([a-zA-Z0-9\/\=\?\&amp;\ \:\/\.\-]*)\]',
965 value = re.sub(r'\[see\ \=\&gt;\ *([a-zA-Z0-9\/\=\?\&amp;\ \:\/\.\-]*)\]',
966 '<div class="metatag" tag="see">see =&gt; \\1 </div>', value)
966 '<div class="metatag" tag="see">see =&gt; \\1 </div>', value)
967 value = re.sub(r'\[license\ \=\&gt;\ *([a-zA-Z0-9\/\=\?\&amp;\ \:\/\.\-]*)\]',
967 value = re.sub(r'\[license\ \=\&gt;\ *([a-zA-Z0-9\/\=\?\&amp;\ \:\/\.\-]*)\]',
968 '<div class="metatag" tag="license"><a href="http:\/\/www.opensource.org/licenses/\\1">\\1</a></div>', value)
968 '<div class="metatag" tag="license"><a href="http:\/\/www.opensource.org/licenses/\\1">\\1</a></div>', value)
969 value = re.sub(r'\[(requires|recommends|conflicts|base)\ \=\&gt;\ *([a-zA-Z0-9\-\/]*)\]',
969 value = re.sub(r'\[(requires|recommends|conflicts|base)\ \=\&gt;\ *([a-zA-Z0-9\-\/]*)\]',
970 '<div class="metatag" tag="\\1">\\1 =&gt; <a href="/\\2">\\2</a></div>', value)
970 '<div class="metatag" tag="\\1">\\1 =&gt; <a href="/\\2">\\2</a></div>', value)
971 value = re.sub(r'\[(lang|language)\ \=\&gt;\ *([a-zA-Z\-\/\#\+]*)\]',
971 value = re.sub(r'\[(lang|language)\ \=\&gt;\ *([a-zA-Z\-\/\#\+]*)\]',
972 '<div class="metatag" tag="lang">\\2</div>', value)
972 '<div class="metatag" tag="lang">\\2</div>', value)
973 value = re.sub(r'\[([a-z]+)\]',
973 value = re.sub(r'\[([a-z]+)\]',
974 '<div class="metatag" tag="\\1">\\1</div>', value)
974 '<div class="metatag" tag="\\1">\\1</div>', value)
975
975
976 return value
976 return value
977
977
978
978
979 def bool2icon(value):
979 def bool2icon(value):
980 """
980 """
981 Returns boolean value of a given value, represented as html element with
981 Returns boolean value of a given value, represented as html element with
982 classes that will represent icons
982 classes that will represent icons
983
983
984 :param value: given value to convert to html node
984 :param value: given value to convert to html node
985 """
985 """
986
986
987 if value: # does bool conversion
987 if value: # does bool conversion
988 return HTML.tag('i', class_="icon-true")
988 return HTML.tag('i', class_="icon-true")
989 else: # not true as bool
989 else: # not true as bool
990 return HTML.tag('i', class_="icon-false")
990 return HTML.tag('i', class_="icon-false")
991
991
992
992
993 #==============================================================================
993 #==============================================================================
994 # PERMS
994 # PERMS
995 #==============================================================================
995 #==============================================================================
996 from rhodecode.lib.auth import HasPermissionAny, HasPermissionAll, \
996 from rhodecode.lib.auth import HasPermissionAny, HasPermissionAll, \
997 HasRepoPermissionAny, HasRepoPermissionAll, HasRepoGroupPermissionAll, \
997 HasRepoPermissionAny, HasRepoPermissionAll, HasRepoGroupPermissionAll, \
998 HasRepoGroupPermissionAny, HasRepoPermissionAnyApi, get_csrf_token, \
998 HasRepoGroupPermissionAny, HasRepoPermissionAnyApi, get_csrf_token, \
999 csrf_token_key
999 csrf_token_key
1000
1000
1001
1001
1002 #==============================================================================
1002 #==============================================================================
1003 # GRAVATAR URL
1003 # GRAVATAR URL
1004 #==============================================================================
1004 #==============================================================================
1005 class InitialsGravatar(object):
1005 class InitialsGravatar(object):
1006 def __init__(self, email_address, first_name, last_name, size=30,
1006 def __init__(self, email_address, first_name, last_name, size=30,
1007 background=None, text_color='#fff'):
1007 background=None, text_color='#fff'):
1008 self.size = size
1008 self.size = size
1009 self.first_name = first_name
1009 self.first_name = first_name
1010 self.last_name = last_name
1010 self.last_name = last_name
1011 self.email_address = email_address
1011 self.email_address = email_address
1012 self.background = background or self.str2color(email_address)
1012 self.background = background or self.str2color(email_address)
1013 self.text_color = text_color
1013 self.text_color = text_color
1014
1014
1015 def get_color_bank(self):
1015 def get_color_bank(self):
1016 """
1016 """
1017 returns a predefined list of colors that gravatars can use.
1017 returns a predefined list of colors that gravatars can use.
1018 Those are randomized distinct colors that guarantee readability and
1018 Those are randomized distinct colors that guarantee readability and
1019 uniqueness.
1019 uniqueness.
1020
1020
1021 generated with: http://phrogz.net/css/distinct-colors.html
1021 generated with: http://phrogz.net/css/distinct-colors.html
1022 """
1022 """
1023 return [
1023 return [
1024 '#bf3030', '#a67f53', '#00ff00', '#5989b3', '#392040', '#d90000',
1024 '#bf3030', '#a67f53', '#00ff00', '#5989b3', '#392040', '#d90000',
1025 '#402910', '#204020', '#79baf2', '#a700b3', '#bf6060', '#7f5320',
1025 '#402910', '#204020', '#79baf2', '#a700b3', '#bf6060', '#7f5320',
1026 '#008000', '#003059', '#ee00ff', '#ff0000', '#8c4b00', '#007300',
1026 '#008000', '#003059', '#ee00ff', '#ff0000', '#8c4b00', '#007300',
1027 '#005fb3', '#de73e6', '#ff4040', '#ffaa00', '#3df255', '#203140',
1027 '#005fb3', '#de73e6', '#ff4040', '#ffaa00', '#3df255', '#203140',
1028 '#47004d', '#591616', '#664400', '#59b365', '#0d2133', '#83008c',
1028 '#47004d', '#591616', '#664400', '#59b365', '#0d2133', '#83008c',
1029 '#592d2d', '#bf9f60', '#73e682', '#1d3f73', '#73006b', '#402020',
1029 '#592d2d', '#bf9f60', '#73e682', '#1d3f73', '#73006b', '#402020',
1030 '#b2862d', '#397341', '#597db3', '#e600d6', '#a60000', '#736039',
1030 '#b2862d', '#397341', '#597db3', '#e600d6', '#a60000', '#736039',
1031 '#00b318', '#79aaf2', '#330d30', '#ff8080', '#403010', '#16591f',
1031 '#00b318', '#79aaf2', '#330d30', '#ff8080', '#403010', '#16591f',
1032 '#002459', '#8c4688', '#e50000', '#ffbf40', '#00732e', '#102340',
1032 '#002459', '#8c4688', '#e50000', '#ffbf40', '#00732e', '#102340',
1033 '#bf60ac', '#8c4646', '#cc8800', '#00a642', '#1d3473', '#b32d98',
1033 '#bf60ac', '#8c4646', '#cc8800', '#00a642', '#1d3473', '#b32d98',
1034 '#660e00', '#ffd580', '#80ffb2', '#7391e6', '#733967', '#d97b6c',
1034 '#660e00', '#ffd580', '#80ffb2', '#7391e6', '#733967', '#d97b6c',
1035 '#8c5e00', '#59b389', '#3967e6', '#590047', '#73281d', '#665200',
1035 '#8c5e00', '#59b389', '#3967e6', '#590047', '#73281d', '#665200',
1036 '#00e67a', '#2d50b3', '#8c2377', '#734139', '#b2982d', '#16593a',
1036 '#00e67a', '#2d50b3', '#8c2377', '#734139', '#b2982d', '#16593a',
1037 '#001859', '#ff00aa', '#a65e53', '#ffcc00', '#0d3321', '#2d3959',
1037 '#001859', '#ff00aa', '#a65e53', '#ffcc00', '#0d3321', '#2d3959',
1038 '#731d56', '#401610', '#4c3d00', '#468c6c', '#002ca6', '#d936a3',
1038 '#731d56', '#401610', '#4c3d00', '#468c6c', '#002ca6', '#d936a3',
1039 '#d94c36', '#403920', '#36d9a3', '#0d1733', '#592d4a', '#993626',
1039 '#d94c36', '#403920', '#36d9a3', '#0d1733', '#592d4a', '#993626',
1040 '#cca300', '#00734d', '#46598c', '#8c005e', '#7f1100', '#8c7000',
1040 '#cca300', '#00734d', '#46598c', '#8c005e', '#7f1100', '#8c7000',
1041 '#00a66f', '#7382e6', '#b32d74', '#d9896c', '#ffe680', '#1d7362',
1041 '#00a66f', '#7382e6', '#b32d74', '#d9896c', '#ffe680', '#1d7362',
1042 '#364cd9', '#73003d', '#d93a00', '#998a4d', '#59b3a1', '#5965b3',
1042 '#364cd9', '#73003d', '#d93a00', '#998a4d', '#59b3a1', '#5965b3',
1043 '#e5007a', '#73341d', '#665f00', '#00b38f', '#0018b3', '#59163a',
1043 '#e5007a', '#73341d', '#665f00', '#00b38f', '#0018b3', '#59163a',
1044 '#b2502d', '#bfb960', '#00ffcc', '#23318c', '#a6537f', '#734939',
1044 '#b2502d', '#bfb960', '#00ffcc', '#23318c', '#a6537f', '#734939',
1045 '#b2a700', '#104036', '#3d3df2', '#402031', '#e56739', '#736f39',
1045 '#b2a700', '#104036', '#3d3df2', '#402031', '#e56739', '#736f39',
1046 '#79f2ea', '#000059', '#401029', '#4c1400', '#ffee00', '#005953',
1046 '#79f2ea', '#000059', '#401029', '#4c1400', '#ffee00', '#005953',
1047 '#101040', '#990052', '#402820', '#403d10', '#00ffee', '#0000d9',
1047 '#101040', '#990052', '#402820', '#403d10', '#00ffee', '#0000d9',
1048 '#ff80c4', '#a66953', '#eeff00', '#00ccbe', '#8080ff', '#e673a1',
1048 '#ff80c4', '#a66953', '#eeff00', '#00ccbe', '#8080ff', '#e673a1',
1049 '#a62c00', '#474d00', '#1a3331', '#46468c', '#733950', '#662900',
1049 '#a62c00', '#474d00', '#1a3331', '#46468c', '#733950', '#662900',
1050 '#858c23', '#238c85', '#0f0073', '#b20047', '#d9986c', '#becc00',
1050 '#858c23', '#238c85', '#0f0073', '#b20047', '#d9986c', '#becc00',
1051 '#396f73', '#281d73', '#ff0066', '#ff6600', '#dee673', '#59adb3',
1051 '#396f73', '#281d73', '#ff0066', '#ff6600', '#dee673', '#59adb3',
1052 '#6559b3', '#590024', '#b2622d', '#98b32d', '#36ced9', '#332d59',
1052 '#6559b3', '#590024', '#b2622d', '#98b32d', '#36ced9', '#332d59',
1053 '#40001a', '#733f1d', '#526600', '#005359', '#242040', '#bf6079',
1053 '#40001a', '#733f1d', '#526600', '#005359', '#242040', '#bf6079',
1054 '#735039', '#cef23d', '#007780', '#5630bf', '#66001b', '#b24700',
1054 '#735039', '#cef23d', '#007780', '#5630bf', '#66001b', '#b24700',
1055 '#acbf60', '#1d6273', '#25008c', '#731d34', '#a67453', '#50592d',
1055 '#acbf60', '#1d6273', '#25008c', '#731d34', '#a67453', '#50592d',
1056 '#00ccff', '#6600ff', '#ff0044', '#4c1f00', '#8a994d', '#79daf2',
1056 '#00ccff', '#6600ff', '#ff0044', '#4c1f00', '#8a994d', '#79daf2',
1057 '#a173e6', '#d93662', '#402310', '#aaff00', '#2d98b3', '#8c40ff',
1057 '#a173e6', '#d93662', '#402310', '#aaff00', '#2d98b3', '#8c40ff',
1058 '#592d39', '#ff8c40', '#354020', '#103640', '#1a0040', '#331a20',
1058 '#592d39', '#ff8c40', '#354020', '#103640', '#1a0040', '#331a20',
1059 '#331400', '#334d00', '#1d5673', '#583973', '#7f0022', '#4c3626',
1059 '#331400', '#334d00', '#1d5673', '#583973', '#7f0022', '#4c3626',
1060 '#88cc00', '#36a3d9', '#3d0073', '#d9364c', '#33241a', '#698c23',
1060 '#88cc00', '#36a3d9', '#3d0073', '#d9364c', '#33241a', '#698c23',
1061 '#5995b3', '#300059', '#e57382', '#7f3300', '#366600', '#00aaff',
1061 '#5995b3', '#300059', '#e57382', '#7f3300', '#366600', '#00aaff',
1062 '#3a1659', '#733941', '#663600', '#74b32d', '#003c59', '#7f53a6',
1062 '#3a1659', '#733941', '#663600', '#74b32d', '#003c59', '#7f53a6',
1063 '#73000f', '#ff8800', '#baf279', '#79caf2', '#291040', '#a6293a',
1063 '#73000f', '#ff8800', '#baf279', '#79caf2', '#291040', '#a6293a',
1064 '#b2742d', '#587339', '#0077b3', '#632699', '#400009', '#d9a66c',
1064 '#b2742d', '#587339', '#0077b3', '#632699', '#400009', '#d9a66c',
1065 '#294010', '#2d4a59', '#aa00ff', '#4c131b', '#b25f00', '#5ce600',
1065 '#294010', '#2d4a59', '#aa00ff', '#4c131b', '#b25f00', '#5ce600',
1066 '#267399', '#a336d9', '#990014', '#664e33', '#86bf60', '#0088ff',
1066 '#267399', '#a336d9', '#990014', '#664e33', '#86bf60', '#0088ff',
1067 '#7700b3', '#593a16', '#073300', '#1d4b73', '#ac60bf', '#e59539',
1067 '#7700b3', '#593a16', '#073300', '#1d4b73', '#ac60bf', '#e59539',
1068 '#4f8c46', '#368dd9', '#5c0073'
1068 '#4f8c46', '#368dd9', '#5c0073'
1069 ]
1069 ]
1070
1070
1071 def rgb_to_hex_color(self, rgb_tuple):
1071 def rgb_to_hex_color(self, rgb_tuple):
1072 """
1072 """
1073 Converts an rgb_tuple passed to an hex color.
1073 Converts an rgb_tuple passed to an hex color.
1074
1074
1075 :param rgb_tuple: tuple with 3 ints represents rgb color space
1075 :param rgb_tuple: tuple with 3 ints represents rgb color space
1076 """
1076 """
1077 return '#' + ("".join(map(chr, rgb_tuple)).encode('hex'))
1077 return '#' + ("".join(map(chr, rgb_tuple)).encode('hex'))
1078
1078
1079 def email_to_int_list(self, email_str):
1079 def email_to_int_list(self, email_str):
1080 """
1080 """
1081 Get every byte of the hex digest value of email and turn it to integer.
1081 Get every byte of the hex digest value of email and turn it to integer.
1082 It's going to be always between 0-255
1082 It's going to be always between 0-255
1083 """
1083 """
1084 digest = md5_safe(email_str.lower())
1084 digest = md5_safe(email_str.lower())
1085 return [int(digest[i * 2:i * 2 + 2], 16) for i in range(16)]
1085 return [int(digest[i * 2:i * 2 + 2], 16) for i in range(16)]
1086
1086
1087 def pick_color_bank_index(self, email_str, color_bank):
1087 def pick_color_bank_index(self, email_str, color_bank):
1088 return self.email_to_int_list(email_str)[0] % len(color_bank)
1088 return self.email_to_int_list(email_str)[0] % len(color_bank)
1089
1089
1090 def str2color(self, email_str):
1090 def str2color(self, email_str):
1091 """
1091 """
1092 Tries to map in a stable algorithm an email to color
1092 Tries to map in a stable algorithm an email to color
1093
1093
1094 :param email_str:
1094 :param email_str:
1095 """
1095 """
1096 color_bank = self.get_color_bank()
1096 color_bank = self.get_color_bank()
1097 # pick position (module it's length so we always find it in the
1097 # pick position (module it's length so we always find it in the
1098 # bank even if it's smaller than 256 values
1098 # bank even if it's smaller than 256 values
1099 pos = self.pick_color_bank_index(email_str, color_bank)
1099 pos = self.pick_color_bank_index(email_str, color_bank)
1100 return color_bank[pos]
1100 return color_bank[pos]
1101
1101
1102 def normalize_email(self, email_address):
1102 def normalize_email(self, email_address):
1103 import unicodedata
1103 import unicodedata
1104 # default host used to fill in the fake/missing email
1104 # default host used to fill in the fake/missing email
1105 default_host = u'localhost'
1105 default_host = u'localhost'
1106
1106
1107 if not email_address:
1107 if not email_address:
1108 email_address = u'%s@%s' % (User.DEFAULT_USER, default_host)
1108 email_address = u'%s@%s' % (User.DEFAULT_USER, default_host)
1109
1109
1110 email_address = safe_unicode(email_address)
1110 email_address = safe_unicode(email_address)
1111
1111
1112 if u'@' not in email_address:
1112 if u'@' not in email_address:
1113 email_address = u'%s@%s' % (email_address, default_host)
1113 email_address = u'%s@%s' % (email_address, default_host)
1114
1114
1115 if email_address.endswith(u'@'):
1115 if email_address.endswith(u'@'):
1116 email_address = u'%s%s' % (email_address, default_host)
1116 email_address = u'%s%s' % (email_address, default_host)
1117
1117
1118 email_address = unicodedata.normalize('NFKD', email_address)\
1118 email_address = unicodedata.normalize('NFKD', email_address)\
1119 .encode('ascii', 'ignore')
1119 .encode('ascii', 'ignore')
1120 return email_address
1120 return email_address
1121
1121
1122 def get_initials(self):
1122 def get_initials(self):
1123 """
1123 """
1124 Returns 2 letter initials calculated based on the input.
1124 Returns 2 letter initials calculated based on the input.
1125 The algorithm picks first given email address, and takes first letter
1125 The algorithm picks first given email address, and takes first letter
1126 of part before @, and then the first letter of server name. In case
1126 of part before @, and then the first letter of server name. In case
1127 the part before @ is in a format of `somestring.somestring2` it replaces
1127 the part before @ is in a format of `somestring.somestring2` it replaces
1128 the server letter with first letter of somestring2
1128 the server letter with first letter of somestring2
1129
1129
1130 In case function was initialized with both first and lastname, this
1130 In case function was initialized with both first and lastname, this
1131 overrides the extraction from email by first letter of the first and
1131 overrides the extraction from email by first letter of the first and
1132 last name. We add special logic to that functionality, In case Full name
1132 last name. We add special logic to that functionality, In case Full name
1133 is compound, like Guido Von Rossum, we use last part of the last name
1133 is compound, like Guido Von Rossum, we use last part of the last name
1134 (Von Rossum) picking `R`.
1134 (Von Rossum) picking `R`.
1135
1135
1136 Function also normalizes the non-ascii characters to they ascii
1136 Function also normalizes the non-ascii characters to they ascii
1137 representation, eg Δ„ => A
1137 representation, eg Δ„ => A
1138 """
1138 """
1139 import unicodedata
1139 import unicodedata
1140 # replace non-ascii to ascii
1140 # replace non-ascii to ascii
1141 first_name = unicodedata.normalize(
1141 first_name = unicodedata.normalize(
1142 'NFKD', safe_unicode(self.first_name)).encode('ascii', 'ignore')
1142 'NFKD', safe_unicode(self.first_name)).encode('ascii', 'ignore')
1143 last_name = unicodedata.normalize(
1143 last_name = unicodedata.normalize(
1144 'NFKD', safe_unicode(self.last_name)).encode('ascii', 'ignore')
1144 'NFKD', safe_unicode(self.last_name)).encode('ascii', 'ignore')
1145
1145
1146 # do NFKD encoding, and also make sure email has proper format
1146 # do NFKD encoding, and also make sure email has proper format
1147 email_address = self.normalize_email(self.email_address)
1147 email_address = self.normalize_email(self.email_address)
1148
1148
1149 # first push the email initials
1149 # first push the email initials
1150 prefix, server = email_address.split('@', 1)
1150 prefix, server = email_address.split('@', 1)
1151
1151
1152 # check if prefix is maybe a 'first_name.last_name' syntax
1152 # check if prefix is maybe a 'first_name.last_name' syntax
1153 _dot_split = prefix.rsplit('.', 1)
1153 _dot_split = prefix.rsplit('.', 1)
1154 if len(_dot_split) == 2:
1154 if len(_dot_split) == 2 and _dot_split[1]:
1155 initials = [_dot_split[0][0], _dot_split[1][0]]
1155 initials = [_dot_split[0][0], _dot_split[1][0]]
1156 else:
1156 else:
1157 initials = [prefix[0], server[0]]
1157 initials = [prefix[0], server[0]]
1158
1158
1159 # then try to replace either first_name or last_name
1159 # then try to replace either first_name or last_name
1160 fn_letter = (first_name or " ")[0].strip()
1160 fn_letter = (first_name or " ")[0].strip()
1161 ln_letter = (last_name.split(' ', 1)[-1] or " ")[0].strip()
1161 ln_letter = (last_name.split(' ', 1)[-1] or " ")[0].strip()
1162
1162
1163 if fn_letter:
1163 if fn_letter:
1164 initials[0] = fn_letter
1164 initials[0] = fn_letter
1165
1165
1166 if ln_letter:
1166 if ln_letter:
1167 initials[1] = ln_letter
1167 initials[1] = ln_letter
1168
1168
1169 return ''.join(initials).upper()
1169 return ''.join(initials).upper()
1170
1170
1171 def get_img_data_by_type(self, font_family, img_type):
1171 def get_img_data_by_type(self, font_family, img_type):
1172 default_user = """
1172 default_user = """
1173 <svg xmlns="http://www.w3.org/2000/svg"
1173 <svg xmlns="http://www.w3.org/2000/svg"
1174 version="1.1" x="0px" y="0px" width="{size}" height="{size}"
1174 version="1.1" x="0px" y="0px" width="{size}" height="{size}"
1175 viewBox="-15 -10 439.165 429.164"
1175 viewBox="-15 -10 439.165 429.164"
1176
1176
1177 xml:space="preserve"
1177 xml:space="preserve"
1178 style="background:{background};" >
1178 style="background:{background};" >
1179
1179
1180 <path d="M204.583,216.671c50.664,0,91.74-48.075,
1180 <path d="M204.583,216.671c50.664,0,91.74-48.075,
1181 91.74-107.378c0-82.237-41.074-107.377-91.74-107.377
1181 91.74-107.378c0-82.237-41.074-107.377-91.74-107.377
1182 c-50.668,0-91.74,25.14-91.74,107.377C112.844,
1182 c-50.668,0-91.74,25.14-91.74,107.377C112.844,
1183 168.596,153.916,216.671,
1183 168.596,153.916,216.671,
1184 204.583,216.671z" fill="{text_color}"/>
1184 204.583,216.671z" fill="{text_color}"/>
1185 <path d="M407.164,374.717L360.88,
1185 <path d="M407.164,374.717L360.88,
1186 270.454c-2.117-4.771-5.836-8.728-10.465-11.138l-71.83-37.392
1186 270.454c-2.117-4.771-5.836-8.728-10.465-11.138l-71.83-37.392
1187 c-1.584-0.823-3.502-0.663-4.926,0.415c-20.316,
1187 c-1.584-0.823-3.502-0.663-4.926,0.415c-20.316,
1188 15.366-44.203,23.488-69.076,23.488c-24.877,
1188 15.366-44.203,23.488-69.076,23.488c-24.877,
1189 0-48.762-8.122-69.078-23.488
1189 0-48.762-8.122-69.078-23.488
1190 c-1.428-1.078-3.346-1.238-4.93-0.415L58.75,
1190 c-1.428-1.078-3.346-1.238-4.93-0.415L58.75,
1191 259.316c-4.631,2.41-8.346,6.365-10.465,11.138L2.001,374.717
1191 259.316c-4.631,2.41-8.346,6.365-10.465,11.138L2.001,374.717
1192 c-3.191,7.188-2.537,15.412,1.75,22.005c4.285,
1192 c-3.191,7.188-2.537,15.412,1.75,22.005c4.285,
1193 6.592,11.537,10.526,19.4,10.526h362.861c7.863,0,15.117-3.936,
1193 6.592,11.537,10.526,19.4,10.526h362.861c7.863,0,15.117-3.936,
1194 19.402-10.527 C409.699,390.129,
1194 19.402-10.527 C409.699,390.129,
1195 410.355,381.902,407.164,374.717z" fill="{text_color}"/>
1195 410.355,381.902,407.164,374.717z" fill="{text_color}"/>
1196 </svg>""".format(
1196 </svg>""".format(
1197 size=self.size,
1197 size=self.size,
1198 background='#979797', # @grey4
1198 background='#979797', # @grey4
1199 text_color=self.text_color,
1199 text_color=self.text_color,
1200 font_family=font_family)
1200 font_family=font_family)
1201
1201
1202 return {
1202 return {
1203 "default_user": default_user
1203 "default_user": default_user
1204 }[img_type]
1204 }[img_type]
1205
1205
1206 def get_img_data(self, svg_type=None):
1206 def get_img_data(self, svg_type=None):
1207 """
1207 """
1208 generates the svg metadata for image
1208 generates the svg metadata for image
1209 """
1209 """
1210
1210
1211 font_family = ','.join([
1211 font_family = ','.join([
1212 'proximanovaregular',
1212 'proximanovaregular',
1213 'Proxima Nova Regular',
1213 'Proxima Nova Regular',
1214 'Proxima Nova',
1214 'Proxima Nova',
1215 'Arial',
1215 'Arial',
1216 'Lucida Grande',
1216 'Lucida Grande',
1217 'sans-serif'
1217 'sans-serif'
1218 ])
1218 ])
1219 if svg_type:
1219 if svg_type:
1220 return self.get_img_data_by_type(font_family, svg_type)
1220 return self.get_img_data_by_type(font_family, svg_type)
1221
1221
1222 initials = self.get_initials()
1222 initials = self.get_initials()
1223 img_data = """
1223 img_data = """
1224 <svg xmlns="http://www.w3.org/2000/svg" pointer-events="none"
1224 <svg xmlns="http://www.w3.org/2000/svg" pointer-events="none"
1225 width="{size}" height="{size}"
1225 width="{size}" height="{size}"
1226 style="width: 100%; height: 100%; background-color: {background}"
1226 style="width: 100%; height: 100%; background-color: {background}"
1227 viewBox="0 0 {size} {size}">
1227 viewBox="0 0 {size} {size}">
1228 <text text-anchor="middle" y="50%" x="50%" dy="0.35em"
1228 <text text-anchor="middle" y="50%" x="50%" dy="0.35em"
1229 pointer-events="auto" fill="{text_color}"
1229 pointer-events="auto" fill="{text_color}"
1230 font-family="{font_family}"
1230 font-family="{font_family}"
1231 style="font-weight: 400; font-size: {f_size}px;">{text}
1231 style="font-weight: 400; font-size: {f_size}px;">{text}
1232 </text>
1232 </text>
1233 </svg>""".format(
1233 </svg>""".format(
1234 size=self.size,
1234 size=self.size,
1235 f_size=self.size/1.85, # scale the text inside the box nicely
1235 f_size=self.size/1.85, # scale the text inside the box nicely
1236 background=self.background,
1236 background=self.background,
1237 text_color=self.text_color,
1237 text_color=self.text_color,
1238 text=initials.upper(),
1238 text=initials.upper(),
1239 font_family=font_family)
1239 font_family=font_family)
1240
1240
1241 return img_data
1241 return img_data
1242
1242
1243 def generate_svg(self, svg_type=None):
1243 def generate_svg(self, svg_type=None):
1244 img_data = self.get_img_data(svg_type)
1244 img_data = self.get_img_data(svg_type)
1245 return "data:image/svg+xml;base64,%s" % img_data.encode('base64')
1245 return "data:image/svg+xml;base64,%s" % img_data.encode('base64')
1246
1246
1247
1247
1248 def initials_gravatar(email_address, first_name, last_name, size=30):
1248 def initials_gravatar(email_address, first_name, last_name, size=30):
1249 svg_type = None
1249 svg_type = None
1250 if email_address == User.DEFAULT_USER_EMAIL:
1250 if email_address == User.DEFAULT_USER_EMAIL:
1251 svg_type = 'default_user'
1251 svg_type = 'default_user'
1252 klass = InitialsGravatar(email_address, first_name, last_name, size)
1252 klass = InitialsGravatar(email_address, first_name, last_name, size)
1253 return klass.generate_svg(svg_type=svg_type)
1253 return klass.generate_svg(svg_type=svg_type)
1254
1254
1255
1255
1256 def gravatar_url(email_address, size=30, request=None):
1256 def gravatar_url(email_address, size=30, request=None):
1257 request = get_current_request()
1257 request = get_current_request()
1258 if request and hasattr(request, 'call_context'):
1258 if request and hasattr(request, 'call_context'):
1259 _use_gravatar = request.call_context.visual.use_gravatar
1259 _use_gravatar = request.call_context.visual.use_gravatar
1260 _gravatar_url = request.call_context.visual.gravatar_url
1260 _gravatar_url = request.call_context.visual.gravatar_url
1261 else:
1261 else:
1262 # doh, we need to re-import those to mock it later
1262 # doh, we need to re-import those to mock it later
1263 from pylons import tmpl_context as c
1263 from pylons import tmpl_context as c
1264
1264
1265 _use_gravatar = c.visual.use_gravatar
1265 _use_gravatar = c.visual.use_gravatar
1266 _gravatar_url = c.visual.gravatar_url
1266 _gravatar_url = c.visual.gravatar_url
1267
1267
1268 _gravatar_url = _gravatar_url or User.DEFAULT_GRAVATAR_URL
1268 _gravatar_url = _gravatar_url or User.DEFAULT_GRAVATAR_URL
1269
1269
1270 email_address = email_address or User.DEFAULT_USER_EMAIL
1270 email_address = email_address or User.DEFAULT_USER_EMAIL
1271 if isinstance(email_address, unicode):
1271 if isinstance(email_address, unicode):
1272 # hashlib crashes on unicode items
1272 # hashlib crashes on unicode items
1273 email_address = safe_str(email_address)
1273 email_address = safe_str(email_address)
1274
1274
1275 # empty email or default user
1275 # empty email or default user
1276 if not email_address or email_address == User.DEFAULT_USER_EMAIL:
1276 if not email_address or email_address == User.DEFAULT_USER_EMAIL:
1277 return initials_gravatar(User.DEFAULT_USER_EMAIL, '', '', size=size)
1277 return initials_gravatar(User.DEFAULT_USER_EMAIL, '', '', size=size)
1278
1278
1279 if _use_gravatar:
1279 if _use_gravatar:
1280 # TODO: Disuse pyramid thread locals. Think about another solution to
1280 # TODO: Disuse pyramid thread locals. Think about another solution to
1281 # get the host and schema here.
1281 # get the host and schema here.
1282 request = get_current_request()
1282 request = get_current_request()
1283 tmpl = safe_str(_gravatar_url)
1283 tmpl = safe_str(_gravatar_url)
1284 tmpl = tmpl.replace('{email}', email_address)\
1284 tmpl = tmpl.replace('{email}', email_address)\
1285 .replace('{md5email}', md5_safe(email_address.lower())) \
1285 .replace('{md5email}', md5_safe(email_address.lower())) \
1286 .replace('{netloc}', request.host)\
1286 .replace('{netloc}', request.host)\
1287 .replace('{scheme}', request.scheme)\
1287 .replace('{scheme}', request.scheme)\
1288 .replace('{size}', safe_str(size))
1288 .replace('{size}', safe_str(size))
1289 return tmpl
1289 return tmpl
1290 else:
1290 else:
1291 return initials_gravatar(email_address, '', '', size=size)
1291 return initials_gravatar(email_address, '', '', size=size)
1292
1292
1293
1293
1294 class Page(_Page):
1294 class Page(_Page):
1295 """
1295 """
1296 Custom pager to match rendering style with paginator
1296 Custom pager to match rendering style with paginator
1297 """
1297 """
1298
1298
1299 def _get_pos(self, cur_page, max_page, items):
1299 def _get_pos(self, cur_page, max_page, items):
1300 edge = (items / 2) + 1
1300 edge = (items / 2) + 1
1301 if (cur_page <= edge):
1301 if (cur_page <= edge):
1302 radius = max(items / 2, items - cur_page)
1302 radius = max(items / 2, items - cur_page)
1303 elif (max_page - cur_page) < edge:
1303 elif (max_page - cur_page) < edge:
1304 radius = (items - 1) - (max_page - cur_page)
1304 radius = (items - 1) - (max_page - cur_page)
1305 else:
1305 else:
1306 radius = items / 2
1306 radius = items / 2
1307
1307
1308 left = max(1, (cur_page - (radius)))
1308 left = max(1, (cur_page - (radius)))
1309 right = min(max_page, cur_page + (radius))
1309 right = min(max_page, cur_page + (radius))
1310 return left, cur_page, right
1310 return left, cur_page, right
1311
1311
1312 def _range(self, regexp_match):
1312 def _range(self, regexp_match):
1313 """
1313 """
1314 Return range of linked pages (e.g. '1 2 [3] 4 5 6 7 8').
1314 Return range of linked pages (e.g. '1 2 [3] 4 5 6 7 8').
1315
1315
1316 Arguments:
1316 Arguments:
1317
1317
1318 regexp_match
1318 regexp_match
1319 A "re" (regular expressions) match object containing the
1319 A "re" (regular expressions) match object containing the
1320 radius of linked pages around the current page in
1320 radius of linked pages around the current page in
1321 regexp_match.group(1) as a string
1321 regexp_match.group(1) as a string
1322
1322
1323 This function is supposed to be called as a callable in
1323 This function is supposed to be called as a callable in
1324 re.sub.
1324 re.sub.
1325
1325
1326 """
1326 """
1327 radius = int(regexp_match.group(1))
1327 radius = int(regexp_match.group(1))
1328
1328
1329 # Compute the first and last page number within the radius
1329 # Compute the first and last page number within the radius
1330 # e.g. '1 .. 5 6 [7] 8 9 .. 12'
1330 # e.g. '1 .. 5 6 [7] 8 9 .. 12'
1331 # -> leftmost_page = 5
1331 # -> leftmost_page = 5
1332 # -> rightmost_page = 9
1332 # -> rightmost_page = 9
1333 leftmost_page, _cur, rightmost_page = self._get_pos(self.page,
1333 leftmost_page, _cur, rightmost_page = self._get_pos(self.page,
1334 self.last_page,
1334 self.last_page,
1335 (radius * 2) + 1)
1335 (radius * 2) + 1)
1336 nav_items = []
1336 nav_items = []
1337
1337
1338 # Create a link to the first page (unless we are on the first page
1338 # Create a link to the first page (unless we are on the first page
1339 # or there would be no need to insert '..' spacers)
1339 # or there would be no need to insert '..' spacers)
1340 if self.page != self.first_page and self.first_page < leftmost_page:
1340 if self.page != self.first_page and self.first_page < leftmost_page:
1341 nav_items.append(self._pagerlink(self.first_page, self.first_page))
1341 nav_items.append(self._pagerlink(self.first_page, self.first_page))
1342
1342
1343 # Insert dots if there are pages between the first page
1343 # Insert dots if there are pages between the first page
1344 # and the currently displayed page range
1344 # and the currently displayed page range
1345 if leftmost_page - self.first_page > 1:
1345 if leftmost_page - self.first_page > 1:
1346 # Wrap in a SPAN tag if nolink_attr is set
1346 # Wrap in a SPAN tag if nolink_attr is set
1347 text = '..'
1347 text = '..'
1348 if self.dotdot_attr:
1348 if self.dotdot_attr:
1349 text = HTML.span(c=text, **self.dotdot_attr)
1349 text = HTML.span(c=text, **self.dotdot_attr)
1350 nav_items.append(text)
1350 nav_items.append(text)
1351
1351
1352 for thispage in xrange(leftmost_page, rightmost_page + 1):
1352 for thispage in xrange(leftmost_page, rightmost_page + 1):
1353 # Hilight the current page number and do not use a link
1353 # Hilight the current page number and do not use a link
1354 if thispage == self.page:
1354 if thispage == self.page:
1355 text = '%s' % (thispage,)
1355 text = '%s' % (thispage,)
1356 # Wrap in a SPAN tag if nolink_attr is set
1356 # Wrap in a SPAN tag if nolink_attr is set
1357 if self.curpage_attr:
1357 if self.curpage_attr:
1358 text = HTML.span(c=text, **self.curpage_attr)
1358 text = HTML.span(c=text, **self.curpage_attr)
1359 nav_items.append(text)
1359 nav_items.append(text)
1360 # Otherwise create just a link to that page
1360 # Otherwise create just a link to that page
1361 else:
1361 else:
1362 text = '%s' % (thispage,)
1362 text = '%s' % (thispage,)
1363 nav_items.append(self._pagerlink(thispage, text))
1363 nav_items.append(self._pagerlink(thispage, text))
1364
1364
1365 # Insert dots if there are pages between the displayed
1365 # Insert dots if there are pages between the displayed
1366 # page numbers and the end of the page range
1366 # page numbers and the end of the page range
1367 if self.last_page - rightmost_page > 1:
1367 if self.last_page - rightmost_page > 1:
1368 text = '..'
1368 text = '..'
1369 # Wrap in a SPAN tag if nolink_attr is set
1369 # Wrap in a SPAN tag if nolink_attr is set
1370 if self.dotdot_attr:
1370 if self.dotdot_attr:
1371 text = HTML.span(c=text, **self.dotdot_attr)
1371 text = HTML.span(c=text, **self.dotdot_attr)
1372 nav_items.append(text)
1372 nav_items.append(text)
1373
1373
1374 # Create a link to the very last page (unless we are on the last
1374 # Create a link to the very last page (unless we are on the last
1375 # page or there would be no need to insert '..' spacers)
1375 # page or there would be no need to insert '..' spacers)
1376 if self.page != self.last_page and rightmost_page < self.last_page:
1376 if self.page != self.last_page and rightmost_page < self.last_page:
1377 nav_items.append(self._pagerlink(self.last_page, self.last_page))
1377 nav_items.append(self._pagerlink(self.last_page, self.last_page))
1378
1378
1379 ## prerender links
1379 ## prerender links
1380 #_page_link = url.current()
1380 #_page_link = url.current()
1381 #nav_items.append(literal('<link rel="prerender" href="%s?page=%s">' % (_page_link, str(int(self.page)+1))))
1381 #nav_items.append(literal('<link rel="prerender" href="%s?page=%s">' % (_page_link, str(int(self.page)+1))))
1382 #nav_items.append(literal('<link rel="prefetch" href="%s?page=%s">' % (_page_link, str(int(self.page)+1))))
1382 #nav_items.append(literal('<link rel="prefetch" href="%s?page=%s">' % (_page_link, str(int(self.page)+1))))
1383 return self.separator.join(nav_items)
1383 return self.separator.join(nav_items)
1384
1384
1385 def pager(self, format='~2~', page_param='page', partial_param='partial',
1385 def pager(self, format='~2~', page_param='page', partial_param='partial',
1386 show_if_single_page=False, separator=' ', onclick=None,
1386 show_if_single_page=False, separator=' ', onclick=None,
1387 symbol_first='<<', symbol_last='>>',
1387 symbol_first='<<', symbol_last='>>',
1388 symbol_previous='<', symbol_next='>',
1388 symbol_previous='<', symbol_next='>',
1389 link_attr={'class': 'pager_link', 'rel': 'prerender'},
1389 link_attr={'class': 'pager_link', 'rel': 'prerender'},
1390 curpage_attr={'class': 'pager_curpage'},
1390 curpage_attr={'class': 'pager_curpage'},
1391 dotdot_attr={'class': 'pager_dotdot'}, **kwargs):
1391 dotdot_attr={'class': 'pager_dotdot'}, **kwargs):
1392
1392
1393 self.curpage_attr = curpage_attr
1393 self.curpage_attr = curpage_attr
1394 self.separator = separator
1394 self.separator = separator
1395 self.pager_kwargs = kwargs
1395 self.pager_kwargs = kwargs
1396 self.page_param = page_param
1396 self.page_param = page_param
1397 self.partial_param = partial_param
1397 self.partial_param = partial_param
1398 self.onclick = onclick
1398 self.onclick = onclick
1399 self.link_attr = link_attr
1399 self.link_attr = link_attr
1400 self.dotdot_attr = dotdot_attr
1400 self.dotdot_attr = dotdot_attr
1401
1401
1402 # Don't show navigator if there is no more than one page
1402 # Don't show navigator if there is no more than one page
1403 if self.page_count == 0 or (self.page_count == 1 and not show_if_single_page):
1403 if self.page_count == 0 or (self.page_count == 1 and not show_if_single_page):
1404 return ''
1404 return ''
1405
1405
1406 from string import Template
1406 from string import Template
1407 # Replace ~...~ in token format by range of pages
1407 # Replace ~...~ in token format by range of pages
1408 result = re.sub(r'~(\d+)~', self._range, format)
1408 result = re.sub(r'~(\d+)~', self._range, format)
1409
1409
1410 # Interpolate '%' variables
1410 # Interpolate '%' variables
1411 result = Template(result).safe_substitute({
1411 result = Template(result).safe_substitute({
1412 'first_page': self.first_page,
1412 'first_page': self.first_page,
1413 'last_page': self.last_page,
1413 'last_page': self.last_page,
1414 'page': self.page,
1414 'page': self.page,
1415 'page_count': self.page_count,
1415 'page_count': self.page_count,
1416 'items_per_page': self.items_per_page,
1416 'items_per_page': self.items_per_page,
1417 'first_item': self.first_item,
1417 'first_item': self.first_item,
1418 'last_item': self.last_item,
1418 'last_item': self.last_item,
1419 'item_count': self.item_count,
1419 'item_count': self.item_count,
1420 'link_first': self.page > self.first_page and \
1420 'link_first': self.page > self.first_page and \
1421 self._pagerlink(self.first_page, symbol_first) or '',
1421 self._pagerlink(self.first_page, symbol_first) or '',
1422 'link_last': self.page < self.last_page and \
1422 'link_last': self.page < self.last_page and \
1423 self._pagerlink(self.last_page, symbol_last) or '',
1423 self._pagerlink(self.last_page, symbol_last) or '',
1424 'link_previous': self.previous_page and \
1424 'link_previous': self.previous_page and \
1425 self._pagerlink(self.previous_page, symbol_previous) \
1425 self._pagerlink(self.previous_page, symbol_previous) \
1426 or HTML.span(symbol_previous, class_="pg-previous disabled"),
1426 or HTML.span(symbol_previous, class_="pg-previous disabled"),
1427 'link_next': self.next_page and \
1427 'link_next': self.next_page and \
1428 self._pagerlink(self.next_page, symbol_next) \
1428 self._pagerlink(self.next_page, symbol_next) \
1429 or HTML.span(symbol_next, class_="pg-next disabled")
1429 or HTML.span(symbol_next, class_="pg-next disabled")
1430 })
1430 })
1431
1431
1432 return literal(result)
1432 return literal(result)
1433
1433
1434
1434
1435 #==============================================================================
1435 #==============================================================================
1436 # REPO PAGER, PAGER FOR REPOSITORY
1436 # REPO PAGER, PAGER FOR REPOSITORY
1437 #==============================================================================
1437 #==============================================================================
1438 class RepoPage(Page):
1438 class RepoPage(Page):
1439
1439
1440 def __init__(self, collection, page=1, items_per_page=20,
1440 def __init__(self, collection, page=1, items_per_page=20,
1441 item_count=None, url=None, **kwargs):
1441 item_count=None, url=None, **kwargs):
1442
1442
1443 """Create a "RepoPage" instance. special pager for paging
1443 """Create a "RepoPage" instance. special pager for paging
1444 repository
1444 repository
1445 """
1445 """
1446 self._url_generator = url
1446 self._url_generator = url
1447
1447
1448 # Safe the kwargs class-wide so they can be used in the pager() method
1448 # Safe the kwargs class-wide so they can be used in the pager() method
1449 self.kwargs = kwargs
1449 self.kwargs = kwargs
1450
1450
1451 # Save a reference to the collection
1451 # Save a reference to the collection
1452 self.original_collection = collection
1452 self.original_collection = collection
1453
1453
1454 self.collection = collection
1454 self.collection = collection
1455
1455
1456 # The self.page is the number of the current page.
1456 # The self.page is the number of the current page.
1457 # The first page has the number 1!
1457 # The first page has the number 1!
1458 try:
1458 try:
1459 self.page = int(page) # make it int() if we get it as a string
1459 self.page = int(page) # make it int() if we get it as a string
1460 except (ValueError, TypeError):
1460 except (ValueError, TypeError):
1461 self.page = 1
1461 self.page = 1
1462
1462
1463 self.items_per_page = items_per_page
1463 self.items_per_page = items_per_page
1464
1464
1465 # Unless the user tells us how many items the collections has
1465 # Unless the user tells us how many items the collections has
1466 # we calculate that ourselves.
1466 # we calculate that ourselves.
1467 if item_count is not None:
1467 if item_count is not None:
1468 self.item_count = item_count
1468 self.item_count = item_count
1469 else:
1469 else:
1470 self.item_count = len(self.collection)
1470 self.item_count = len(self.collection)
1471
1471
1472 # Compute the number of the first and last available page
1472 # Compute the number of the first and last available page
1473 if self.item_count > 0:
1473 if self.item_count > 0:
1474 self.first_page = 1
1474 self.first_page = 1
1475 self.page_count = int(math.ceil(float(self.item_count) /
1475 self.page_count = int(math.ceil(float(self.item_count) /
1476 self.items_per_page))
1476 self.items_per_page))
1477 self.last_page = self.first_page + self.page_count - 1
1477 self.last_page = self.first_page + self.page_count - 1
1478
1478
1479 # Make sure that the requested page number is the range of
1479 # Make sure that the requested page number is the range of
1480 # valid pages
1480 # valid pages
1481 if self.page > self.last_page:
1481 if self.page > self.last_page:
1482 self.page = self.last_page
1482 self.page = self.last_page
1483 elif self.page < self.first_page:
1483 elif self.page < self.first_page:
1484 self.page = self.first_page
1484 self.page = self.first_page
1485
1485
1486 # Note: the number of items on this page can be less than
1486 # Note: the number of items on this page can be less than
1487 # items_per_page if the last page is not full
1487 # items_per_page if the last page is not full
1488 self.first_item = max(0, (self.item_count) - (self.page *
1488 self.first_item = max(0, (self.item_count) - (self.page *
1489 items_per_page))
1489 items_per_page))
1490 self.last_item = ((self.item_count - 1) - items_per_page *
1490 self.last_item = ((self.item_count - 1) - items_per_page *
1491 (self.page - 1))
1491 (self.page - 1))
1492
1492
1493 self.items = list(self.collection[self.first_item:self.last_item + 1])
1493 self.items = list(self.collection[self.first_item:self.last_item + 1])
1494
1494
1495 # Links to previous and next page
1495 # Links to previous and next page
1496 if self.page > self.first_page:
1496 if self.page > self.first_page:
1497 self.previous_page = self.page - 1
1497 self.previous_page = self.page - 1
1498 else:
1498 else:
1499 self.previous_page = None
1499 self.previous_page = None
1500
1500
1501 if self.page < self.last_page:
1501 if self.page < self.last_page:
1502 self.next_page = self.page + 1
1502 self.next_page = self.page + 1
1503 else:
1503 else:
1504 self.next_page = None
1504 self.next_page = None
1505
1505
1506 # No items available
1506 # No items available
1507 else:
1507 else:
1508 self.first_page = None
1508 self.first_page = None
1509 self.page_count = 0
1509 self.page_count = 0
1510 self.last_page = None
1510 self.last_page = None
1511 self.first_item = None
1511 self.first_item = None
1512 self.last_item = None
1512 self.last_item = None
1513 self.previous_page = None
1513 self.previous_page = None
1514 self.next_page = None
1514 self.next_page = None
1515 self.items = []
1515 self.items = []
1516
1516
1517 # This is a subclass of the 'list' type. Initialise the list now.
1517 # This is a subclass of the 'list' type. Initialise the list now.
1518 list.__init__(self, reversed(self.items))
1518 list.__init__(self, reversed(self.items))
1519
1519
1520
1520
1521 def breadcrumb_repo_link(repo):
1521 def breadcrumb_repo_link(repo):
1522 """
1522 """
1523 Makes a breadcrumbs path link to repo
1523 Makes a breadcrumbs path link to repo
1524
1524
1525 ex::
1525 ex::
1526 group >> subgroup >> repo
1526 group >> subgroup >> repo
1527
1527
1528 :param repo: a Repository instance
1528 :param repo: a Repository instance
1529 """
1529 """
1530
1530
1531 path = [
1531 path = [
1532 link_to(group.name, route_path('repo_group_home', repo_group_name=group.group_name))
1532 link_to(group.name, route_path('repo_group_home', repo_group_name=group.group_name))
1533 for group in repo.groups_with_parents
1533 for group in repo.groups_with_parents
1534 ] + [
1534 ] + [
1535 link_to(repo.just_name, route_path('repo_summary', repo_name=repo.repo_name))
1535 link_to(repo.just_name, route_path('repo_summary', repo_name=repo.repo_name))
1536 ]
1536 ]
1537
1537
1538 return literal(' &raquo; '.join(path))
1538 return literal(' &raquo; '.join(path))
1539
1539
1540
1540
1541 def format_byte_size_binary(file_size):
1541 def format_byte_size_binary(file_size):
1542 """
1542 """
1543 Formats file/folder sizes to standard.
1543 Formats file/folder sizes to standard.
1544 """
1544 """
1545 if file_size is None:
1545 if file_size is None:
1546 file_size = 0
1546 file_size = 0
1547
1547
1548 formatted_size = format_byte_size(file_size, binary=True)
1548 formatted_size = format_byte_size(file_size, binary=True)
1549 return formatted_size
1549 return formatted_size
1550
1550
1551
1551
1552 def urlify_text(text_, safe=True):
1552 def urlify_text(text_, safe=True):
1553 """
1553 """
1554 Extrac urls from text and make html links out of them
1554 Extrac urls from text and make html links out of them
1555
1555
1556 :param text_:
1556 :param text_:
1557 """
1557 """
1558
1558
1559 url_pat = re.compile(r'''(http[s]?://(?:[a-zA-Z]|[0-9]|[$-_@#.&+]'''
1559 url_pat = re.compile(r'''(http[s]?://(?:[a-zA-Z]|[0-9]|[$-_@#.&+]'''
1560 '''|[!*\(\),]|(?:%[0-9a-fA-F][0-9a-fA-F]))+)''')
1560 '''|[!*\(\),]|(?:%[0-9a-fA-F][0-9a-fA-F]))+)''')
1561
1561
1562 def url_func(match_obj):
1562 def url_func(match_obj):
1563 url_full = match_obj.groups()[0]
1563 url_full = match_obj.groups()[0]
1564 return '<a href="%(url)s">%(url)s</a>' % ({'url': url_full})
1564 return '<a href="%(url)s">%(url)s</a>' % ({'url': url_full})
1565 _newtext = url_pat.sub(url_func, text_)
1565 _newtext = url_pat.sub(url_func, text_)
1566 if safe:
1566 if safe:
1567 return literal(_newtext)
1567 return literal(_newtext)
1568 return _newtext
1568 return _newtext
1569
1569
1570
1570
1571 def urlify_commits(text_, repository):
1571 def urlify_commits(text_, repository):
1572 """
1572 """
1573 Extract commit ids from text and make link from them
1573 Extract commit ids from text and make link from them
1574
1574
1575 :param text_:
1575 :param text_:
1576 :param repository: repo name to build the URL with
1576 :param repository: repo name to build the URL with
1577 """
1577 """
1578
1578
1579 URL_PAT = re.compile(r'(^|\s)([0-9a-fA-F]{12,40})($|\s)')
1579 URL_PAT = re.compile(r'(^|\s)([0-9a-fA-F]{12,40})($|\s)')
1580
1580
1581 def url_func(match_obj):
1581 def url_func(match_obj):
1582 commit_id = match_obj.groups()[1]
1582 commit_id = match_obj.groups()[1]
1583 pref = match_obj.groups()[0]
1583 pref = match_obj.groups()[0]
1584 suf = match_obj.groups()[2]
1584 suf = match_obj.groups()[2]
1585
1585
1586 tmpl = (
1586 tmpl = (
1587 '%(pref)s<a class="%(cls)s" href="%(url)s">'
1587 '%(pref)s<a class="%(cls)s" href="%(url)s">'
1588 '%(commit_id)s</a>%(suf)s'
1588 '%(commit_id)s</a>%(suf)s'
1589 )
1589 )
1590 return tmpl % {
1590 return tmpl % {
1591 'pref': pref,
1591 'pref': pref,
1592 'cls': 'revision-link',
1592 'cls': 'revision-link',
1593 'url': route_url('repo_commit', repo_name=repository,
1593 'url': route_url('repo_commit', repo_name=repository,
1594 commit_id=commit_id),
1594 commit_id=commit_id),
1595 'commit_id': commit_id,
1595 'commit_id': commit_id,
1596 'suf': suf
1596 'suf': suf
1597 }
1597 }
1598
1598
1599 newtext = URL_PAT.sub(url_func, text_)
1599 newtext = URL_PAT.sub(url_func, text_)
1600
1600
1601 return newtext
1601 return newtext
1602
1602
1603
1603
1604 def _process_url_func(match_obj, repo_name, uid, entry,
1604 def _process_url_func(match_obj, repo_name, uid, entry,
1605 return_raw_data=False, link_format='html'):
1605 return_raw_data=False, link_format='html'):
1606 pref = ''
1606 pref = ''
1607 if match_obj.group().startswith(' '):
1607 if match_obj.group().startswith(' '):
1608 pref = ' '
1608 pref = ' '
1609
1609
1610 issue_id = ''.join(match_obj.groups())
1610 issue_id = ''.join(match_obj.groups())
1611
1611
1612 if link_format == 'html':
1612 if link_format == 'html':
1613 tmpl = (
1613 tmpl = (
1614 '%(pref)s<a class="%(cls)s" href="%(url)s">'
1614 '%(pref)s<a class="%(cls)s" href="%(url)s">'
1615 '%(issue-prefix)s%(id-repr)s'
1615 '%(issue-prefix)s%(id-repr)s'
1616 '</a>')
1616 '</a>')
1617 elif link_format == 'rst':
1617 elif link_format == 'rst':
1618 tmpl = '`%(issue-prefix)s%(id-repr)s <%(url)s>`_'
1618 tmpl = '`%(issue-prefix)s%(id-repr)s <%(url)s>`_'
1619 elif link_format == 'markdown':
1619 elif link_format == 'markdown':
1620 tmpl = '[%(issue-prefix)s%(id-repr)s](%(url)s)'
1620 tmpl = '[%(issue-prefix)s%(id-repr)s](%(url)s)'
1621 else:
1621 else:
1622 raise ValueError('Bad link_format:{}'.format(link_format))
1622 raise ValueError('Bad link_format:{}'.format(link_format))
1623
1623
1624 (repo_name_cleaned,
1624 (repo_name_cleaned,
1625 parent_group_name) = RepoGroupModel().\
1625 parent_group_name) = RepoGroupModel().\
1626 _get_group_name_and_parent(repo_name)
1626 _get_group_name_and_parent(repo_name)
1627
1627
1628 # variables replacement
1628 # variables replacement
1629 named_vars = {
1629 named_vars = {
1630 'id': issue_id,
1630 'id': issue_id,
1631 'repo': repo_name,
1631 'repo': repo_name,
1632 'repo_name': repo_name_cleaned,
1632 'repo_name': repo_name_cleaned,
1633 'group_name': parent_group_name
1633 'group_name': parent_group_name
1634 }
1634 }
1635 # named regex variables
1635 # named regex variables
1636 named_vars.update(match_obj.groupdict())
1636 named_vars.update(match_obj.groupdict())
1637 _url = string.Template(entry['url']).safe_substitute(**named_vars)
1637 _url = string.Template(entry['url']).safe_substitute(**named_vars)
1638
1638
1639 data = {
1639 data = {
1640 'pref': pref,
1640 'pref': pref,
1641 'cls': 'issue-tracker-link',
1641 'cls': 'issue-tracker-link',
1642 'url': _url,
1642 'url': _url,
1643 'id-repr': issue_id,
1643 'id-repr': issue_id,
1644 'issue-prefix': entry['pref'],
1644 'issue-prefix': entry['pref'],
1645 'serv': entry['url'],
1645 'serv': entry['url'],
1646 }
1646 }
1647 if return_raw_data:
1647 if return_raw_data:
1648 return {
1648 return {
1649 'id': issue_id,
1649 'id': issue_id,
1650 'url': _url
1650 'url': _url
1651 }
1651 }
1652 return tmpl % data
1652 return tmpl % data
1653
1653
1654
1654
1655 def process_patterns(text_string, repo_name, link_format='html'):
1655 def process_patterns(text_string, repo_name, link_format='html'):
1656 allowed_formats = ['html', 'rst', 'markdown']
1656 allowed_formats = ['html', 'rst', 'markdown']
1657 if link_format not in allowed_formats:
1657 if link_format not in allowed_formats:
1658 raise ValueError('Link format can be only one of:{} got {}'.format(
1658 raise ValueError('Link format can be only one of:{} got {}'.format(
1659 allowed_formats, link_format))
1659 allowed_formats, link_format))
1660
1660
1661 repo = None
1661 repo = None
1662 if repo_name:
1662 if repo_name:
1663 # Retrieving repo_name to avoid invalid repo_name to explode on
1663 # Retrieving repo_name to avoid invalid repo_name to explode on
1664 # IssueTrackerSettingsModel but still passing invalid name further down
1664 # IssueTrackerSettingsModel but still passing invalid name further down
1665 repo = Repository.get_by_repo_name(repo_name, cache=True)
1665 repo = Repository.get_by_repo_name(repo_name, cache=True)
1666
1666
1667 settings_model = IssueTrackerSettingsModel(repo=repo)
1667 settings_model = IssueTrackerSettingsModel(repo=repo)
1668 active_entries = settings_model.get_settings(cache=True)
1668 active_entries = settings_model.get_settings(cache=True)
1669
1669
1670 issues_data = []
1670 issues_data = []
1671 newtext = text_string
1671 newtext = text_string
1672
1672
1673 for uid, entry in active_entries.items():
1673 for uid, entry in active_entries.items():
1674 log.debug('found issue tracker entry with uid %s' % (uid,))
1674 log.debug('found issue tracker entry with uid %s' % (uid,))
1675
1675
1676 if not (entry['pat'] and entry['url']):
1676 if not (entry['pat'] and entry['url']):
1677 log.debug('skipping due to missing data')
1677 log.debug('skipping due to missing data')
1678 continue
1678 continue
1679
1679
1680 log.debug('issue tracker entry: uid: `%s` PAT:%s URL:%s PREFIX:%s'
1680 log.debug('issue tracker entry: uid: `%s` PAT:%s URL:%s PREFIX:%s'
1681 % (uid, entry['pat'], entry['url'], entry['pref']))
1681 % (uid, entry['pat'], entry['url'], entry['pref']))
1682
1682
1683 try:
1683 try:
1684 pattern = re.compile(r'%s' % entry['pat'])
1684 pattern = re.compile(r'%s' % entry['pat'])
1685 except re.error:
1685 except re.error:
1686 log.exception(
1686 log.exception(
1687 'issue tracker pattern: `%s` failed to compile',
1687 'issue tracker pattern: `%s` failed to compile',
1688 entry['pat'])
1688 entry['pat'])
1689 continue
1689 continue
1690
1690
1691 data_func = partial(
1691 data_func = partial(
1692 _process_url_func, repo_name=repo_name, entry=entry, uid=uid,
1692 _process_url_func, repo_name=repo_name, entry=entry, uid=uid,
1693 return_raw_data=True)
1693 return_raw_data=True)
1694
1694
1695 for match_obj in pattern.finditer(text_string):
1695 for match_obj in pattern.finditer(text_string):
1696 issues_data.append(data_func(match_obj))
1696 issues_data.append(data_func(match_obj))
1697
1697
1698 url_func = partial(
1698 url_func = partial(
1699 _process_url_func, repo_name=repo_name, entry=entry, uid=uid,
1699 _process_url_func, repo_name=repo_name, entry=entry, uid=uid,
1700 link_format=link_format)
1700 link_format=link_format)
1701
1701
1702 newtext = pattern.sub(url_func, newtext)
1702 newtext = pattern.sub(url_func, newtext)
1703 log.debug('processed prefix:uid `%s`' % (uid,))
1703 log.debug('processed prefix:uid `%s`' % (uid,))
1704
1704
1705 return newtext, issues_data
1705 return newtext, issues_data
1706
1706
1707
1707
1708 def urlify_commit_message(commit_text, repository=None):
1708 def urlify_commit_message(commit_text, repository=None):
1709 """
1709 """
1710 Parses given text message and makes proper links.
1710 Parses given text message and makes proper links.
1711 issues are linked to given issue-server, and rest is a commit link
1711 issues are linked to given issue-server, and rest is a commit link
1712
1712
1713 :param commit_text:
1713 :param commit_text:
1714 :param repository:
1714 :param repository:
1715 """
1715 """
1716 from pylons import url # doh, we need to re-import url to mock it later
1716 from pylons import url # doh, we need to re-import url to mock it later
1717
1717
1718 def escaper(string):
1718 def escaper(string):
1719 return string.replace('<', '&lt;').replace('>', '&gt;')
1719 return string.replace('<', '&lt;').replace('>', '&gt;')
1720
1720
1721 newtext = escaper(commit_text)
1721 newtext = escaper(commit_text)
1722
1722
1723 # extract http/https links and make them real urls
1723 # extract http/https links and make them real urls
1724 newtext = urlify_text(newtext, safe=False)
1724 newtext = urlify_text(newtext, safe=False)
1725
1725
1726 # urlify commits - extract commit ids and make link out of them, if we have
1726 # urlify commits - extract commit ids and make link out of them, if we have
1727 # the scope of repository present.
1727 # the scope of repository present.
1728 if repository:
1728 if repository:
1729 newtext = urlify_commits(newtext, repository)
1729 newtext = urlify_commits(newtext, repository)
1730
1730
1731 # process issue tracker patterns
1731 # process issue tracker patterns
1732 newtext, issues = process_patterns(newtext, repository or '')
1732 newtext, issues = process_patterns(newtext, repository or '')
1733
1733
1734 return literal(newtext)
1734 return literal(newtext)
1735
1735
1736
1736
1737 def render_binary(repo_name, file_obj):
1737 def render_binary(repo_name, file_obj):
1738 """
1738 """
1739 Choose how to render a binary file
1739 Choose how to render a binary file
1740 """
1740 """
1741 filename = file_obj.name
1741 filename = file_obj.name
1742
1742
1743 # images
1743 # images
1744 for ext in ['*.png', '*.jpg', '*.ico', '*.gif']:
1744 for ext in ['*.png', '*.jpg', '*.ico', '*.gif']:
1745 if fnmatch.fnmatch(filename, pat=ext):
1745 if fnmatch.fnmatch(filename, pat=ext):
1746 alt = filename
1746 alt = filename
1747 src = route_path(
1747 src = route_path(
1748 'repo_file_raw', repo_name=repo_name,
1748 'repo_file_raw', repo_name=repo_name,
1749 commit_id=file_obj.commit.raw_id, f_path=file_obj.path)
1749 commit_id=file_obj.commit.raw_id, f_path=file_obj.path)
1750 return literal('<img class="rendered-binary" alt="{}" src="{}">'.format(alt, src))
1750 return literal('<img class="rendered-binary" alt="{}" src="{}">'.format(alt, src))
1751
1751
1752
1752
1753 def renderer_from_filename(filename, exclude=None):
1753 def renderer_from_filename(filename, exclude=None):
1754 """
1754 """
1755 choose a renderer based on filename, this works only for text based files
1755 choose a renderer based on filename, this works only for text based files
1756 """
1756 """
1757
1757
1758 # ipython
1758 # ipython
1759 for ext in ['*.ipynb']:
1759 for ext in ['*.ipynb']:
1760 if fnmatch.fnmatch(filename, pat=ext):
1760 if fnmatch.fnmatch(filename, pat=ext):
1761 return 'jupyter'
1761 return 'jupyter'
1762
1762
1763 is_markup = MarkupRenderer.renderer_from_filename(filename, exclude=exclude)
1763 is_markup = MarkupRenderer.renderer_from_filename(filename, exclude=exclude)
1764 if is_markup:
1764 if is_markup:
1765 return is_markup
1765 return is_markup
1766 return None
1766 return None
1767
1767
1768
1768
1769 def render(source, renderer='rst', mentions=False, relative_urls=None,
1769 def render(source, renderer='rst', mentions=False, relative_urls=None,
1770 repo_name=None):
1770 repo_name=None):
1771
1771
1772 def maybe_convert_relative_links(html_source):
1772 def maybe_convert_relative_links(html_source):
1773 if relative_urls:
1773 if relative_urls:
1774 return relative_links(html_source, relative_urls)
1774 return relative_links(html_source, relative_urls)
1775 return html_source
1775 return html_source
1776
1776
1777 if renderer == 'rst':
1777 if renderer == 'rst':
1778 if repo_name:
1778 if repo_name:
1779 # process patterns on comments if we pass in repo name
1779 # process patterns on comments if we pass in repo name
1780 source, issues = process_patterns(
1780 source, issues = process_patterns(
1781 source, repo_name, link_format='rst')
1781 source, repo_name, link_format='rst')
1782
1782
1783 return literal(
1783 return literal(
1784 '<div class="rst-block">%s</div>' %
1784 '<div class="rst-block">%s</div>' %
1785 maybe_convert_relative_links(
1785 maybe_convert_relative_links(
1786 MarkupRenderer.rst(source, mentions=mentions)))
1786 MarkupRenderer.rst(source, mentions=mentions)))
1787 elif renderer == 'markdown':
1787 elif renderer == 'markdown':
1788 if repo_name:
1788 if repo_name:
1789 # process patterns on comments if we pass in repo name
1789 # process patterns on comments if we pass in repo name
1790 source, issues = process_patterns(
1790 source, issues = process_patterns(
1791 source, repo_name, link_format='markdown')
1791 source, repo_name, link_format='markdown')
1792
1792
1793 return literal(
1793 return literal(
1794 '<div class="markdown-block">%s</div>' %
1794 '<div class="markdown-block">%s</div>' %
1795 maybe_convert_relative_links(
1795 maybe_convert_relative_links(
1796 MarkupRenderer.markdown(source, flavored=True,
1796 MarkupRenderer.markdown(source, flavored=True,
1797 mentions=mentions)))
1797 mentions=mentions)))
1798 elif renderer == 'jupyter':
1798 elif renderer == 'jupyter':
1799 return literal(
1799 return literal(
1800 '<div class="ipynb">%s</div>' %
1800 '<div class="ipynb">%s</div>' %
1801 maybe_convert_relative_links(
1801 maybe_convert_relative_links(
1802 MarkupRenderer.jupyter(source)))
1802 MarkupRenderer.jupyter(source)))
1803
1803
1804 # None means just show the file-source
1804 # None means just show the file-source
1805 return None
1805 return None
1806
1806
1807
1807
1808 def commit_status(repo, commit_id):
1808 def commit_status(repo, commit_id):
1809 return ChangesetStatusModel().get_status(repo, commit_id)
1809 return ChangesetStatusModel().get_status(repo, commit_id)
1810
1810
1811
1811
1812 def commit_status_lbl(commit_status):
1812 def commit_status_lbl(commit_status):
1813 return dict(ChangesetStatus.STATUSES).get(commit_status)
1813 return dict(ChangesetStatus.STATUSES).get(commit_status)
1814
1814
1815
1815
1816 def commit_time(repo_name, commit_id):
1816 def commit_time(repo_name, commit_id):
1817 repo = Repository.get_by_repo_name(repo_name)
1817 repo = Repository.get_by_repo_name(repo_name)
1818 commit = repo.get_commit(commit_id=commit_id)
1818 commit = repo.get_commit(commit_id=commit_id)
1819 return commit.date
1819 return commit.date
1820
1820
1821
1821
1822 def get_permission_name(key):
1822 def get_permission_name(key):
1823 return dict(Permission.PERMS).get(key)
1823 return dict(Permission.PERMS).get(key)
1824
1824
1825
1825
1826 def journal_filter_help(request):
1826 def journal_filter_help(request):
1827 _ = request.translate
1827 _ = request.translate
1828
1828
1829 return _(
1829 return _(
1830 'Example filter terms:\n' +
1830 'Example filter terms:\n' +
1831 ' repository:vcs\n' +
1831 ' repository:vcs\n' +
1832 ' username:marcin\n' +
1832 ' username:marcin\n' +
1833 ' username:(NOT marcin)\n' +
1833 ' username:(NOT marcin)\n' +
1834 ' action:*push*\n' +
1834 ' action:*push*\n' +
1835 ' ip:127.0.0.1\n' +
1835 ' ip:127.0.0.1\n' +
1836 ' date:20120101\n' +
1836 ' date:20120101\n' +
1837 ' date:[20120101100000 TO 20120102]\n' +
1837 ' date:[20120101100000 TO 20120102]\n' +
1838 '\n' +
1838 '\n' +
1839 'Generate wildcards using \'*\' character:\n' +
1839 'Generate wildcards using \'*\' character:\n' +
1840 ' "repository:vcs*" - search everything starting with \'vcs\'\n' +
1840 ' "repository:vcs*" - search everything starting with \'vcs\'\n' +
1841 ' "repository:*vcs*" - search for repository containing \'vcs\'\n' +
1841 ' "repository:*vcs*" - search for repository containing \'vcs\'\n' +
1842 '\n' +
1842 '\n' +
1843 'Optional AND / OR operators in queries\n' +
1843 'Optional AND / OR operators in queries\n' +
1844 ' "repository:vcs OR repository:test"\n' +
1844 ' "repository:vcs OR repository:test"\n' +
1845 ' "username:test AND repository:test*"\n'
1845 ' "username:test AND repository:test*"\n'
1846 )
1846 )
1847
1847
1848
1848
1849 def search_filter_help(searcher, request):
1849 def search_filter_help(searcher, request):
1850 _ = request.translate
1850 _ = request.translate
1851
1851
1852 terms = ''
1852 terms = ''
1853 return _(
1853 return _(
1854 'Example filter terms for `{searcher}` search:\n' +
1854 'Example filter terms for `{searcher}` search:\n' +
1855 '{terms}\n' +
1855 '{terms}\n' +
1856 'Generate wildcards using \'*\' character:\n' +
1856 'Generate wildcards using \'*\' character:\n' +
1857 ' "repo_name:vcs*" - search everything starting with \'vcs\'\n' +
1857 ' "repo_name:vcs*" - search everything starting with \'vcs\'\n' +
1858 ' "repo_name:*vcs*" - search for repository containing \'vcs\'\n' +
1858 ' "repo_name:*vcs*" - search for repository containing \'vcs\'\n' +
1859 '\n' +
1859 '\n' +
1860 'Optional AND / OR operators in queries\n' +
1860 'Optional AND / OR operators in queries\n' +
1861 ' "repo_name:vcs OR repo_name:test"\n' +
1861 ' "repo_name:vcs OR repo_name:test"\n' +
1862 ' "owner:test AND repo_name:test*"\n' +
1862 ' "owner:test AND repo_name:test*"\n' +
1863 'More: {search_doc}'
1863 'More: {search_doc}'
1864 ).format(searcher=searcher.name,
1864 ).format(searcher=searcher.name,
1865 terms=terms, search_doc=searcher.query_lang_doc)
1865 terms=terms, search_doc=searcher.query_lang_doc)
1866
1866
1867
1867
1868 def not_mapped_error(repo_name):
1868 def not_mapped_error(repo_name):
1869 from rhodecode.translation import _
1869 from rhodecode.translation import _
1870 flash(_('%s repository is not mapped to db perhaps'
1870 flash(_('%s repository is not mapped to db perhaps'
1871 ' it was created or renamed from the filesystem'
1871 ' it was created or renamed from the filesystem'
1872 ' please run the application again'
1872 ' please run the application again'
1873 ' in order to rescan repositories') % repo_name, category='error')
1873 ' in order to rescan repositories') % repo_name, category='error')
1874
1874
1875
1875
1876 def ip_range(ip_addr):
1876 def ip_range(ip_addr):
1877 from rhodecode.model.db import UserIpMap
1877 from rhodecode.model.db import UserIpMap
1878 s, e = UserIpMap._get_ip_range(ip_addr)
1878 s, e = UserIpMap._get_ip_range(ip_addr)
1879 return '%s - %s' % (s, e)
1879 return '%s - %s' % (s, e)
1880
1880
1881
1881
1882 def form(url, method='post', needs_csrf_token=True, **attrs):
1882 def form(url, method='post', needs_csrf_token=True, **attrs):
1883 """Wrapper around webhelpers.tags.form to prevent CSRF attacks."""
1883 """Wrapper around webhelpers.tags.form to prevent CSRF attacks."""
1884 if method.lower() != 'get' and needs_csrf_token:
1884 if method.lower() != 'get' and needs_csrf_token:
1885 raise Exception(
1885 raise Exception(
1886 'Forms to POST/PUT/DELETE endpoints should have (in general) a ' +
1886 'Forms to POST/PUT/DELETE endpoints should have (in general) a ' +
1887 'CSRF token. If the endpoint does not require such token you can ' +
1887 'CSRF token. If the endpoint does not require such token you can ' +
1888 'explicitly set the parameter needs_csrf_token to false.')
1888 'explicitly set the parameter needs_csrf_token to false.')
1889
1889
1890 return wh_form(url, method=method, **attrs)
1890 return wh_form(url, method=method, **attrs)
1891
1891
1892
1892
1893 def secure_form(url, method="POST", multipart=False, **attrs):
1893 def secure_form(url, method="POST", multipart=False, **attrs):
1894 """Start a form tag that points the action to an url. This
1894 """Start a form tag that points the action to an url. This
1895 form tag will also include the hidden field containing
1895 form tag will also include the hidden field containing
1896 the auth token.
1896 the auth token.
1897
1897
1898 The url options should be given either as a string, or as a
1898 The url options should be given either as a string, or as a
1899 ``url()`` function. The method for the form defaults to POST.
1899 ``url()`` function. The method for the form defaults to POST.
1900
1900
1901 Options:
1901 Options:
1902
1902
1903 ``multipart``
1903 ``multipart``
1904 If set to True, the enctype is set to "multipart/form-data".
1904 If set to True, the enctype is set to "multipart/form-data".
1905 ``method``
1905 ``method``
1906 The method to use when submitting the form, usually either
1906 The method to use when submitting the form, usually either
1907 "GET" or "POST". If "PUT", "DELETE", or another verb is used, a
1907 "GET" or "POST". If "PUT", "DELETE", or another verb is used, a
1908 hidden input with name _method is added to simulate the verb
1908 hidden input with name _method is added to simulate the verb
1909 over POST.
1909 over POST.
1910
1910
1911 """
1911 """
1912 from webhelpers.pylonslib.secure_form import insecure_form
1912 from webhelpers.pylonslib.secure_form import insecure_form
1913 form = insecure_form(url, method, multipart, **attrs)
1913 form = insecure_form(url, method, multipart, **attrs)
1914
1914
1915 session = None
1915 session = None
1916 # TODO(marcink): after pyramid migration require request variable ALWAYS
1916 # TODO(marcink): after pyramid migration require request variable ALWAYS
1917 if 'request' in attrs:
1917 if 'request' in attrs:
1918 session = attrs['request'].session
1918 session = attrs['request'].session
1919
1919
1920 token = literal(
1920 token = literal(
1921 '<input type="hidden" id="{}" name="{}" value="{}">'.format(
1921 '<input type="hidden" id="{}" name="{}" value="{}">'.format(
1922 csrf_token_key, csrf_token_key, get_csrf_token(session)))
1922 csrf_token_key, csrf_token_key, get_csrf_token(session)))
1923
1923
1924 return literal("%s\n%s" % (form, token))
1924 return literal("%s\n%s" % (form, token))
1925
1925
1926
1926
1927 def dropdownmenu(name, selected, options, enable_filter=False, **attrs):
1927 def dropdownmenu(name, selected, options, enable_filter=False, **attrs):
1928 select_html = select(name, selected, options, **attrs)
1928 select_html = select(name, selected, options, **attrs)
1929 select2 = """
1929 select2 = """
1930 <script>
1930 <script>
1931 $(document).ready(function() {
1931 $(document).ready(function() {
1932 $('#%s').select2({
1932 $('#%s').select2({
1933 containerCssClass: 'drop-menu',
1933 containerCssClass: 'drop-menu',
1934 dropdownCssClass: 'drop-menu-dropdown',
1934 dropdownCssClass: 'drop-menu-dropdown',
1935 dropdownAutoWidth: true%s
1935 dropdownAutoWidth: true%s
1936 });
1936 });
1937 });
1937 });
1938 </script>
1938 </script>
1939 """
1939 """
1940 filter_option = """,
1940 filter_option = """,
1941 minimumResultsForSearch: -1
1941 minimumResultsForSearch: -1
1942 """
1942 """
1943 input_id = attrs.get('id') or name
1943 input_id = attrs.get('id') or name
1944 filter_enabled = "" if enable_filter else filter_option
1944 filter_enabled = "" if enable_filter else filter_option
1945 select_script = literal(select2 % (input_id, filter_enabled))
1945 select_script = literal(select2 % (input_id, filter_enabled))
1946
1946
1947 return literal(select_html+select_script)
1947 return literal(select_html+select_script)
1948
1948
1949
1949
1950 def get_visual_attr(tmpl_context_var, attr_name):
1950 def get_visual_attr(tmpl_context_var, attr_name):
1951 """
1951 """
1952 A safe way to get a variable from visual variable of template context
1952 A safe way to get a variable from visual variable of template context
1953
1953
1954 :param tmpl_context_var: instance of tmpl_context, usually present as `c`
1954 :param tmpl_context_var: instance of tmpl_context, usually present as `c`
1955 :param attr_name: name of the attribute we fetch from the c.visual
1955 :param attr_name: name of the attribute we fetch from the c.visual
1956 """
1956 """
1957 visual = getattr(tmpl_context_var, 'visual', None)
1957 visual = getattr(tmpl_context_var, 'visual', None)
1958 if not visual:
1958 if not visual:
1959 return
1959 return
1960 else:
1960 else:
1961 return getattr(visual, attr_name, None)
1961 return getattr(visual, attr_name, None)
1962
1962
1963
1963
1964 def get_last_path_part(file_node):
1964 def get_last_path_part(file_node):
1965 if not file_node.path:
1965 if not file_node.path:
1966 return u''
1966 return u''
1967
1967
1968 path = safe_unicode(file_node.path.split('/')[-1])
1968 path = safe_unicode(file_node.path.split('/')[-1])
1969 return u'../' + path
1969 return u'../' + path
1970
1970
1971
1971
1972 def route_url(*args, **kwargs):
1972 def route_url(*args, **kwargs):
1973 """
1973 """
1974 Wrapper around pyramids `route_url` (fully qualified url) function.
1974 Wrapper around pyramids `route_url` (fully qualified url) function.
1975 It is used to generate URLs from within pylons views or templates.
1975 It is used to generate URLs from within pylons views or templates.
1976 This will be removed when pyramid migration if finished.
1976 This will be removed when pyramid migration if finished.
1977 """
1977 """
1978 req = get_current_request()
1978 req = get_current_request()
1979 return req.route_url(*args, **kwargs)
1979 return req.route_url(*args, **kwargs)
1980
1980
1981
1981
1982 def route_path(*args, **kwargs):
1982 def route_path(*args, **kwargs):
1983 """
1983 """
1984 Wrapper around pyramids `route_path` function. It is used to generate
1984 Wrapper around pyramids `route_path` function. It is used to generate
1985 URLs from within pylons views or templates. This will be removed when
1985 URLs from within pylons views or templates. This will be removed when
1986 pyramid migration if finished.
1986 pyramid migration if finished.
1987 """
1987 """
1988 req = get_current_request()
1988 req = get_current_request()
1989 return req.route_path(*args, **kwargs)
1989 return req.route_path(*args, **kwargs)
1990
1990
1991
1991
1992 def route_path_or_none(*args, **kwargs):
1992 def route_path_or_none(*args, **kwargs):
1993 try:
1993 try:
1994 return route_path(*args, **kwargs)
1994 return route_path(*args, **kwargs)
1995 except KeyError:
1995 except KeyError:
1996 return None
1996 return None
1997
1997
1998
1998
1999 def static_url(*args, **kwds):
1999 def static_url(*args, **kwds):
2000 """
2000 """
2001 Wrapper around pyramids `route_path` function. It is used to generate
2001 Wrapper around pyramids `route_path` function. It is used to generate
2002 URLs from within pylons views or templates. This will be removed when
2002 URLs from within pylons views or templates. This will be removed when
2003 pyramid migration if finished.
2003 pyramid migration if finished.
2004 """
2004 """
2005 req = get_current_request()
2005 req = get_current_request()
2006 return req.static_url(*args, **kwds)
2006 return req.static_url(*args, **kwds)
2007
2007
2008
2008
2009 def resource_path(*args, **kwds):
2009 def resource_path(*args, **kwds):
2010 """
2010 """
2011 Wrapper around pyramids `route_path` function. It is used to generate
2011 Wrapper around pyramids `route_path` function. It is used to generate
2012 URLs from within pylons views or templates. This will be removed when
2012 URLs from within pylons views or templates. This will be removed when
2013 pyramid migration if finished.
2013 pyramid migration if finished.
2014 """
2014 """
2015 req = get_current_request()
2015 req = get_current_request()
2016 return req.resource_path(*args, **kwds)
2016 return req.resource_path(*args, **kwds)
2017
2017
2018
2018
2019 def api_call_example(method, args):
2019 def api_call_example(method, args):
2020 """
2020 """
2021 Generates an API call example via CURL
2021 Generates an API call example via CURL
2022 """
2022 """
2023 args_json = json.dumps(OrderedDict([
2023 args_json = json.dumps(OrderedDict([
2024 ('id', 1),
2024 ('id', 1),
2025 ('auth_token', 'SECRET'),
2025 ('auth_token', 'SECRET'),
2026 ('method', method),
2026 ('method', method),
2027 ('args', args)
2027 ('args', args)
2028 ]))
2028 ]))
2029 return literal(
2029 return literal(
2030 "curl {api_url} -X POST -H 'content-type:text/plain' --data-binary '{data}'"
2030 "curl {api_url} -X POST -H 'content-type:text/plain' --data-binary '{data}'"
2031 "<br/><br/>SECRET can be found in <a href=\"{token_url}\">auth-tokens</a> page, "
2031 "<br/><br/>SECRET can be found in <a href=\"{token_url}\">auth-tokens</a> page, "
2032 "and needs to be of `api calls` role."
2032 "and needs to be of `api calls` role."
2033 .format(
2033 .format(
2034 api_url=route_url('apiv2'),
2034 api_url=route_url('apiv2'),
2035 token_url=route_url('my_account_auth_tokens'),
2035 token_url=route_url('my_account_auth_tokens'),
2036 data=args_json))
2036 data=args_json))
2037
2037
2038
2038
2039 def notification_description(notification, request):
2039 def notification_description(notification, request):
2040 """
2040 """
2041 Generate notification human readable description based on notification type
2041 Generate notification human readable description based on notification type
2042 """
2042 """
2043 from rhodecode.model.notification import NotificationModel
2043 from rhodecode.model.notification import NotificationModel
2044 return NotificationModel().make_description(
2044 return NotificationModel().make_description(
2045 notification, translate=request.translate)
2045 notification, translate=request.translate)
@@ -1,539 +1,540 b''
1 # -*- coding: utf-8 -*-
1 # -*- coding: utf-8 -*-
2
2
3 # Copyright (C) 2010-2017 RhodeCode GmbH
3 # Copyright (C) 2010-2017 RhodeCode GmbH
4 #
4 #
5 # This program is free software: you can redistribute it and/or modify
5 # This program is free software: you can redistribute it and/or modify
6 # it under the terms of the GNU Affero General Public License, version 3
6 # it under the terms of the GNU Affero General Public License, version 3
7 # (only), as published by the Free Software Foundation.
7 # (only), as published by the Free Software Foundation.
8 #
8 #
9 # This program is distributed in the hope that it will be useful,
9 # This program is distributed in the hope that it will be useful,
10 # but WITHOUT ANY WARRANTY; without even the implied warranty of
10 # but WITHOUT ANY WARRANTY; without even the implied warranty of
11 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
11 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 # GNU General Public License for more details.
12 # GNU General Public License for more details.
13 #
13 #
14 # You should have received a copy of the GNU Affero General Public License
14 # You should have received a copy of the GNU Affero General Public License
15 # along with this program. If not, see <http://www.gnu.org/licenses/>.
15 # along with this program. If not, see <http://www.gnu.org/licenses/>.
16 #
16 #
17 # This program is dual-licensed. If you wish to learn more about the
17 # This program is dual-licensed. If you wish to learn more about the
18 # RhodeCode Enterprise Edition, including its added features, Support services,
18 # RhodeCode Enterprise Edition, including its added features, Support services,
19 # and proprietary license terms, please see https://rhodecode.com/licenses/
19 # and proprietary license terms, please see https://rhodecode.com/licenses/
20
20
21
21
22 """
22 """
23 Package for testing various lib/helper functions in rhodecode
23 Package for testing various lib/helper functions in rhodecode
24 """
24 """
25
25
26 import datetime
26 import datetime
27 import string
27 import string
28 import mock
28 import mock
29 import pytest
29 import pytest
30
30
31 from rhodecode.tests import no_newline_id_generator
31 from rhodecode.tests import no_newline_id_generator
32 from rhodecode.tests.utils import run_test_concurrently
32 from rhodecode.tests.utils import run_test_concurrently
33 from rhodecode.lib.helpers import InitialsGravatar
33 from rhodecode.lib.helpers import InitialsGravatar
34
34
35 from rhodecode.lib.utils2 import AttributeDict
35 from rhodecode.lib.utils2 import AttributeDict
36 from rhodecode.model.db import Repository
36 from rhodecode.model.db import Repository
37
37
38
38
39 def _urls_for_proto(proto):
39 def _urls_for_proto(proto):
40 return [
40 return [
41 ('%s://127.0.0.1' % proto, ['%s://' % proto, '127.0.0.1'],
41 ('%s://127.0.0.1' % proto, ['%s://' % proto, '127.0.0.1'],
42 '%s://127.0.0.1' % proto),
42 '%s://127.0.0.1' % proto),
43 ('%s://marcink@127.0.0.1' % proto, ['%s://' % proto, '127.0.0.1'],
43 ('%s://marcink@127.0.0.1' % proto, ['%s://' % proto, '127.0.0.1'],
44 '%s://127.0.0.1' % proto),
44 '%s://127.0.0.1' % proto),
45 ('%s://marcink:pass@127.0.0.1' % proto, ['%s://' % proto, '127.0.0.1'],
45 ('%s://marcink:pass@127.0.0.1' % proto, ['%s://' % proto, '127.0.0.1'],
46 '%s://127.0.0.1' % proto),
46 '%s://127.0.0.1' % proto),
47 ('%s://127.0.0.1:8080' % proto, ['%s://' % proto, '127.0.0.1', '8080'],
47 ('%s://127.0.0.1:8080' % proto, ['%s://' % proto, '127.0.0.1', '8080'],
48 '%s://127.0.0.1:8080' % proto),
48 '%s://127.0.0.1:8080' % proto),
49 ('%s://domain.org' % proto, ['%s://' % proto, 'domain.org'],
49 ('%s://domain.org' % proto, ['%s://' % proto, 'domain.org'],
50 '%s://domain.org' % proto),
50 '%s://domain.org' % proto),
51 ('%s://user:pass@domain.org:8080' % proto,
51 ('%s://user:pass@domain.org:8080' % proto,
52 ['%s://' % proto, 'domain.org', '8080'],
52 ['%s://' % proto, 'domain.org', '8080'],
53 '%s://domain.org:8080' % proto),
53 '%s://domain.org:8080' % proto),
54 ]
54 ]
55
55
56 TEST_URLS = _urls_for_proto('http') + _urls_for_proto('https')
56 TEST_URLS = _urls_for_proto('http') + _urls_for_proto('https')
57
57
58
58
59 @pytest.mark.parametrize("test_url, expected, expected_creds", TEST_URLS)
59 @pytest.mark.parametrize("test_url, expected, expected_creds", TEST_URLS)
60 def test_uri_filter(test_url, expected, expected_creds):
60 def test_uri_filter(test_url, expected, expected_creds):
61 from rhodecode.lib.utils2 import uri_filter
61 from rhodecode.lib.utils2 import uri_filter
62 assert uri_filter(test_url) == expected
62 assert uri_filter(test_url) == expected
63
63
64
64
65 @pytest.mark.parametrize("test_url, expected, expected_creds", TEST_URLS)
65 @pytest.mark.parametrize("test_url, expected, expected_creds", TEST_URLS)
66 def test_credentials_filter(test_url, expected, expected_creds):
66 def test_credentials_filter(test_url, expected, expected_creds):
67 from rhodecode.lib.utils2 import credentials_filter
67 from rhodecode.lib.utils2 import credentials_filter
68 assert credentials_filter(test_url) == expected_creds
68 assert credentials_filter(test_url) == expected_creds
69
69
70
70
71 @pytest.mark.parametrize("str_bool, expected", [
71 @pytest.mark.parametrize("str_bool, expected", [
72 ('t', True),
72 ('t', True),
73 ('true', True),
73 ('true', True),
74 ('y', True),
74 ('y', True),
75 ('yes', True),
75 ('yes', True),
76 ('on', True),
76 ('on', True),
77 ('1', True),
77 ('1', True),
78 ('Y', True),
78 ('Y', True),
79 ('yeS', True),
79 ('yeS', True),
80 ('Y', True),
80 ('Y', True),
81 ('TRUE', True),
81 ('TRUE', True),
82 ('T', True),
82 ('T', True),
83 ('False', False),
83 ('False', False),
84 ('F', False),
84 ('F', False),
85 ('FALSE', False),
85 ('FALSE', False),
86 ('0', False),
86 ('0', False),
87 ('-1', False),
87 ('-1', False),
88 ('', False)
88 ('', False)
89 ])
89 ])
90 def test_str2bool(str_bool, expected):
90 def test_str2bool(str_bool, expected):
91 from rhodecode.lib.utils2 import str2bool
91 from rhodecode.lib.utils2 import str2bool
92 assert str2bool(str_bool) == expected
92 assert str2bool(str_bool) == expected
93
93
94
94
95 @pytest.mark.parametrize("text, expected", reduce(lambda a1,a2:a1+a2, [
95 @pytest.mark.parametrize("text, expected", reduce(lambda a1,a2:a1+a2, [
96 [
96 [
97 (pref+"", []),
97 (pref+"", []),
98 (pref+"Hi there @marcink", ['marcink']),
98 (pref+"Hi there @marcink", ['marcink']),
99 (pref+"Hi there @marcink and @bob", ['bob', 'marcink']),
99 (pref+"Hi there @marcink and @bob", ['bob', 'marcink']),
100 (pref+"Hi there @marcink\n", ['marcink']),
100 (pref+"Hi there @marcink\n", ['marcink']),
101 (pref+"Hi there @marcink and @bob\n", ['bob', 'marcink']),
101 (pref+"Hi there @marcink and @bob\n", ['bob', 'marcink']),
102 (pref+"Hi there marcin@rhodecode.com", []),
102 (pref+"Hi there marcin@rhodecode.com", []),
103 (pref+"Hi there @john.malcovic and @bob\n", ['bob', 'john.malcovic']),
103 (pref+"Hi there @john.malcovic and @bob\n", ['bob', 'john.malcovic']),
104 (pref+"This needs to be reviewed: (@marcink,@john)", ["john", "marcink"]),
104 (pref+"This needs to be reviewed: (@marcink,@john)", ["john", "marcink"]),
105 (pref+"This needs to be reviewed: (@marcink, @john)", ["john", "marcink"]),
105 (pref+"This needs to be reviewed: (@marcink, @john)", ["john", "marcink"]),
106 (pref+"This needs to be reviewed: [@marcink,@john]", ["john", "marcink"]),
106 (pref+"This needs to be reviewed: [@marcink,@john]", ["john", "marcink"]),
107 (pref+"This needs to be reviewed: (@marcink @john)", ["john", "marcink"]),
107 (pref+"This needs to be reviewed: (@marcink @john)", ["john", "marcink"]),
108 (pref+"@john @mary, please review", ["john", "mary"]),
108 (pref+"@john @mary, please review", ["john", "mary"]),
109 (pref+"@john,@mary, please review", ["john", "mary"]),
109 (pref+"@john,@mary, please review", ["john", "mary"]),
110 (pref+"Hej @123, @22john,@mary, please review", ['123', '22john', 'mary']),
110 (pref+"Hej @123, @22john,@mary, please review", ['123', '22john', 'mary']),
111 (pref+"@first hi there @marcink here's my email marcin@email.com "
111 (pref+"@first hi there @marcink here's my email marcin@email.com "
112 "@lukaszb check @one_more22 it pls @ ttwelve @D[] @one@two@three ", ['first', 'lukaszb', 'marcink', 'one', 'one_more22']),
112 "@lukaszb check @one_more22 it pls @ ttwelve @D[] @one@two@three ", ['first', 'lukaszb', 'marcink', 'one', 'one_more22']),
113 (pref+"@MARCIN @maRCiN @2one_more22 @john please see this http://org.pl", ['2one_more22', 'john', 'MARCIN', 'maRCiN']),
113 (pref+"@MARCIN @maRCiN @2one_more22 @john please see this http://org.pl", ['2one_more22', 'john', 'MARCIN', 'maRCiN']),
114 (pref+"@marian.user just do it @marco-polo and next extract @marco_polo", ['marco-polo', 'marco_polo', 'marian.user']),
114 (pref+"@marian.user just do it @marco-polo and next extract @marco_polo", ['marco-polo', 'marco_polo', 'marian.user']),
115 (pref+"user.dot hej ! not-needed maril@domain.org", []),
115 (pref+"user.dot hej ! not-needed maril@domain.org", []),
116 (pref+"\n@marcin", ['marcin']),
116 (pref+"\n@marcin", ['marcin']),
117 ]
117 ]
118 for pref in ['', '\n', 'hi !', '\t', '\n\n']]), ids=no_newline_id_generator)
118 for pref in ['', '\n', 'hi !', '\t', '\n\n']]), ids=no_newline_id_generator)
119 def test_mention_extractor(text, expected):
119 def test_mention_extractor(text, expected):
120 from rhodecode.lib.utils2 import extract_mentioned_users
120 from rhodecode.lib.utils2 import extract_mentioned_users
121 got = extract_mentioned_users(text)
121 got = extract_mentioned_users(text)
122 assert sorted(got, key=lambda x: x.lower()) == got
122 assert sorted(got, key=lambda x: x.lower()) == got
123 assert set(expected) == set(got)
123 assert set(expected) == set(got)
124
124
125 @pytest.mark.parametrize("age_args, expected, kw", [
125 @pytest.mark.parametrize("age_args, expected, kw", [
126 ({}, u'just now', {}),
126 ({}, u'just now', {}),
127 ({'seconds': -1}, u'1 second ago', {}),
127 ({'seconds': -1}, u'1 second ago', {}),
128 ({'seconds': -60 * 2}, u'2 minutes ago', {}),
128 ({'seconds': -60 * 2}, u'2 minutes ago', {}),
129 ({'hours': -1}, u'1 hour ago', {}),
129 ({'hours': -1}, u'1 hour ago', {}),
130 ({'hours': -24}, u'1 day ago', {}),
130 ({'hours': -24}, u'1 day ago', {}),
131 ({'hours': -24 * 5}, u'5 days ago', {}),
131 ({'hours': -24 * 5}, u'5 days ago', {}),
132 ({'months': -1}, u'1 month ago', {}),
132 ({'months': -1}, u'1 month ago', {}),
133 ({'months': -1, 'days': -2}, u'1 month and 2 days ago', {}),
133 ({'months': -1, 'days': -2}, u'1 month and 2 days ago', {}),
134 ({'years': -1, 'months': -1}, u'1 year and 1 month ago', {}),
134 ({'years': -1, 'months': -1}, u'1 year and 1 month ago', {}),
135 ({}, u'just now', {'short_format': True}),
135 ({}, u'just now', {'short_format': True}),
136 ({'seconds': -1}, u'1sec ago', {'short_format': True}),
136 ({'seconds': -1}, u'1sec ago', {'short_format': True}),
137 ({'seconds': -60 * 2}, u'2min ago', {'short_format': True}),
137 ({'seconds': -60 * 2}, u'2min ago', {'short_format': True}),
138 ({'hours': -1}, u'1h ago', {'short_format': True}),
138 ({'hours': -1}, u'1h ago', {'short_format': True}),
139 ({'hours': -24}, u'1d ago', {'short_format': True}),
139 ({'hours': -24}, u'1d ago', {'short_format': True}),
140 ({'hours': -24 * 5}, u'5d ago', {'short_format': True}),
140 ({'hours': -24 * 5}, u'5d ago', {'short_format': True}),
141 ({'months': -1}, u'1m ago', {'short_format': True}),
141 ({'months': -1}, u'1m ago', {'short_format': True}),
142 ({'months': -1, 'days': -2}, u'1m, 2d ago', {'short_format': True}),
142 ({'months': -1, 'days': -2}, u'1m, 2d ago', {'short_format': True}),
143 ({'years': -1, 'months': -1}, u'1y, 1m ago', {'short_format': True}),
143 ({'years': -1, 'months': -1}, u'1y, 1m ago', {'short_format': True}),
144 ])
144 ])
145 def test_age(age_args, expected, kw, pylonsapp):
145 def test_age(age_args, expected, kw, pylonsapp):
146 from rhodecode.lib.utils2 import age
146 from rhodecode.lib.utils2 import age
147 from dateutil import relativedelta
147 from dateutil import relativedelta
148 n = datetime.datetime(year=2012, month=5, day=17)
148 n = datetime.datetime(year=2012, month=5, day=17)
149 delt = lambda *args, **kwargs: relativedelta.relativedelta(*args, **kwargs)
149 delt = lambda *args, **kwargs: relativedelta.relativedelta(*args, **kwargs)
150
150
151 def translate(elem):
151 def translate(elem):
152 return elem.interpolate()
152 return elem.interpolate()
153
153
154 assert translate(age(n + delt(**age_args), now=n, **kw)) == expected
154 assert translate(age(n + delt(**age_args), now=n, **kw)) == expected
155
155
156
156
157 @pytest.mark.parametrize("age_args, expected, kw", [
157 @pytest.mark.parametrize("age_args, expected, kw", [
158 ({}, u'just now', {}),
158 ({}, u'just now', {}),
159 ({'seconds': 1}, u'in 1 second', {}),
159 ({'seconds': 1}, u'in 1 second', {}),
160 ({'seconds': 60 * 2}, u'in 2 minutes', {}),
160 ({'seconds': 60 * 2}, u'in 2 minutes', {}),
161 ({'hours': 1}, u'in 1 hour', {}),
161 ({'hours': 1}, u'in 1 hour', {}),
162 ({'hours': 24}, u'in 1 day', {}),
162 ({'hours': 24}, u'in 1 day', {}),
163 ({'hours': 24 * 5}, u'in 5 days', {}),
163 ({'hours': 24 * 5}, u'in 5 days', {}),
164 ({'months': 1}, u'in 1 month', {}),
164 ({'months': 1}, u'in 1 month', {}),
165 ({'months': 1, 'days': 1}, u'in 1 month and 1 day', {}),
165 ({'months': 1, 'days': 1}, u'in 1 month and 1 day', {}),
166 ({'years': 1, 'months': 1}, u'in 1 year and 1 month', {}),
166 ({'years': 1, 'months': 1}, u'in 1 year and 1 month', {}),
167 ({}, u'just now', {'short_format': True}),
167 ({}, u'just now', {'short_format': True}),
168 ({'seconds': 1}, u'in 1sec', {'short_format': True}),
168 ({'seconds': 1}, u'in 1sec', {'short_format': True}),
169 ({'seconds': 60 * 2}, u'in 2min', {'short_format': True}),
169 ({'seconds': 60 * 2}, u'in 2min', {'short_format': True}),
170 ({'hours': 1}, u'in 1h', {'short_format': True}),
170 ({'hours': 1}, u'in 1h', {'short_format': True}),
171 ({'hours': 24}, u'in 1d', {'short_format': True}),
171 ({'hours': 24}, u'in 1d', {'short_format': True}),
172 ({'hours': 24 * 5}, u'in 5d', {'short_format': True}),
172 ({'hours': 24 * 5}, u'in 5d', {'short_format': True}),
173 ({'months': 1}, u'in 1m', {'short_format': True}),
173 ({'months': 1}, u'in 1m', {'short_format': True}),
174 ({'months': 1, 'days': 1}, u'in 1m, 1d', {'short_format': True}),
174 ({'months': 1, 'days': 1}, u'in 1m, 1d', {'short_format': True}),
175 ({'years': 1, 'months': 1}, u'in 1y, 1m', {'short_format': True}),
175 ({'years': 1, 'months': 1}, u'in 1y, 1m', {'short_format': True}),
176 ])
176 ])
177 def test_age_in_future(age_args, expected, kw, pylonsapp):
177 def test_age_in_future(age_args, expected, kw, pylonsapp):
178 from rhodecode.lib.utils2 import age
178 from rhodecode.lib.utils2 import age
179 from dateutil import relativedelta
179 from dateutil import relativedelta
180 n = datetime.datetime(year=2012, month=5, day=17)
180 n = datetime.datetime(year=2012, month=5, day=17)
181 delt = lambda *args, **kwargs: relativedelta.relativedelta(*args, **kwargs)
181 delt = lambda *args, **kwargs: relativedelta.relativedelta(*args, **kwargs)
182
182
183 def translate(elem):
183 def translate(elem):
184 return elem.interpolate()
184 return elem.interpolate()
185
185
186 assert translate(age(n + delt(**age_args), now=n, **kw)) == expected
186 assert translate(age(n + delt(**age_args), now=n, **kw)) == expected
187
187
188
188
189 def test_tag_exctrator():
189 def test_tag_exctrator():
190 sample = (
190 sample = (
191 "hello pta[tag] gog [[]] [[] sda ero[or]d [me =>>< sa]"
191 "hello pta[tag] gog [[]] [[] sda ero[or]d [me =>>< sa]"
192 "[requires] [stale] [see<>=>] [see => http://url.com]"
192 "[requires] [stale] [see<>=>] [see => http://url.com]"
193 "[requires => url] [lang => python] [just a tag] <html_tag first='abc' attr=\"my.url?attr=&another=\"></html_tag>"
193 "[requires => url] [lang => python] [just a tag] <html_tag first='abc' attr=\"my.url?attr=&another=\"></html_tag>"
194 "[,d] [ => ULR ] [obsolete] [desc]]"
194 "[,d] [ => ULR ] [obsolete] [desc]]"
195 )
195 )
196 from rhodecode.lib.helpers import desc_stylize, escaped_stylize
196 from rhodecode.lib.helpers import desc_stylize, escaped_stylize
197 res = desc_stylize(sample)
197 res = desc_stylize(sample)
198 assert '<div class="metatag" tag="tag">tag</div>' in res
198 assert '<div class="metatag" tag="tag">tag</div>' in res
199 assert '<div class="metatag" tag="obsolete">obsolete</div>' in res
199 assert '<div class="metatag" tag="obsolete">obsolete</div>' in res
200 assert '<div class="metatag" tag="stale">stale</div>' in res
200 assert '<div class="metatag" tag="stale">stale</div>' in res
201 assert '<div class="metatag" tag="lang">python</div>' in res
201 assert '<div class="metatag" tag="lang">python</div>' in res
202 assert '<div class="metatag" tag="requires">requires =&gt; <a href="/url">url</a></div>' in res
202 assert '<div class="metatag" tag="requires">requires =&gt; <a href="/url">url</a></div>' in res
203 assert '<div class="metatag" tag="tag">tag</div>' in res
203 assert '<div class="metatag" tag="tag">tag</div>' in res
204 assert '<html_tag first=\'abc\' attr=\"my.url?attr=&another=\"></html_tag>' in res
204 assert '<html_tag first=\'abc\' attr=\"my.url?attr=&another=\"></html_tag>' in res
205
205
206 res_encoded = escaped_stylize(sample)
206 res_encoded = escaped_stylize(sample)
207 assert '<div class="metatag" tag="tag">tag</div>' in res_encoded
207 assert '<div class="metatag" tag="tag">tag</div>' in res_encoded
208 assert '<div class="metatag" tag="obsolete">obsolete</div>' in res_encoded
208 assert '<div class="metatag" tag="obsolete">obsolete</div>' in res_encoded
209 assert '<div class="metatag" tag="stale">stale</div>' in res_encoded
209 assert '<div class="metatag" tag="stale">stale</div>' in res_encoded
210 assert '<div class="metatag" tag="lang">python</div>' in res_encoded
210 assert '<div class="metatag" tag="lang">python</div>' in res_encoded
211 assert '<div class="metatag" tag="requires">requires =&gt; <a href="/url">url</a></div>' in res_encoded
211 assert '<div class="metatag" tag="requires">requires =&gt; <a href="/url">url</a></div>' in res_encoded
212 assert '<div class="metatag" tag="tag">tag</div>' in res_encoded
212 assert '<div class="metatag" tag="tag">tag</div>' in res_encoded
213 assert '&lt;html_tag first=&#39;abc&#39; attr=&#34;my.url?attr=&amp;another=&#34;&gt;&lt;/html_tag&gt;' in res_encoded
213 assert '&lt;html_tag first=&#39;abc&#39; attr=&#34;my.url?attr=&amp;another=&#34;&gt;&lt;/html_tag&gt;' in res_encoded
214
214
215
215
216 @pytest.mark.parametrize("tmpl_url, email, expected", [
216 @pytest.mark.parametrize("tmpl_url, email, expected", [
217 ('http://test.com/{email}', 'test@foo.com', 'http://test.com/test@foo.com'),
217 ('http://test.com/{email}', 'test@foo.com', 'http://test.com/test@foo.com'),
218
218
219 ('http://test.com/{md5email}', 'test@foo.com', 'http://test.com/3cb7232fcc48743000cb86d0d5022bd9'),
219 ('http://test.com/{md5email}', 'test@foo.com', 'http://test.com/3cb7232fcc48743000cb86d0d5022bd9'),
220 ('http://test.com/{md5email}', 'testΔ…Δ‡@foo.com', 'http://test.com/978debb907a3c55cd741872ab293ef30'),
220 ('http://test.com/{md5email}', 'testΔ…Δ‡@foo.com', 'http://test.com/978debb907a3c55cd741872ab293ef30'),
221
221
222 ('http://testX.com/{md5email}?s={size}', 'test@foo.com', 'http://testX.com/3cb7232fcc48743000cb86d0d5022bd9?s=24'),
222 ('http://testX.com/{md5email}?s={size}', 'test@foo.com', 'http://testX.com/3cb7232fcc48743000cb86d0d5022bd9?s=24'),
223 ('http://testX.com/{md5email}?s={size}', 'testΔ…Δ‡@foo.com', 'http://testX.com/978debb907a3c55cd741872ab293ef30?s=24'),
223 ('http://testX.com/{md5email}?s={size}', 'testΔ…Δ‡@foo.com', 'http://testX.com/978debb907a3c55cd741872ab293ef30?s=24'),
224
224
225 ('{scheme}://{netloc}/{md5email}/{size}', 'test@foo.com', 'https://server.com/3cb7232fcc48743000cb86d0d5022bd9/24'),
225 ('{scheme}://{netloc}/{md5email}/{size}', 'test@foo.com', 'https://server.com/3cb7232fcc48743000cb86d0d5022bd9/24'),
226 ('{scheme}://{netloc}/{md5email}/{size}', 'testΔ…Δ‡@foo.com', 'https://server.com/978debb907a3c55cd741872ab293ef30/24'),
226 ('{scheme}://{netloc}/{md5email}/{size}', 'testΔ…Δ‡@foo.com', 'https://server.com/978debb907a3c55cd741872ab293ef30/24'),
227
227
228 ('http://test.com/{email}', 'testΔ…Δ‡@foo.com', 'http://test.com/testΔ…Δ‡@foo.com'),
228 ('http://test.com/{email}', 'testΔ…Δ‡@foo.com', 'http://test.com/testΔ…Δ‡@foo.com'),
229 ('http://test.com/{email}?size={size}', 'test@foo.com', 'http://test.com/test@foo.com?size=24'),
229 ('http://test.com/{email}?size={size}', 'test@foo.com', 'http://test.com/test@foo.com?size=24'),
230 ('http://test.com/{email}?size={size}', 'testΔ…Δ‡@foo.com', 'http://test.com/testΔ…Δ‡@foo.com?size=24'),
230 ('http://test.com/{email}?size={size}', 'testΔ…Δ‡@foo.com', 'http://test.com/testΔ…Δ‡@foo.com?size=24'),
231 ])
231 ])
232 def test_gravatar_url_builder(tmpl_url, email, expected, request_stub):
232 def test_gravatar_url_builder(tmpl_url, email, expected, request_stub):
233 from rhodecode.lib.helpers import gravatar_url
233 from rhodecode.lib.helpers import gravatar_url
234
234
235 # mock pyramid.threadlocals
235 # mock pyramid.threadlocals
236 def fake_get_current_request():
236 def fake_get_current_request():
237 request_stub.scheme = 'https'
237 request_stub.scheme = 'https'
238 request_stub.host = 'server.com'
238 request_stub.host = 'server.com'
239 return request_stub
239 return request_stub
240
240
241 # mock pylons.tmpl_context
241 # mock pylons.tmpl_context
242 def fake_tmpl_context(_url):
242 def fake_tmpl_context(_url):
243 _c = AttributeDict()
243 _c = AttributeDict()
244 _c.visual = AttributeDict()
244 _c.visual = AttributeDict()
245 _c.visual.use_gravatar = True
245 _c.visual.use_gravatar = True
246 _c.visual.gravatar_url = _url
246 _c.visual.gravatar_url = _url
247
247
248 return _c
248 return _c
249
249
250 with mock.patch('rhodecode.lib.helpers.get_current_request',
250 with mock.patch('rhodecode.lib.helpers.get_current_request',
251 fake_get_current_request):
251 fake_get_current_request):
252 fake = fake_tmpl_context(_url=tmpl_url)
252 fake = fake_tmpl_context(_url=tmpl_url)
253 with mock.patch('pylons.tmpl_context', fake):
253 with mock.patch('pylons.tmpl_context', fake):
254 grav = gravatar_url(email_address=email, size=24)
254 grav = gravatar_url(email_address=email, size=24)
255 assert grav == expected
255 assert grav == expected
256
256
257
257
258 @pytest.mark.parametrize(
258 @pytest.mark.parametrize(
259 "email, first_name, last_name, expected_initials, expected_color", [
259 "email, first_name, last_name, expected_initials, expected_color", [
260
260
261 ('test@rhodecode.com', '', '', 'TR', '#8a994d'),
261 ('test@rhodecode.com', '', '', 'TR', '#8a994d'),
262 ('marcin.kuzminski@rhodecode.com', '', '', 'MK', '#6559b3'),
262 ('marcin.kuzminski@rhodecode.com', '', '', 'MK', '#6559b3'),
263 # special cases of email
263 # special cases of email
264 ('john.van.dam@rhodecode.com', '', '', 'JD', '#526600'),
264 ('john.van.dam@rhodecode.com', '', '', 'JD', '#526600'),
265 ('Guido.van.Rossum@rhodecode.com', '', '', 'GR', '#990052'),
265 ('Guido.van.Rossum@rhodecode.com', '', '', 'GR', '#990052'),
266 ('Guido.van.Rossum@rhodecode.com', 'Guido', 'Van Rossum', 'GR', '#990052'),
266 ('Guido.van.Rossum@rhodecode.com', 'Guido', 'Van Rossum', 'GR', '#990052'),
267
267
268 ('rhodecode+Guido.van.Rossum@rhodecode.com', '', '', 'RR', '#46598c'),
268 ('rhodecode+Guido.van.Rossum@rhodecode.com', '', '', 'RR', '#46598c'),
269 ('pclouds@rhodecode.com', 'Nguyα»…n ThΓ‘i', 'Tgọc Duy', 'ND', '#665200'),
269 ('pclouds@rhodecode.com', 'Nguyα»…n ThΓ‘i', 'Tgọc Duy', 'ND', '#665200'),
270
270
271 ('john-brown@foo.com', '', '', 'JF', '#73006b'),
271 ('john-brown@foo.com', '', '', 'JF', '#73006b'),
272 ('admin@rhodecode.com', 'Marcin', 'Kuzminski', 'MK', '#104036'),
272 ('admin@rhodecode.com', 'Marcin', 'Kuzminski', 'MK', '#104036'),
273 # partials
273 # partials
274 ('admin@rhodecode.com', 'Marcin', '', 'MR', '#104036'), # fn+email
274 ('admin@rhodecode.com', 'Marcin', '', 'MR', '#104036'), # fn+email
275 ('admin@rhodecode.com', '', 'Kuzminski', 'AK', '#104036'), # em+ln
275 ('admin@rhodecode.com', '', 'Kuzminski', 'AK', '#104036'), # em+ln
276 # non-ascii
276 # non-ascii
277 ('admin@rhodecode.com', 'Marcin', 'Śuzminski', 'MS', '#104036'),
277 ('admin@rhodecode.com', 'Marcin', 'Śuzminski', 'MS', '#104036'),
278 ('marcin.Ε›uzminski@rhodecode.com', '', '', 'MS', '#73000f'),
278 ('marcin.Ε›uzminski@rhodecode.com', '', '', 'MS', '#73000f'),
279
279
280 # special cases, LDAP can provide those...
280 # special cases, LDAP can provide those...
281 ('admin@', 'Marcin', 'Śuzminski', 'MS', '#aa00ff'),
281 ('admin@', 'Marcin', 'Śuzminski', 'MS', '#aa00ff'),
282 ('marcin.Ε›uzminski', '', '', 'MS', '#402020'),
282 ('marcin.Ε›uzminski', '', '', 'MS', '#402020'),
283 ('null', '', '', 'NL', '#8c4646'),
283 ('null', '', '', 'NL', '#8c4646'),
284 ('some.@abc.com', 'some', '', 'SA', '#664e33')
284 ])
285 ])
285 def test_initials_gravatar_pick_of_initials_and_color_algo(
286 def test_initials_gravatar_pick_of_initials_and_color_algo(
286 email, first_name, last_name, expected_initials, expected_color):
287 email, first_name, last_name, expected_initials, expected_color):
287 instance = InitialsGravatar(email, first_name, last_name)
288 instance = InitialsGravatar(email, first_name, last_name)
288 assert instance.get_initials() == expected_initials
289 assert instance.get_initials() == expected_initials
289 assert instance.str2color(email) == expected_color
290 assert instance.str2color(email) == expected_color
290
291
291
292
292 def test_initials_gravatar_mapping_algo():
293 def test_initials_gravatar_mapping_algo():
293 pos = set()
294 pos = set()
294 instance = InitialsGravatar('', '', '')
295 instance = InitialsGravatar('', '', '')
295 iterations = 0
296 iterations = 0
296
297
297 variations = []
298 variations = []
298 for letter1 in string.ascii_letters:
299 for letter1 in string.ascii_letters:
299 for letter2 in string.ascii_letters[::-1][:10]:
300 for letter2 in string.ascii_letters[::-1][:10]:
300 for letter3 in string.ascii_letters[:10]:
301 for letter3 in string.ascii_letters[:10]:
301 variations.append(
302 variations.append(
302 '%s@rhodecode.com' % (letter1+letter2+letter3))
303 '%s@rhodecode.com' % (letter1+letter2+letter3))
303
304
304 max_variations = 4096
305 max_variations = 4096
305 for email in variations[:max_variations]:
306 for email in variations[:max_variations]:
306 iterations += 1
307 iterations += 1
307 pos.add(
308 pos.add(
308 instance.pick_color_bank_index(email,
309 instance.pick_color_bank_index(email,
309 instance.get_color_bank()))
310 instance.get_color_bank()))
310
311
311 # we assume that we have match all 256 possible positions,
312 # we assume that we have match all 256 possible positions,
312 # in reasonable amount of different email addresses
313 # in reasonable amount of different email addresses
313 assert len(pos) == 256
314 assert len(pos) == 256
314 assert iterations == max_variations
315 assert iterations == max_variations
315
316
316
317
317 @pytest.mark.parametrize("tmpl, repo_name, overrides, prefix, expected", [
318 @pytest.mark.parametrize("tmpl, repo_name, overrides, prefix, expected", [
318 (Repository.DEFAULT_CLONE_URI, 'group/repo1', {}, '', 'http://vps1:8000/group/repo1'),
319 (Repository.DEFAULT_CLONE_URI, 'group/repo1', {}, '', 'http://vps1:8000/group/repo1'),
319 (Repository.DEFAULT_CLONE_URI, 'group/repo1', {'user': 'marcink'}, '', 'http://marcink@vps1:8000/group/repo1'),
320 (Repository.DEFAULT_CLONE_URI, 'group/repo1', {'user': 'marcink'}, '', 'http://marcink@vps1:8000/group/repo1'),
320 (Repository.DEFAULT_CLONE_URI, 'group/repo1', {}, '/rc', 'http://vps1:8000/rc/group/repo1'),
321 (Repository.DEFAULT_CLONE_URI, 'group/repo1', {}, '/rc', 'http://vps1:8000/rc/group/repo1'),
321 (Repository.DEFAULT_CLONE_URI, 'group/repo1', {'user': 'user'}, '/rc', 'http://user@vps1:8000/rc/group/repo1'),
322 (Repository.DEFAULT_CLONE_URI, 'group/repo1', {'user': 'user'}, '/rc', 'http://user@vps1:8000/rc/group/repo1'),
322 (Repository.DEFAULT_CLONE_URI, 'group/repo1', {'user': 'marcink'}, '/rc', 'http://marcink@vps1:8000/rc/group/repo1'),
323 (Repository.DEFAULT_CLONE_URI, 'group/repo1', {'user': 'marcink'}, '/rc', 'http://marcink@vps1:8000/rc/group/repo1'),
323 (Repository.DEFAULT_CLONE_URI, 'group/repo1', {'user': 'user'}, '/rc/', 'http://user@vps1:8000/rc/group/repo1'),
324 (Repository.DEFAULT_CLONE_URI, 'group/repo1', {'user': 'user'}, '/rc/', 'http://user@vps1:8000/rc/group/repo1'),
324 (Repository.DEFAULT_CLONE_URI, 'group/repo1', {'user': 'marcink'}, '/rc/', 'http://marcink@vps1:8000/rc/group/repo1'),
325 (Repository.DEFAULT_CLONE_URI, 'group/repo1', {'user': 'marcink'}, '/rc/', 'http://marcink@vps1:8000/rc/group/repo1'),
325 ('{scheme}://{user}@{netloc}/_{repoid}', 'group/repo1', {}, '', 'http://vps1:8000/_23'),
326 ('{scheme}://{user}@{netloc}/_{repoid}', 'group/repo1', {}, '', 'http://vps1:8000/_23'),
326 ('{scheme}://{user}@{netloc}/_{repoid}', 'group/repo1', {'user': 'marcink'}, '', 'http://marcink@vps1:8000/_23'),
327 ('{scheme}://{user}@{netloc}/_{repoid}', 'group/repo1', {'user': 'marcink'}, '', 'http://marcink@vps1:8000/_23'),
327 ('http://{user}@{netloc}/_{repoid}', 'group/repo1', {'user': 'marcink'}, '', 'http://marcink@vps1:8000/_23'),
328 ('http://{user}@{netloc}/_{repoid}', 'group/repo1', {'user': 'marcink'}, '', 'http://marcink@vps1:8000/_23'),
328 ('http://{netloc}/_{repoid}', 'group/repo1', {'user': 'marcink'}, '', 'http://vps1:8000/_23'),
329 ('http://{netloc}/_{repoid}', 'group/repo1', {'user': 'marcink'}, '', 'http://vps1:8000/_23'),
329 ('https://{user}@proxy1.server.com/{repo}', 'group/repo1', {'user': 'marcink'}, '', 'https://marcink@proxy1.server.com/group/repo1'),
330 ('https://{user}@proxy1.server.com/{repo}', 'group/repo1', {'user': 'marcink'}, '', 'https://marcink@proxy1.server.com/group/repo1'),
330 ('https://{user}@proxy1.server.com/{repo}', 'group/repo1', {}, '', 'https://proxy1.server.com/group/repo1'),
331 ('https://{user}@proxy1.server.com/{repo}', 'group/repo1', {}, '', 'https://proxy1.server.com/group/repo1'),
331 ('https://proxy1.server.com/{user}/{repo}', 'group/repo1', {'user': 'marcink'}, '', 'https://proxy1.server.com/marcink/group/repo1'),
332 ('https://proxy1.server.com/{user}/{repo}', 'group/repo1', {'user': 'marcink'}, '', 'https://proxy1.server.com/marcink/group/repo1'),
332 ])
333 ])
333 def test_clone_url_generator(tmpl, repo_name, overrides, prefix, expected):
334 def test_clone_url_generator(tmpl, repo_name, overrides, prefix, expected):
334 from rhodecode.lib.utils2 import get_clone_url
335 from rhodecode.lib.utils2 import get_clone_url
335
336
336 class RequestStub(object):
337 class RequestStub(object):
337 def request_url(self, name):
338 def request_url(self, name):
338 return 'http://vps1:8000' + prefix
339 return 'http://vps1:8000' + prefix
339
340
340 def route_url(self, name):
341 def route_url(self, name):
341 return self.request_url(name)
342 return self.request_url(name)
342
343
343 clone_url = get_clone_url(
344 clone_url = get_clone_url(
344 request=RequestStub(),
345 request=RequestStub(),
345 uri_tmpl=tmpl,
346 uri_tmpl=tmpl,
346 repo_name=repo_name, repo_id=23, **overrides)
347 repo_name=repo_name, repo_id=23, **overrides)
347 assert clone_url == expected
348 assert clone_url == expected
348
349
349
350
350 def _quick_url(text, tmpl="""<a class="revision-link" href="%s">%s</a>""", url_=None):
351 def _quick_url(text, tmpl="""<a class="revision-link" href="%s">%s</a>""", url_=None):
351 """
352 """
352 Changes `some text url[foo]` => `some text <a href="/">foo</a>
353 Changes `some text url[foo]` => `some text <a href="/">foo</a>
353
354
354 :param text:
355 :param text:
355 """
356 """
356 import re
357 import re
357 # quickly change expected url[] into a link
358 # quickly change expected url[] into a link
358 URL_PAT = re.compile(r'(?:url\[)(.+?)(?:\])')
359 URL_PAT = re.compile(r'(?:url\[)(.+?)(?:\])')
359
360
360 def url_func(match_obj):
361 def url_func(match_obj):
361 _url = match_obj.groups()[0]
362 _url = match_obj.groups()[0]
362 return tmpl % (url_ or '/some-url', _url)
363 return tmpl % (url_ or '/some-url', _url)
363 return URL_PAT.sub(url_func, text)
364 return URL_PAT.sub(url_func, text)
364
365
365
366
366 @pytest.mark.parametrize("sample, expected", [
367 @pytest.mark.parametrize("sample, expected", [
367 ("",
368 ("",
368 ""),
369 ""),
369 ("git-svn-id: https://svn.apache.org/repos/asf/libcloud/trunk@1441655 13f79535-47bb-0310-9956-ffa450edef68",
370 ("git-svn-id: https://svn.apache.org/repos/asf/libcloud/trunk@1441655 13f79535-47bb-0310-9956-ffa450edef68",
370 "git-svn-id: https://svn.apache.org/repos/asf/libcloud/trunk@1441655 13f79535-47bb-0310-9956-ffa450edef68"),
371 "git-svn-id: https://svn.apache.org/repos/asf/libcloud/trunk@1441655 13f79535-47bb-0310-9956-ffa450edef68"),
371 ("from rev 000000000000",
372 ("from rev 000000000000",
372 "from rev url[000000000000]"),
373 "from rev url[000000000000]"),
373 ("from rev 000000000000123123 also rev 000000000000",
374 ("from rev 000000000000123123 also rev 000000000000",
374 "from rev url[000000000000123123] also rev url[000000000000]"),
375 "from rev url[000000000000123123] also rev url[000000000000]"),
375 ("this should-000 00",
376 ("this should-000 00",
376 "this should-000 00"),
377 "this should-000 00"),
377 ("longtextffffffffff rev 123123123123",
378 ("longtextffffffffff rev 123123123123",
378 "longtextffffffffff rev url[123123123123]"),
379 "longtextffffffffff rev url[123123123123]"),
379 ("rev ffffffffffffffffffffffffffffffffffffffffffffffffff",
380 ("rev ffffffffffffffffffffffffffffffffffffffffffffffffff",
380 "rev ffffffffffffffffffffffffffffffffffffffffffffffffff"),
381 "rev ffffffffffffffffffffffffffffffffffffffffffffffffff"),
381 ("ffffffffffff some text traalaa",
382 ("ffffffffffff some text traalaa",
382 "url[ffffffffffff] some text traalaa"),
383 "url[ffffffffffff] some text traalaa"),
383 ("""Multi line
384 ("""Multi line
384 123123123123
385 123123123123
385 some text 123123123123
386 some text 123123123123
386 sometimes !
387 sometimes !
387 """,
388 """,
388 """Multi line
389 """Multi line
389 url[123123123123]
390 url[123123123123]
390 some text url[123123123123]
391 some text url[123123123123]
391 sometimes !
392 sometimes !
392 """)
393 """)
393 ], ids=no_newline_id_generator)
394 ], ids=no_newline_id_generator)
394 def test_urlify_commits(sample, expected):
395 def test_urlify_commits(sample, expected):
395 def fake_url(self, *args, **kwargs):
396 def fake_url(self, *args, **kwargs):
396 return '/some-url'
397 return '/some-url'
397
398
398 expected = _quick_url(expected)
399 expected = _quick_url(expected)
399
400
400 with mock.patch('rhodecode.lib.helpers.route_url', fake_url):
401 with mock.patch('rhodecode.lib.helpers.route_url', fake_url):
401 from rhodecode.lib.helpers import urlify_commits
402 from rhodecode.lib.helpers import urlify_commits
402 assert urlify_commits(sample, 'repo_name') == expected
403 assert urlify_commits(sample, 'repo_name') == expected
403
404
404
405
405 @pytest.mark.parametrize("sample, expected, url_", [
406 @pytest.mark.parametrize("sample, expected, url_", [
406 ("",
407 ("",
407 "",
408 "",
408 ""),
409 ""),
409 ("https://svn.apache.org/repos",
410 ("https://svn.apache.org/repos",
410 "url[https://svn.apache.org/repos]",
411 "url[https://svn.apache.org/repos]",
411 "https://svn.apache.org/repos"),
412 "https://svn.apache.org/repos"),
412 ("http://svn.apache.org/repos",
413 ("http://svn.apache.org/repos",
413 "url[http://svn.apache.org/repos]",
414 "url[http://svn.apache.org/repos]",
414 "http://svn.apache.org/repos"),
415 "http://svn.apache.org/repos"),
415 ("from rev a also rev http://google.com",
416 ("from rev a also rev http://google.com",
416 "from rev a also rev url[http://google.com]",
417 "from rev a also rev url[http://google.com]",
417 "http://google.com"),
418 "http://google.com"),
418 ("""Multi line
419 ("""Multi line
419 https://foo.bar.com
420 https://foo.bar.com
420 some text lalala""",
421 some text lalala""",
421 """Multi line
422 """Multi line
422 url[https://foo.bar.com]
423 url[https://foo.bar.com]
423 some text lalala""",
424 some text lalala""",
424 "https://foo.bar.com")
425 "https://foo.bar.com")
425 ], ids=no_newline_id_generator)
426 ], ids=no_newline_id_generator)
426 def test_urlify_test(sample, expected, url_):
427 def test_urlify_test(sample, expected, url_):
427 from rhodecode.lib.helpers import urlify_text
428 from rhodecode.lib.helpers import urlify_text
428 expected = _quick_url(expected, tmpl="""<a href="%s">%s</a>""", url_=url_)
429 expected = _quick_url(expected, tmpl="""<a href="%s">%s</a>""", url_=url_)
429 assert urlify_text(sample) == expected
430 assert urlify_text(sample) == expected
430
431
431
432
432 @pytest.mark.parametrize("test, expected", [
433 @pytest.mark.parametrize("test, expected", [
433 ("", None),
434 ("", None),
434 ("/_2", '2'),
435 ("/_2", '2'),
435 ("_2", '2'),
436 ("_2", '2'),
436 ("/_2/", '2'),
437 ("/_2/", '2'),
437 ("_2/", '2'),
438 ("_2/", '2'),
438
439
439 ("/_21", '21'),
440 ("/_21", '21'),
440 ("_21", '21'),
441 ("_21", '21'),
441 ("/_21/", '21'),
442 ("/_21/", '21'),
442 ("_21/", '21'),
443 ("_21/", '21'),
443
444
444 ("/_21/foobar", '21'),
445 ("/_21/foobar", '21'),
445 ("_21/121", '21'),
446 ("_21/121", '21'),
446 ("/_21/_12", '21'),
447 ("/_21/_12", '21'),
447 ("_21/rc/foo", '21'),
448 ("_21/rc/foo", '21'),
448
449
449 ])
450 ])
450 def test_get_repo_by_id(test, expected):
451 def test_get_repo_by_id(test, expected):
451 from rhodecode.model.repo import RepoModel
452 from rhodecode.model.repo import RepoModel
452 _test = RepoModel()._extract_id_from_repo_name(test)
453 _test = RepoModel()._extract_id_from_repo_name(test)
453 assert _test == expected
454 assert _test == expected
454
455
455
456
456 @pytest.mark.parametrize("test_repo_name, repo_type", [
457 @pytest.mark.parametrize("test_repo_name, repo_type", [
457 ("test_repo_1", None),
458 ("test_repo_1", None),
458 ("repo_group/foobar", None),
459 ("repo_group/foobar", None),
459 ("test_non_asci_Δ…Δ‡Δ™", None),
460 ("test_non_asci_Δ…Δ‡Δ™", None),
460 (u"test_non_asci_unicode_Δ…Δ‡Δ™", None),
461 (u"test_non_asci_unicode_Δ…Δ‡Δ™", None),
461 ])
462 ])
462 def test_invalidation_context(pylonsapp, test_repo_name, repo_type):
463 def test_invalidation_context(pylonsapp, test_repo_name, repo_type):
463 from beaker.cache import cache_region
464 from beaker.cache import cache_region
464 from rhodecode.lib import caches
465 from rhodecode.lib import caches
465 from rhodecode.model.db import CacheKey
466 from rhodecode.model.db import CacheKey
466
467
467 @cache_region('long_term')
468 @cache_region('long_term')
468 def _dummy_func(cache_key):
469 def _dummy_func(cache_key):
469 return 'result'
470 return 'result'
470
471
471 invalidator_context = CacheKey.repo_context_cache(
472 invalidator_context = CacheKey.repo_context_cache(
472 _dummy_func, test_repo_name, 'repo')
473 _dummy_func, test_repo_name, 'repo')
473
474
474 with invalidator_context as context:
475 with invalidator_context as context:
475 invalidated = context.invalidate()
476 invalidated = context.invalidate()
476 result = context.compute()
477 result = context.compute()
477
478
478 assert invalidated == True
479 assert invalidated == True
479 assert 'result' == result
480 assert 'result' == result
480 assert isinstance(context, caches.FreshRegionCache)
481 assert isinstance(context, caches.FreshRegionCache)
481
482
482 assert 'InvalidationContext' in repr(invalidator_context)
483 assert 'InvalidationContext' in repr(invalidator_context)
483
484
484 with invalidator_context as context:
485 with invalidator_context as context:
485 context.invalidate()
486 context.invalidate()
486 result = context.compute()
487 result = context.compute()
487
488
488 assert 'result' == result
489 assert 'result' == result
489 assert isinstance(context, caches.ActiveRegionCache)
490 assert isinstance(context, caches.ActiveRegionCache)
490
491
491
492
492 def test_invalidation_context_exception_in_compute(pylonsapp):
493 def test_invalidation_context_exception_in_compute(pylonsapp):
493 from rhodecode.model.db import CacheKey
494 from rhodecode.model.db import CacheKey
494 from beaker.cache import cache_region
495 from beaker.cache import cache_region
495
496
496 @cache_region('long_term')
497 @cache_region('long_term')
497 def _dummy_func(cache_key):
498 def _dummy_func(cache_key):
498 # this causes error since it doesn't get any params
499 # this causes error since it doesn't get any params
499 raise Exception('ups')
500 raise Exception('ups')
500
501
501 invalidator_context = CacheKey.repo_context_cache(
502 invalidator_context = CacheKey.repo_context_cache(
502 _dummy_func, 'test_repo_2', 'repo')
503 _dummy_func, 'test_repo_2', 'repo')
503
504
504 with pytest.raises(Exception):
505 with pytest.raises(Exception):
505 with invalidator_context as context:
506 with invalidator_context as context:
506 context.invalidate()
507 context.invalidate()
507 context.compute()
508 context.compute()
508
509
509
510
510 @pytest.mark.parametrize('execution_number', range(5))
511 @pytest.mark.parametrize('execution_number', range(5))
511 def test_cache_invalidation_race_condition(execution_number, pylonsapp):
512 def test_cache_invalidation_race_condition(execution_number, pylonsapp):
512 import time
513 import time
513 from beaker.cache import cache_region
514 from beaker.cache import cache_region
514 from rhodecode.model.db import CacheKey
515 from rhodecode.model.db import CacheKey
515
516
516 if CacheKey.metadata.bind.url.get_backend_name() == "mysql":
517 if CacheKey.metadata.bind.url.get_backend_name() == "mysql":
517 reason = (
518 reason = (
518 'Fails on MariaDB due to some locking issues. Investigation'
519 'Fails on MariaDB due to some locking issues. Investigation'
519 ' needed')
520 ' needed')
520 pytest.xfail(reason=reason)
521 pytest.xfail(reason=reason)
521
522
522 @run_test_concurrently(25)
523 @run_test_concurrently(25)
523 def test_create_and_delete_cache_keys():
524 def test_create_and_delete_cache_keys():
524 time.sleep(0.2)
525 time.sleep(0.2)
525
526
526 @cache_region('long_term')
527 @cache_region('long_term')
527 def _dummy_func(cache_key):
528 def _dummy_func(cache_key):
528 return 'result'
529 return 'result'
529
530
530 invalidator_context = CacheKey.repo_context_cache(
531 invalidator_context = CacheKey.repo_context_cache(
531 _dummy_func, 'test_repo_1', 'repo')
532 _dummy_func, 'test_repo_1', 'repo')
532
533
533 with invalidator_context as context:
534 with invalidator_context as context:
534 context.invalidate()
535 context.invalidate()
535 context.compute()
536 context.compute()
536
537
537 CacheKey.set_invalidate('test_repo_1', delete=True)
538 CacheKey.set_invalidate('test_repo_1', delete=True)
538
539
539 test_create_and_delete_cache_keys()
540 test_create_and_delete_cache_keys()
General Comments 0
You need to be logged in to leave comments. Login now