|
|
|
|
|
|
|
|
# Copyright (C) 2016-2023 RhodeCode GmbH
|
|
|
#
|
|
|
# This program is free software: you can redistribute it and/or modify
|
|
|
# it under the terms of the GNU Affero General Public License, version 3
|
|
|
# (only), as published by the Free Software Foundation.
|
|
|
#
|
|
|
# This program is distributed in the hope that it will be useful,
|
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
# GNU General Public License for more details.
|
|
|
#
|
|
|
# You should have received a copy of the GNU Affero General Public License
|
|
|
# along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
|
#
|
|
|
# This program is dual-licensed. If you wish to learn more about the
|
|
|
# RhodeCode Enterprise Edition, including its added features, Support services,
|
|
|
# and proprietary license terms, please see https://rhodecode.com/licenses/
|
|
|
|
|
|
import pytest
|
|
|
from pygments.lexers import get_lexer_by_name
|
|
|
|
|
|
from rhodecode.tests import no_newline_id_generator
|
|
|
from rhodecode.lib.codeblocks import (
|
|
|
tokenize_string, split_token_stream, rollup_tokenstream,
|
|
|
render_tokenstream)
|
|
|
|
|
|
|
|
|
class TestTokenizeString(object):
|
|
|
|
|
|
python_code = '''
|
|
|
import this
|
|
|
|
|
|
var = 6
|
|
|
print("this")
|
|
|
|
|
|
'''
|
|
|
|
|
|
def test_tokenize_as_python(self):
|
|
|
lexer = get_lexer_by_name('python')
|
|
|
tokens = list(tokenize_string(self.python_code, lexer))
|
|
|
expected_tokens = [
|
|
|
('w', '\n'),
|
|
|
('', ' '),
|
|
|
('kn', 'import'),
|
|
|
('', ' '),
|
|
|
('nn', 'this'),
|
|
|
('w', '\n'),
|
|
|
('w', '\n'),
|
|
|
('', ' '),
|
|
|
('n', 'var'),
|
|
|
('', ' '),
|
|
|
('o', '='),
|
|
|
('', ' '),
|
|
|
('mi', '6'),
|
|
|
('w', '\n'),
|
|
|
('', ' '),
|
|
|
('nb', 'print'),
|
|
|
('p', '('),
|
|
|
('s2', '"'),
|
|
|
('s2', 'this'),
|
|
|
('s2', '"'),
|
|
|
('p', ')'),
|
|
|
('w', '\n'),
|
|
|
('w', '\n'),
|
|
|
('', ' ')
|
|
|
]
|
|
|
|
|
|
assert tokens == expected_tokens
|
|
|
|
|
|
def test_tokenize_as_text(self):
|
|
|
lexer = get_lexer_by_name('text')
|
|
|
tokens = list(tokenize_string(self.python_code, lexer))
|
|
|
|
|
|
assert tokens == [
|
|
|
('',
|
|
|
'\n import this\n\n var = 6\n print("this")\n\n ')
|
|
|
]
|
|
|
|
|
|
|
|
|
class TestSplitTokenStream(object):
|
|
|
|
|
|
def test_split_token_stream(self):
|
|
|
tokens = [('type1', 'some\ntext'), ('type2', 'more\n')]
|
|
|
content = [x + y for x, y in tokens]
|
|
|
lines = list(split_token_stream(tokens, content))
|
|
|
|
|
|
assert lines == [
|
|
|
[('type1', 'some')],
|
|
|
[('type1', 'text'), ('type2', 'more')],
|
|
|
[('type2', '')],
|
|
|
]
|
|
|
|
|
|
def test_split_token_stream_single(self):
|
|
|
tokens = [('type1', '\n')]
|
|
|
content = [x + y for x, y in tokens]
|
|
|
lines = list(split_token_stream(tokens, content))
|
|
|
assert lines == [
|
|
|
[('type1', '')],
|
|
|
[('type1', '')],
|
|
|
]
|
|
|
|
|
|
def test_split_token_stream_single_repeat(self):
|
|
|
tokens = [('type1', '\n\n\n')]
|
|
|
content = [x + y for x, y in tokens]
|
|
|
lines = list(split_token_stream(tokens, content))
|
|
|
assert lines == [
|
|
|
[('type1', '')],
|
|
|
[('type1', '')],
|
|
|
[('type1', '')],
|
|
|
[('type1', '')],
|
|
|
]
|
|
|
|
|
|
def test_split_token_stream_multiple_repeat(self):
|
|
|
tokens = [('type1', '\n\n'), ('type2', '\n\n')]
|
|
|
content = [x + y for x, y in tokens]
|
|
|
|
|
|
lines = list(split_token_stream(tokens, content))
|
|
|
assert lines == [
|
|
|
[('type1', '')],
|
|
|
[('type1', '')],
|
|
|
[('type1', ''), ('type2', '')],
|
|
|
[('type2', '')],
|
|
|
[('type2', '')],
|
|
|
]
|
|
|
|
|
|
def test_no_tokens_by_content(self):
|
|
|
tokens = []
|
|
|
content = '\ufeff'
|
|
|
lines = list(split_token_stream(tokens, content))
|
|
|
assert lines == [
|
|
|
[('', content)],
|
|
|
]
|
|
|
|
|
|
def test_no_tokens_by_valid_content(self):
|
|
|
from pygments.lexers.css import CssLexer
|
|
|
content = '\ufeff table.dataTable'
|
|
|
tokens = tokenize_string(content, CssLexer())
|
|
|
|
|
|
lines = list(split_token_stream(tokens, content))
|
|
|
assert lines == [
|
|
|
[('w', ' '),
|
|
|
('nt', 'table'),
|
|
|
('p', '.'),
|
|
|
('nc', 'dataTable')],
|
|
|
]
|
|
|
|
|
|
|
|
|
class TestRollupTokens(object):
|
|
|
|
|
|
@pytest.mark.parametrize('tokenstream,output', [
|
|
|
([],
|
|
|
[]),
|
|
|
([('A', 'hell'), ('A', 'o')], [
|
|
|
('A', [
|
|
|
('', 'hello')]),
|
|
|
]),
|
|
|
([('A', 'hell'), ('B', 'o')], [
|
|
|
('A', [
|
|
|
('', 'hell')]),
|
|
|
('B', [
|
|
|
('', 'o')]),
|
|
|
]),
|
|
|
([('A', 'hel'), ('A', 'lo'), ('B', ' '), ('A', 'there')], [
|
|
|
('A', [
|
|
|
('', 'hello')]),
|
|
|
('B', [
|
|
|
('', ' ')]),
|
|
|
('A', [
|
|
|
('', 'there')]),
|
|
|
]),
|
|
|
])
|
|
|
def test_rollup_tokenstream_without_ops(self, tokenstream, output):
|
|
|
assert list(rollup_tokenstream(tokenstream)) == output
|
|
|
|
|
|
@pytest.mark.parametrize('tokenstream,output', [
|
|
|
([],
|
|
|
[]),
|
|
|
([('A', '', 'hell'), ('A', '', 'o')], [
|
|
|
('A', [
|
|
|
('', 'hello')]),
|
|
|
]),
|
|
|
([('A', '', 'hell'), ('B', '', 'o')], [
|
|
|
('A', [
|
|
|
('', 'hell')]),
|
|
|
('B', [
|
|
|
('', 'o')]),
|
|
|
]),
|
|
|
([('A', '', 'h'), ('B', '', 'e'), ('C', '', 'y')], [
|
|
|
('A', [
|
|
|
('', 'h')]),
|
|
|
('B', [
|
|
|
('', 'e')]),
|
|
|
('C', [
|
|
|
('', 'y')]),
|
|
|
]),
|
|
|
([('A', '', 'h'), ('A', '', 'e'), ('C', '', 'y')], [
|
|
|
('A', [
|
|
|
('', 'he')]),
|
|
|
('C', [
|
|
|
('', 'y')]),
|
|
|
]),
|
|
|
([('A', 'ins', 'h'), ('A', 'ins', 'e')], [
|
|
|
('A', [
|
|
|
('ins', 'he')
|
|
|
]),
|
|
|
]),
|
|
|
([('A', 'ins', 'h'), ('A', 'del', 'e')], [
|
|
|
('A', [
|
|
|
('ins', 'h'),
|
|
|
('del', 'e')
|
|
|
]),
|
|
|
]),
|
|
|
([('A', 'ins', 'h'), ('B', 'del', 'e'), ('B', 'del', 'y')], [
|
|
|
('A', [
|
|
|
('ins', 'h'),
|
|
|
]),
|
|
|
('B', [
|
|
|
('del', 'ey'),
|
|
|
]),
|
|
|
]),
|
|
|
([('A', 'ins', 'h'), ('A', 'del', 'e'), ('B', 'del', 'y')], [
|
|
|
('A', [
|
|
|
('ins', 'h'),
|
|
|
('del', 'e'),
|
|
|
]),
|
|
|
('B', [
|
|
|
('del', 'y'),
|
|
|
]),
|
|
|
]),
|
|
|
([('A', '', 'some'), ('A', 'ins', 'new'), ('A', '', 'name')], [
|
|
|
('A', [
|
|
|
('', 'some'),
|
|
|
('ins', 'new'),
|
|
|
('', 'name'),
|
|
|
]),
|
|
|
]),
|
|
|
])
|
|
|
def test_rollup_tokenstream_with_ops(self, tokenstream, output):
|
|
|
assert list(rollup_tokenstream(tokenstream)) == output
|
|
|
|
|
|
|
|
|
class TestRenderTokenStream(object):
|
|
|
|
|
|
@pytest.mark.parametrize('tokenstream,output', [
|
|
|
(
|
|
|
[],
|
|
|
'',
|
|
|
),
|
|
|
(
|
|
|
[('', '', '')],
|
|
|
'<span></span>',
|
|
|
),
|
|
|
(
|
|
|
[('', '', 'text')],
|
|
|
'<span>text</span>',
|
|
|
),
|
|
|
(
|
|
|
[('A', '', '')],
|
|
|
'<span class="A"></span>',
|
|
|
),
|
|
|
(
|
|
|
[('A', '', 'hello')],
|
|
|
'<span class="A">hello</span>',
|
|
|
),
|
|
|
(
|
|
|
[('A', '', 'hel'), ('A', '', 'lo')],
|
|
|
'<span class="A">hello</span>',
|
|
|
),
|
|
|
(
|
|
|
[('A', '', 'two\n'), ('A', '', 'lines')],
|
|
|
'<span class="A">two\nlines</span>',
|
|
|
),
|
|
|
(
|
|
|
[('A', '', '\nthree\n'), ('A', '', 'lines')],
|
|
|
'<span class="A">\nthree\nlines</span>',
|
|
|
),
|
|
|
(
|
|
|
[('', '', '\n'), ('A', '', 'line')],
|
|
|
'<span>\n</span><span class="A">line</span>',
|
|
|
),
|
|
|
(
|
|
|
[('', 'ins', '\n'), ('A', '', 'line')],
|
|
|
'<span><ins>\n</ins></span><span class="A">line</span>',
|
|
|
),
|
|
|
(
|
|
|
[('A', '', 'hel'), ('A', 'ins', 'lo')],
|
|
|
'<span class="A">hel<ins>lo</ins></span>',
|
|
|
),
|
|
|
(
|
|
|
[('A', '', 'hel'), ('A', 'ins', 'l'), ('A', 'ins', 'o')],
|
|
|
'<span class="A">hel<ins>lo</ins></span>',
|
|
|
),
|
|
|
(
|
|
|
[('A', '', 'hel'), ('A', 'ins', 'l'), ('A', 'del', 'o')],
|
|
|
'<span class="A">hel<ins>l</ins><del>o</del></span>',
|
|
|
),
|
|
|
(
|
|
|
[('A', '', 'hel'), ('B', '', 'lo')],
|
|
|
'<span class="A">hel</span><span class="B">lo</span>',
|
|
|
),
|
|
|
(
|
|
|
[('A', '', 'hel'), ('B', 'ins', 'lo')],
|
|
|
'<span class="A">hel</span><span class="B"><ins>lo</ins></span>',
|
|
|
),
|
|
|
], ids=no_newline_id_generator)
|
|
|
def test_render_tokenstream_with_ops(self, tokenstream, output):
|
|
|
html = render_tokenstream(tokenstream)
|
|
|
assert html == output
|
|
|
|
|
|
@pytest.mark.parametrize('tokenstream,output', [
|
|
|
(
|
|
|
[('A', 'hel'), ('A', 'lo')],
|
|
|
'<span class="A">hello</span>',
|
|
|
),
|
|
|
(
|
|
|
[('A', 'hel'), ('A', 'l'), ('A', 'o')],
|
|
|
'<span class="A">hello</span>',
|
|
|
),
|
|
|
(
|
|
|
[('A', 'hel'), ('A', 'l'), ('A', 'o')],
|
|
|
'<span class="A">hello</span>',
|
|
|
),
|
|
|
(
|
|
|
[('A', 'hel'), ('B', 'lo')],
|
|
|
'<span class="A">hel</span><span class="B">lo</span>',
|
|
|
),
|
|
|
(
|
|
|
[('A', 'hel'), ('B', 'lo')],
|
|
|
'<span class="A">hel</span><span class="B">lo</span>',
|
|
|
),
|
|
|
])
|
|
|
def test_render_tokenstream_without_ops(self, tokenstream, output):
|
|
|
html = render_tokenstream(tokenstream)
|
|
|
assert html == output
|
|
|
|