# Copyright (C) 2016-2020 RhodeCode GmbH # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License, version 3 # (only), as published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see . # # This program is dual-licensed. If you wish to learn more about the # RhodeCode Enterprise Edition, including its added features, Support services, # and proprietary license terms, please see https://rhodecode.com/licenses/ import pytest from pygments.lexers import get_lexer_by_name from rhodecode.tests import no_newline_id_generator from rhodecode.lib.codeblocks import ( tokenize_string, split_token_stream, rollup_tokenstream, render_tokenstream) class TestTokenizeString(object): python_code = ''' import this var = 6 print("this") ''' def test_tokenize_as_python(self): lexer = get_lexer_by_name('python') tokens = list(tokenize_string(self.python_code, lexer)) expected_tokens = [ ('w', '\n'), ('', ' '), ('kn', 'import'), ('', ' '), ('nn', 'this'), ('w', '\n'), ('w', '\n'), ('', ' '), ('n', 'var'), ('', ' '), ('o', '='), ('', ' '), ('mi', '6'), ('w', '\n'), ('', ' '), ('nb', 'print'), ('p', '('), ('s2', '"'), ('s2', 'this'), ('s2', '"'), ('p', ')'), ('w', '\n'), ('w', '\n'), ('', ' ') ] assert tokens == expected_tokens def test_tokenize_as_text(self): lexer = get_lexer_by_name('text') tokens = list(tokenize_string(self.python_code, lexer)) assert tokens == [ ('', '\n import this\n\n var = 6\n print("this")\n\n ') ] class TestSplitTokenStream(object): def test_split_token_stream(self): tokens = [('type1', 'some\ntext'), ('type2', 'more\n')] content = [x + y for x, y in tokens] lines = list(split_token_stream(tokens, content)) assert lines == [ [('type1', 'some')], [('type1', 'text'), ('type2', 'more')], [('type2', '')], ] def test_split_token_stream_single(self): tokens = [('type1', '\n')] content = [x + y for x, y in tokens] lines = list(split_token_stream(tokens, content)) assert lines == [ [('type1', '')], [('type1', '')], ] def test_split_token_stream_single_repeat(self): tokens = [('type1', '\n\n\n')] content = [x + y for x, y in tokens] lines = list(split_token_stream(tokens, content)) assert lines == [ [('type1', '')], [('type1', '')], [('type1', '')], [('type1', '')], ] def test_split_token_stream_multiple_repeat(self): tokens = [('type1', '\n\n'), ('type2', '\n\n')] content = [x + y for x, y in tokens] lines = list(split_token_stream(tokens, content)) assert lines == [ [('type1', '')], [('type1', '')], [('type1', ''), ('type2', '')], [('type2', '')], [('type2', '')], ] def test_no_tokens_by_content(self): tokens = [] content = '\ufeff' lines = list(split_token_stream(tokens, content)) assert lines == [ [('', content)], ] def test_no_tokens_by_valid_content(self): from pygments.lexers.css import CssLexer content = '\ufeff table.dataTable' tokens = tokenize_string(content, CssLexer()) lines = list(split_token_stream(tokens, content)) assert lines == [ [('w', ' '), ('nt', 'table'), ('p', '.'), ('nc', 'dataTable')], ] class TestRollupTokens(object): @pytest.mark.parametrize('tokenstream,output', [ ([], []), ([('A', 'hell'), ('A', 'o')], [ ('A', [ ('', 'hello')]), ]), ([('A', 'hell'), ('B', 'o')], [ ('A', [ ('', 'hell')]), ('B', [ ('', 'o')]), ]), ([('A', 'hel'), ('A', 'lo'), ('B', ' '), ('A', 'there')], [ ('A', [ ('', 'hello')]), ('B', [ ('', ' ')]), ('A', [ ('', 'there')]), ]), ]) def test_rollup_tokenstream_without_ops(self, tokenstream, output): assert list(rollup_tokenstream(tokenstream)) == output @pytest.mark.parametrize('tokenstream,output', [ ([], []), ([('A', '', 'hell'), ('A', '', 'o')], [ ('A', [ ('', 'hello')]), ]), ([('A', '', 'hell'), ('B', '', 'o')], [ ('A', [ ('', 'hell')]), ('B', [ ('', 'o')]), ]), ([('A', '', 'h'), ('B', '', 'e'), ('C', '', 'y')], [ ('A', [ ('', 'h')]), ('B', [ ('', 'e')]), ('C', [ ('', 'y')]), ]), ([('A', '', 'h'), ('A', '', 'e'), ('C', '', 'y')], [ ('A', [ ('', 'he')]), ('C', [ ('', 'y')]), ]), ([('A', 'ins', 'h'), ('A', 'ins', 'e')], [ ('A', [ ('ins', 'he') ]), ]), ([('A', 'ins', 'h'), ('A', 'del', 'e')], [ ('A', [ ('ins', 'h'), ('del', 'e') ]), ]), ([('A', 'ins', 'h'), ('B', 'del', 'e'), ('B', 'del', 'y')], [ ('A', [ ('ins', 'h'), ]), ('B', [ ('del', 'ey'), ]), ]), ([('A', 'ins', 'h'), ('A', 'del', 'e'), ('B', 'del', 'y')], [ ('A', [ ('ins', 'h'), ('del', 'e'), ]), ('B', [ ('del', 'y'), ]), ]), ([('A', '', 'some'), ('A', 'ins', 'new'), ('A', '', 'name')], [ ('A', [ ('', 'some'), ('ins', 'new'), ('', 'name'), ]), ]), ]) def test_rollup_tokenstream_with_ops(self, tokenstream, output): assert list(rollup_tokenstream(tokenstream)) == output class TestRenderTokenStream(object): @pytest.mark.parametrize('tokenstream,output', [ ( [], '', ), ( [('', '', '')], '', ), ( [('', '', 'text')], 'text', ), ( [('A', '', '')], '', ), ( [('A', '', 'hello')], 'hello', ), ( [('A', '', 'hel'), ('A', '', 'lo')], 'hello', ), ( [('A', '', 'two\n'), ('A', '', 'lines')], 'two\nlines', ), ( [('A', '', '\nthree\n'), ('A', '', 'lines')], '\nthree\nlines', ), ( [('', '', '\n'), ('A', '', 'line')], '\nline', ), ( [('', 'ins', '\n'), ('A', '', 'line')], '\nline', ), ( [('A', '', 'hel'), ('A', 'ins', 'lo')], 'hello', ), ( [('A', '', 'hel'), ('A', 'ins', 'l'), ('A', 'ins', 'o')], 'hello', ), ( [('A', '', 'hel'), ('A', 'ins', 'l'), ('A', 'del', 'o')], 'hello', ), ( [('A', '', 'hel'), ('B', '', 'lo')], 'hello', ), ( [('A', '', 'hel'), ('B', 'ins', 'lo')], 'hello', ), ], ids=no_newline_id_generator) def test_render_tokenstream_with_ops(self, tokenstream, output): html = render_tokenstream(tokenstream) assert html == output @pytest.mark.parametrize('tokenstream,output', [ ( [('A', 'hel'), ('A', 'lo')], 'hello', ), ( [('A', 'hel'), ('A', 'l'), ('A', 'o')], 'hello', ), ( [('A', 'hel'), ('A', 'l'), ('A', 'o')], 'hello', ), ( [('A', 'hel'), ('B', 'lo')], 'hello', ), ( [('A', 'hel'), ('B', 'lo')], 'hello', ), ]) def test_render_tokenstream_without_ops(self, tokenstream, output): html = render_tokenstream(tokenstream) assert html == output