# -*- coding: utf-8 -*- # Copyright (C) 2016-2019 RhodeCode GmbH # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License, version 3 # (only), as published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see . # # This program is dual-licensed. If you wish to learn more about the # RhodeCode Enterprise Edition, including its added features, Support services, # and proprietary license terms, please see https://rhodecode.com/licenses/ import pytest from pygments.lexers import get_lexer_by_name from rhodecode.tests import no_newline_id_generator from rhodecode.lib.codeblocks import ( tokenize_string, split_token_stream, rollup_tokenstream, render_tokenstream) class TestTokenizeString(object): python_code = ''' import this var = 6 print("this") ''' def test_tokenize_as_python(self): lexer = get_lexer_by_name('python') tokens = list(tokenize_string(self.python_code, lexer)) assert tokens == [ ('', u'\n'), ('', u' '), ('kn', u'import'), ('', u' '), ('nn', u'this'), ('', u'\n'), ('', u'\n'), ('', u' '), ('n', u'var'), ('', u' '), ('o', u'='), ('', u' '), ('mi', u'6'), ('', u'\n'), ('', u' '), ('k', u'print'), ('p', u'('), ('s2', u'"'), ('s2', u'this'), ('s2', u'"'), ('p', u')'), ('', u'\n'), ('', u'\n'), ('', u' ') ] def test_tokenize_as_text(self): lexer = get_lexer_by_name('text') tokens = list(tokenize_string(self.python_code, lexer)) assert tokens == [ ('', u'\n import this\n\n var = 6\n print("this")\n\n ') ] class TestSplitTokenStream(object): def test_split_token_stream(self): tokens = [('type1', 'some\ntext'), ('type2', 'more\n')] content = [x + y for x, y in tokens] lines = list(split_token_stream(tokens, content)) assert lines == [ [('type1', u'some')], [('type1', u'text'), ('type2', u'more')], [('type2', u'')], ] def test_split_token_stream_single(self): tokens = [('type1', '\n')] content = [x + y for x, y in tokens] lines = list(split_token_stream(tokens, content)) assert lines == [ [('type1', '')], [('type1', '')], ] def test_split_token_stream_single_repeat(self): tokens = [('type1', '\n\n\n')] content = [x + y for x, y in tokens] lines = list(split_token_stream(tokens, content)) assert lines == [ [('type1', '')], [('type1', '')], [('type1', '')], [('type1', '')], ] def test_split_token_stream_multiple_repeat(self): tokens = [('type1', '\n\n'), ('type2', '\n\n')] content = [x + y for x, y in tokens] lines = list(split_token_stream(tokens, content)) assert lines == [ [('type1', '')], [('type1', '')], [('type1', ''), ('type2', '')], [('type2', '')], [('type2', '')], ] def test_no_tokens_by_content(self): tokens = [] content = u'\ufeff' lines = list(split_token_stream(tokens, content)) assert lines == [ [('', content)], ] def test_no_tokens_by_valid_content(self): from pygments.lexers.css import CssLexer content = u'\ufeff table.dataTable' tokens = tokenize_string(content, CssLexer()) lines = list(split_token_stream(tokens, content)) assert lines == [ [('', u' '), ('nt', u'table'), ('p', u'.'), ('nc', u'dataTable')], ] class TestRollupTokens(object): @pytest.mark.parametrize('tokenstream,output', [ ([], []), ([('A', 'hell'), ('A', 'o')], [ ('A', [ ('', 'hello')]), ]), ([('A', 'hell'), ('B', 'o')], [ ('A', [ ('', 'hell')]), ('B', [ ('', 'o')]), ]), ([('A', 'hel'), ('A', 'lo'), ('B', ' '), ('A', 'there')], [ ('A', [ ('', 'hello')]), ('B', [ ('', ' ')]), ('A', [ ('', 'there')]), ]), ]) def test_rollup_tokenstream_without_ops(self, tokenstream, output): assert list(rollup_tokenstream(tokenstream)) == output @pytest.mark.parametrize('tokenstream,output', [ ([], []), ([('A', '', 'hell'), ('A', '', 'o')], [ ('A', [ ('', 'hello')]), ]), ([('A', '', 'hell'), ('B', '', 'o')], [ ('A', [ ('', 'hell')]), ('B', [ ('', 'o')]), ]), ([('A', '', 'h'), ('B', '', 'e'), ('C', '', 'y')], [ ('A', [ ('', 'h')]), ('B', [ ('', 'e')]), ('C', [ ('', 'y')]), ]), ([('A', '', 'h'), ('A', '', 'e'), ('C', '', 'y')], [ ('A', [ ('', 'he')]), ('C', [ ('', 'y')]), ]), ([('A', 'ins', 'h'), ('A', 'ins', 'e')], [ ('A', [ ('ins', 'he') ]), ]), ([('A', 'ins', 'h'), ('A', 'del', 'e')], [ ('A', [ ('ins', 'h'), ('del', 'e') ]), ]), ([('A', 'ins', 'h'), ('B', 'del', 'e'), ('B', 'del', 'y')], [ ('A', [ ('ins', 'h'), ]), ('B', [ ('del', 'ey'), ]), ]), ([('A', 'ins', 'h'), ('A', 'del', 'e'), ('B', 'del', 'y')], [ ('A', [ ('ins', 'h'), ('del', 'e'), ]), ('B', [ ('del', 'y'), ]), ]), ([('A', '', 'some'), ('A', 'ins', 'new'), ('A', '', 'name')], [ ('A', [ ('', 'some'), ('ins', 'new'), ('', 'name'), ]), ]), ]) def test_rollup_tokenstream_with_ops(self, tokenstream, output): assert list(rollup_tokenstream(tokenstream)) == output class TestRenderTokenStream(object): @pytest.mark.parametrize('tokenstream,output', [ ( [], '', ), ( [('', '', u'')], '', ), ( [('', '', u'text')], 'text', ), ( [('A', '', u'')], '', ), ( [('A', '', u'hello')], 'hello', ), ( [('A', '', u'hel'), ('A', '', u'lo')], 'hello', ), ( [('A', '', u'two\n'), ('A', '', u'lines')], 'two\nlines', ), ( [('A', '', u'\nthree\n'), ('A', '', u'lines')], '\nthree\nlines', ), ( [('', '', u'\n'), ('A', '', u'line')], '\nline', ), ( [('', 'ins', u'\n'), ('A', '', u'line')], '\nline', ), ( [('A', '', u'hel'), ('A', 'ins', u'lo')], 'hello', ), ( [('A', '', u'hel'), ('A', 'ins', u'l'), ('A', 'ins', u'o')], 'hello', ), ( [('A', '', u'hel'), ('A', 'ins', u'l'), ('A', 'del', u'o')], 'hello', ), ( [('A', '', u'hel'), ('B', '', u'lo')], 'hello', ), ( [('A', '', u'hel'), ('B', 'ins', u'lo')], 'hello', ), ], ids=no_newline_id_generator) def test_render_tokenstream_with_ops(self, tokenstream, output): html = render_tokenstream(tokenstream) assert html == output @pytest.mark.parametrize('tokenstream,output', [ ( [('A', u'hel'), ('A', u'lo')], 'hello', ), ( [('A', u'hel'), ('A', u'l'), ('A', u'o')], 'hello', ), ( [('A', u'hel'), ('A', u'l'), ('A', u'o')], 'hello', ), ( [('A', u'hel'), ('B', u'lo')], 'hello', ), ( [('A', u'hel'), ('B', u'lo')], 'hello', ), ]) def test_render_tokenstream_without_ops(self, tokenstream, output): html = render_tokenstream(tokenstream) assert html == output