Show More
@@ -0,0 +1,111 b'' | |||
|
1 | """Test lexers module""" | |
|
2 | #----------------------------------------------------------------------------- | |
|
3 | # Copyright (C) 2014 The IPython Development Team | |
|
4 | # | |
|
5 | # Distributed under the terms of the BSD License. The full license is in | |
|
6 | # the file COPYING, distributed as part of this software. | |
|
7 | #----------------------------------------------------------------------------- | |
|
8 | ||
|
9 | #----------------------------------------------------------------------------- | |
|
10 | # Imports | |
|
11 | #----------------------------------------------------------------------------- | |
|
12 | from pygments.token import Token | |
|
13 | ||
|
14 | from IPython.nbconvert.tests.base import TestsBase | |
|
15 | from .. import lexers | |
|
16 | ||
|
17 | ||
|
18 | #----------------------------------------------------------------------------- | |
|
19 | # Classes and functions | |
|
20 | #----------------------------------------------------------------------------- | |
|
21 | class TestLexers(TestsBase): | |
|
22 | """Collection of lexers tests""" | |
|
23 | def setUp(self): | |
|
24 | self.lexer = lexers.IPythonLexer() | |
|
25 | ||
|
26 | def testIPythonLexer(self): | |
|
27 | fragment = '!echo $HOME\n' | |
|
28 | tokens = [ | |
|
29 | (Token.Operator, '!'), | |
|
30 | (Token.Name.Builtin, 'echo'), | |
|
31 | (Token.Text, ' '), | |
|
32 | (Token.Name.Variable, '$HOME'), | |
|
33 | (Token.Text, '\n'), | |
|
34 | ] | |
|
35 | self.assertEqual(tokens, list(self.lexer.get_tokens(fragment))) | |
|
36 | ||
|
37 | fragment_2 = '!' + fragment | |
|
38 | tokens_2 = [ | |
|
39 | (Token.Operator, '!!'), | |
|
40 | ] + tokens[1:] | |
|
41 | self.assertEqual(tokens_2, list(self.lexer.get_tokens(fragment_2))) | |
|
42 | ||
|
43 | fragment_2 = '\t %%!\n' + fragment[1:] | |
|
44 | tokens_2 = [ | |
|
45 | (Token.Text, '\t '), | |
|
46 | (Token.Operator, '%%!'), | |
|
47 | (Token.Text, '\n'), | |
|
48 | ] + tokens[1:] | |
|
49 | self.assertEqual(tokens_2, list(self.lexer.get_tokens(fragment_2))) | |
|
50 | ||
|
51 | fragment_2 = 'x = ' + fragment | |
|
52 | tokens_2 = [ | |
|
53 | (Token.Name, 'x'), | |
|
54 | (Token.Text, ' '), | |
|
55 | (Token.Operator, '='), | |
|
56 | (Token.Text, ' '), | |
|
57 | ] + tokens | |
|
58 | self.assertEqual(tokens_2, list(self.lexer.get_tokens(fragment_2))) | |
|
59 | ||
|
60 | fragment_2 = 'x, = ' + fragment | |
|
61 | tokens_2 = [ | |
|
62 | (Token.Name, 'x'), | |
|
63 | (Token.Punctuation, ','), | |
|
64 | (Token.Text, ' '), | |
|
65 | (Token.Operator, '='), | |
|
66 | (Token.Text, ' '), | |
|
67 | ] + tokens | |
|
68 | self.assertEqual(tokens_2, list(self.lexer.get_tokens(fragment_2))) | |
|
69 | ||
|
70 | fragment_2 = 'x, = %sx ' + fragment[1:] | |
|
71 | tokens_2 = [ | |
|
72 | (Token.Name, 'x'), | |
|
73 | (Token.Punctuation, ','), | |
|
74 | (Token.Text, ' '), | |
|
75 | (Token.Operator, '='), | |
|
76 | (Token.Text, ' '), | |
|
77 | (Token.Operator, '%'), | |
|
78 | (Token.Keyword, 'sx'), | |
|
79 | (Token.Text, ' '), | |
|
80 | ] + tokens[1:] | |
|
81 | self.assertEqual(tokens_2, list(self.lexer.get_tokens(fragment_2))) | |
|
82 | ||
|
83 | fragment_2 = 'f = %R function () {}\n' | |
|
84 | tokens_2 = [ | |
|
85 | (Token.Name, 'f'), | |
|
86 | (Token.Text, ' '), | |
|
87 | (Token.Operator, '='), | |
|
88 | (Token.Text, ' '), | |
|
89 | (Token.Operator, '%'), | |
|
90 | (Token.Keyword, 'R'), | |
|
91 | (Token.Text, ' function () {}\n'), | |
|
92 | ] | |
|
93 | self.assertEqual(tokens_2, list(self.lexer.get_tokens(fragment_2))) | |
|
94 | ||
|
95 | fragment_2 = '\t%%xyz\n$foo\n' | |
|
96 | tokens_2 = [ | |
|
97 | (Token.Text, '\t'), | |
|
98 | (Token.Operator, '%%'), | |
|
99 | (Token.Keyword, 'xyz'), | |
|
100 | (Token.Text, '\n$foo\n'), | |
|
101 | ] | |
|
102 | self.assertEqual(tokens_2, list(self.lexer.get_tokens(fragment_2))) | |
|
103 | ||
|
104 | fragment_2 = '%system?\n' | |
|
105 | tokens_2 = [ | |
|
106 | (Token.Operator, '%'), | |
|
107 | (Token.Keyword, 'system'), | |
|
108 | (Token.Operator, '?'), | |
|
109 | (Token.Text, '\n'), | |
|
110 | ] | |
|
111 | self.assertEqual(tokens_2, list(self.lexer.get_tokens(fragment_2))) |
@@ -49,10 +49,15 b' from IPython.testing.skipdoctest import skip_doctest' | |||
|
49 | 49 | line_re = re.compile('.*?\n') |
|
50 | 50 | |
|
51 | 51 | ipython_tokens = [ |
|
52 |
(r |
|
|
52 | (r"(?s)(\s*)(%%)(\w+)(.*)", bygroups(Text, Operator, Keyword, Text)), | |
|
53 | (r'(?s)(^\s*)(%%!)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(BashLexer))), | |
|
54 | (r"(%%?)(\w+)(\?\??)$", bygroups(Operator, Keyword, Operator)), | |
|
55 | (r"\b(\?\??)(\s*)$", bygroups(Operator, Text)), | |
|
56 | (r'(%)(sx|sc|system)(.*)(\n)', bygroups(Operator, Keyword, | |
|
53 | 57 | using(BashLexer), Text)), |
|
54 |
(r'(\ |
|
|
55 | (r'^(!)(.+)(\n)', bygroups(Operator, using(BashLexer), Text)), | |
|
58 | (r'(%)(\w+)(.*\n)', bygroups(Operator, Keyword, Text)), | |
|
59 | (r'^(!!)(.+)(\n)', bygroups(Operator, using(BashLexer), Text)), | |
|
60 | (r'(!)(.+)(\n)', bygroups(Operator, using(BashLexer), Text)), | |
|
56 | 61 | ] |
|
57 | 62 | |
|
58 | 63 | def build_ipy_lexer(python3): |
General Comments 0
You need to be logged in to leave comments.
Login now