##// END OF EJS Templates
Fix IPythonConsoleLexer doctest
Nikita Kniazev -
Show More
@@ -1,532 +1,526 b''
1 1 # -*- coding: utf-8 -*-
2 2 """
3 3 Defines a variety of Pygments lexers for highlighting IPython code.
4 4
5 5 This includes:
6 6
7 7 IPythonLexer, IPython3Lexer
8 8 Lexers for pure IPython (python + magic/shell commands)
9 9
10 10 IPythonPartialTracebackLexer, IPythonTracebackLexer
11 11 Supports 2.x and 3.x via keyword `python3`. The partial traceback
12 12 lexer reads everything but the Python code appearing in a traceback.
13 13 The full lexer combines the partial lexer with an IPython lexer.
14 14
15 15 IPythonConsoleLexer
16 16 A lexer for IPython console sessions, with support for tracebacks.
17 17
18 18 IPyLexer
19 19 A friendly lexer which examines the first line of text and from it,
20 20 decides whether to use an IPython lexer or an IPython console lexer.
21 21 This is probably the only lexer that needs to be explicitly added
22 22 to Pygments.
23 23
24 24 """
25 25 #-----------------------------------------------------------------------------
26 26 # Copyright (c) 2013, the IPython Development Team.
27 27 #
28 28 # Distributed under the terms of the Modified BSD License.
29 29 #
30 30 # The full license is in the file COPYING.txt, distributed with this software.
31 31 #-----------------------------------------------------------------------------
32 32
33 33 # Standard library
34 34 import re
35 35
36 36 # Third party
37 37 from pygments.lexers import (
38 38 BashLexer, HtmlLexer, JavascriptLexer, RubyLexer, PerlLexer, PythonLexer,
39 39 Python3Lexer, TexLexer)
40 40 from pygments.lexer import (
41 41 Lexer, DelegatingLexer, RegexLexer, do_insertions, bygroups, using,
42 42 )
43 43 from pygments.token import (
44 44 Generic, Keyword, Literal, Name, Operator, Other, Text, Error,
45 45 )
46 46 from pygments.util import get_bool_opt
47 47
48 48 # Local
49 49
50 50 line_re = re.compile('.*?\n')
51 51
52 52 __all__ = ['build_ipy_lexer', 'IPython3Lexer', 'IPythonLexer',
53 53 'IPythonPartialTracebackLexer', 'IPythonTracebackLexer',
54 54 'IPythonConsoleLexer', 'IPyLexer']
55 55
56 56
57 57 def build_ipy_lexer(python3):
58 58 """Builds IPython lexers depending on the value of `python3`.
59 59
60 60 The lexer inherits from an appropriate Python lexer and then adds
61 61 information about IPython specific keywords (i.e. magic commands,
62 62 shell commands, etc.)
63 63
64 64 Parameters
65 65 ----------
66 66 python3 : bool
67 67 If `True`, then build an IPython lexer from a Python 3 lexer.
68 68
69 69 """
70 70 # It would be nice to have a single IPython lexer class which takes
71 71 # a boolean `python3`. But since there are two Python lexer classes,
72 72 # we will also have two IPython lexer classes.
73 73 if python3:
74 74 PyLexer = Python3Lexer
75 75 name = 'IPython3'
76 76 aliases = ['ipython3']
77 77 doc = """IPython3 Lexer"""
78 78 else:
79 79 PyLexer = PythonLexer
80 80 name = 'IPython'
81 81 aliases = ['ipython2', 'ipython']
82 82 doc = """IPython Lexer"""
83 83
84 84 ipython_tokens = [
85 85 (r'(?s)(\s*)(%%capture)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
86 86 (r'(?s)(\s*)(%%debug)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
87 87 (r'(?is)(\s*)(%%html)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(HtmlLexer))),
88 88 (r'(?s)(\s*)(%%javascript)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(JavascriptLexer))),
89 89 (r'(?s)(\s*)(%%js)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(JavascriptLexer))),
90 90 (r'(?s)(\s*)(%%latex)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(TexLexer))),
91 91 (r'(?s)(\s*)(%%perl)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PerlLexer))),
92 92 (r'(?s)(\s*)(%%prun)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
93 93 (r'(?s)(\s*)(%%pypy)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
94 94 (r'(?s)(\s*)(%%python)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
95 95 (r'(?s)(\s*)(%%python2)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PythonLexer))),
96 96 (r'(?s)(\s*)(%%python3)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(Python3Lexer))),
97 97 (r'(?s)(\s*)(%%ruby)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(RubyLexer))),
98 98 (r'(?s)(\s*)(%%time)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
99 99 (r'(?s)(\s*)(%%timeit)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
100 100 (r'(?s)(\s*)(%%writefile)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
101 101 (r'(?s)(\s*)(%%file)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
102 102 (r"(?s)(\s*)(%%)(\w+)(.*)", bygroups(Text, Operator, Keyword, Text)),
103 103 (r'(?s)(^\s*)(%%!)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(BashLexer))),
104 104 (r"(%%?)(\w+)(\?\??)$", bygroups(Operator, Keyword, Operator)),
105 105 (r"\b(\?\??)(\s*)$", bygroups(Operator, Text)),
106 106 (r'(%)(sx|sc|system)(.*)(\n)', bygroups(Operator, Keyword,
107 107 using(BashLexer), Text)),
108 108 (r'(%)(\w+)(.*\n)', bygroups(Operator, Keyword, Text)),
109 109 (r'^(!!)(.+)(\n)', bygroups(Operator, using(BashLexer), Text)),
110 110 (r'(!)(?!=)(.+)(\n)', bygroups(Operator, using(BashLexer), Text)),
111 111 (r'^(\s*)(\?\??)(\s*%{0,2}[\w\.\*]*)', bygroups(Text, Operator, Text)),
112 112 (r'(\s*%{0,2}[\w\.\*]*)(\?\??)(\s*)$', bygroups(Text, Operator, Text)),
113 113 ]
114 114
115 115 tokens = PyLexer.tokens.copy()
116 116 tokens['root'] = ipython_tokens + tokens['root']
117 117
118 118 attrs = {'name': name, 'aliases': aliases, 'filenames': [],
119 119 '__doc__': doc, 'tokens': tokens}
120 120
121 121 return type(name, (PyLexer,), attrs)
122 122
123 123
124 124 IPython3Lexer = build_ipy_lexer(python3=True)
125 125 IPythonLexer = build_ipy_lexer(python3=False)
126 126
127 127
128 128 class IPythonPartialTracebackLexer(RegexLexer):
129 129 """
130 130 Partial lexer for IPython tracebacks.
131 131
132 132 Handles all the non-python output.
133 133
134 134 """
135 135 name = 'IPython Partial Traceback'
136 136
137 137 tokens = {
138 138 'root': [
139 139 # Tracebacks for syntax errors have a different style.
140 140 # For both types of tracebacks, we mark the first line with
141 141 # Generic.Traceback. For syntax errors, we mark the filename
142 142 # as we mark the filenames for non-syntax tracebacks.
143 143 #
144 144 # These two regexps define how IPythonConsoleLexer finds a
145 145 # traceback.
146 146 #
147 147 ## Non-syntax traceback
148 148 (r'^(\^C)?(-+\n)', bygroups(Error, Generic.Traceback)),
149 149 ## Syntax traceback
150 150 (r'^( File)(.*)(, line )(\d+\n)',
151 151 bygroups(Generic.Traceback, Name.Namespace,
152 152 Generic.Traceback, Literal.Number.Integer)),
153 153
154 154 # (Exception Identifier)(Whitespace)(Traceback Message)
155 155 (r'(?u)(^[^\d\W]\w*)(\s*)(Traceback.*?\n)',
156 156 bygroups(Name.Exception, Generic.Whitespace, Text)),
157 157 # (Module/Filename)(Text)(Callee)(Function Signature)
158 158 # Better options for callee and function signature?
159 159 (r'(.*)( in )(.*)(\(.*\)\n)',
160 160 bygroups(Name.Namespace, Text, Name.Entity, Name.Tag)),
161 161 # Regular line: (Whitespace)(Line Number)(Python Code)
162 162 (r'(\s*?)(\d+)(.*?\n)',
163 163 bygroups(Generic.Whitespace, Literal.Number.Integer, Other)),
164 164 # Emphasized line: (Arrow)(Line Number)(Python Code)
165 165 # Using Exception token so arrow color matches the Exception.
166 166 (r'(-*>?\s?)(\d+)(.*?\n)',
167 167 bygroups(Name.Exception, Literal.Number.Integer, Other)),
168 168 # (Exception Identifier)(Message)
169 169 (r'(?u)(^[^\d\W]\w*)(:.*?\n)',
170 170 bygroups(Name.Exception, Text)),
171 171 # Tag everything else as Other, will be handled later.
172 172 (r'.*\n', Other),
173 173 ],
174 174 }
175 175
176 176
177 177 class IPythonTracebackLexer(DelegatingLexer):
178 178 """
179 179 IPython traceback lexer.
180 180
181 181 For doctests, the tracebacks can be snipped as much as desired with the
182 182 exception to the lines that designate a traceback. For non-syntax error
183 183 tracebacks, this is the line of hyphens. For syntax error tracebacks,
184 184 this is the line which lists the File and line number.
185 185
186 186 """
187 187 # The lexer inherits from DelegatingLexer. The "root" lexer is an
188 188 # appropriate IPython lexer, which depends on the value of the boolean
189 189 # `python3`. First, we parse with the partial IPython traceback lexer.
190 190 # Then, any code marked with the "Other" token is delegated to the root
191 191 # lexer.
192 192 #
193 193 name = 'IPython Traceback'
194 194 aliases = ['ipythontb']
195 195
196 196 def __init__(self, **options):
197 197 self.python3 = get_bool_opt(options, 'python3', False)
198 198 if self.python3:
199 199 self.aliases = ['ipython3tb']
200 200 else:
201 201 self.aliases = ['ipython2tb', 'ipythontb']
202 202
203 203 if self.python3:
204 204 IPyLexer = IPython3Lexer
205 205 else:
206 206 IPyLexer = IPythonLexer
207 207
208 208 DelegatingLexer.__init__(self, IPyLexer,
209 209 IPythonPartialTracebackLexer, **options)
210 210
211 211 class IPythonConsoleLexer(Lexer):
212 212 """
213 213 An IPython console lexer for IPython code-blocks and doctests, such as:
214 214
215 215 .. code-block:: rst
216 216
217 217 .. code-block:: ipythonconsole
218 218
219 219 In [1]: a = 'foo'
220 220
221 221 In [2]: a
222 222 Out[2]: 'foo'
223 223
224 In [3]: print a
224 In [3]: print(a)
225 225 foo
226 226
227 In [4]: 1 / 0
228
229 227
230 228 Support is also provided for IPython exceptions:
231 229
232 230 .. code-block:: rst
233 231
234 232 .. code-block:: ipythonconsole
235 233
236 234 In [1]: raise Exception
237
238 ---------------------------------------------------------------------------
239 Exception Traceback (most recent call last)
240 <ipython-input-1-fca2ab0ca76b> in <module>
241 ----> 1 raise Exception
242
243 Exception:
235 Traceback (most recent call last):
236 ...
237 Exception
244 238
245 239 """
246 240 name = 'IPython console session'
247 241 aliases = ['ipythonconsole']
248 242 mimetypes = ['text/x-ipython-console']
249 243
250 244 # The regexps used to determine what is input and what is output.
251 245 # The default prompts for IPython are:
252 246 #
253 247 # in = 'In [#]: '
254 248 # continuation = ' .D.: '
255 249 # template = 'Out[#]: '
256 250 #
257 251 # Where '#' is the 'prompt number' or 'execution count' and 'D'
258 252 # D is a number of dots matching the width of the execution count
259 253 #
260 254 in1_regex = r'In \[[0-9]+\]: '
261 255 in2_regex = r' \.\.+\.: '
262 256 out_regex = r'Out\[[0-9]+\]: '
263 257
264 258 #: The regex to determine when a traceback starts.
265 259 ipytb_start = re.compile(r'^(\^C)?(-+\n)|^( File)(.*)(, line )(\d+\n)')
266 260
267 261 def __init__(self, **options):
268 262 """Initialize the IPython console lexer.
269 263
270 264 Parameters
271 265 ----------
272 266 python3 : bool
273 267 If `True`, then the console inputs are parsed using a Python 3
274 268 lexer. Otherwise, they are parsed using a Python 2 lexer.
275 269 in1_regex : RegexObject
276 270 The compiled regular expression used to detect the start
277 271 of inputs. Although the IPython configuration setting may have a
278 272 trailing whitespace, do not include it in the regex. If `None`,
279 273 then the default input prompt is assumed.
280 274 in2_regex : RegexObject
281 275 The compiled regular expression used to detect the continuation
282 276 of inputs. Although the IPython configuration setting may have a
283 277 trailing whitespace, do not include it in the regex. If `None`,
284 278 then the default input prompt is assumed.
285 279 out_regex : RegexObject
286 280 The compiled regular expression used to detect outputs. If `None`,
287 281 then the default output prompt is assumed.
288 282
289 283 """
290 284 self.python3 = get_bool_opt(options, 'python3', False)
291 285 if self.python3:
292 286 self.aliases = ['ipython3console']
293 287 else:
294 288 self.aliases = ['ipython2console', 'ipythonconsole']
295 289
296 290 in1_regex = options.get('in1_regex', self.in1_regex)
297 291 in2_regex = options.get('in2_regex', self.in2_regex)
298 292 out_regex = options.get('out_regex', self.out_regex)
299 293
300 294 # So that we can work with input and output prompts which have been
301 295 # rstrip'd (possibly by editors) we also need rstrip'd variants. If
302 296 # we do not do this, then such prompts will be tagged as 'output'.
303 297 # The reason can't just use the rstrip'd variants instead is because
304 298 # we want any whitespace associated with the prompt to be inserted
305 299 # with the token. This allows formatted code to be modified so as hide
306 300 # the appearance of prompts, with the whitespace included. One example
307 301 # use of this is in copybutton.js from the standard lib Python docs.
308 302 in1_regex_rstrip = in1_regex.rstrip() + '\n'
309 303 in2_regex_rstrip = in2_regex.rstrip() + '\n'
310 304 out_regex_rstrip = out_regex.rstrip() + '\n'
311 305
312 306 # Compile and save them all.
313 307 attrs = ['in1_regex', 'in2_regex', 'out_regex',
314 308 'in1_regex_rstrip', 'in2_regex_rstrip', 'out_regex_rstrip']
315 309 for attr in attrs:
316 310 self.__setattr__(attr, re.compile(locals()[attr]))
317 311
318 312 Lexer.__init__(self, **options)
319 313
320 314 if self.python3:
321 315 pylexer = IPython3Lexer
322 316 tblexer = IPythonTracebackLexer
323 317 else:
324 318 pylexer = IPythonLexer
325 319 tblexer = IPythonTracebackLexer
326 320
327 321 self.pylexer = pylexer(**options)
328 322 self.tblexer = tblexer(**options)
329 323
330 324 self.reset()
331 325
332 326 def reset(self):
333 327 self.mode = 'output'
334 328 self.index = 0
335 329 self.buffer = u''
336 330 self.insertions = []
337 331
338 332 def buffered_tokens(self):
339 333 """
340 334 Generator of unprocessed tokens after doing insertions and before
341 335 changing to a new state.
342 336
343 337 """
344 338 if self.mode == 'output':
345 339 tokens = [(0, Generic.Output, self.buffer)]
346 340 elif self.mode == 'input':
347 341 tokens = self.pylexer.get_tokens_unprocessed(self.buffer)
348 342 else: # traceback
349 343 tokens = self.tblexer.get_tokens_unprocessed(self.buffer)
350 344
351 345 for i, t, v in do_insertions(self.insertions, tokens):
352 346 # All token indexes are relative to the buffer.
353 347 yield self.index + i, t, v
354 348
355 349 # Clear it all
356 350 self.index += len(self.buffer)
357 351 self.buffer = u''
358 352 self.insertions = []
359 353
360 354 def get_mci(self, line):
361 355 """
362 356 Parses the line and returns a 3-tuple: (mode, code, insertion).
363 357
364 358 `mode` is the next mode (or state) of the lexer, and is always equal
365 359 to 'input', 'output', or 'tb'.
366 360
367 361 `code` is a portion of the line that should be added to the buffer
368 362 corresponding to the next mode and eventually lexed by another lexer.
369 363 For example, `code` could be Python code if `mode` were 'input'.
370 364
371 365 `insertion` is a 3-tuple (index, token, text) representing an
372 366 unprocessed "token" that will be inserted into the stream of tokens
373 367 that are created from the buffer once we change modes. This is usually
374 368 the input or output prompt.
375 369
376 370 In general, the next mode depends on current mode and on the contents
377 371 of `line`.
378 372
379 373 """
380 374 # To reduce the number of regex match checks, we have multiple
381 375 # 'if' blocks instead of 'if-elif' blocks.
382 376
383 377 # Check for possible end of input
384 378 in2_match = self.in2_regex.match(line)
385 379 in2_match_rstrip = self.in2_regex_rstrip.match(line)
386 380 if (in2_match and in2_match.group().rstrip() == line.rstrip()) or \
387 381 in2_match_rstrip:
388 382 end_input = True
389 383 else:
390 384 end_input = False
391 385 if end_input and self.mode != 'tb':
392 386 # Only look for an end of input when not in tb mode.
393 387 # An ellipsis could appear within the traceback.
394 388 mode = 'output'
395 389 code = u''
396 390 insertion = (0, Generic.Prompt, line)
397 391 return mode, code, insertion
398 392
399 393 # Check for output prompt
400 394 out_match = self.out_regex.match(line)
401 395 out_match_rstrip = self.out_regex_rstrip.match(line)
402 396 if out_match or out_match_rstrip:
403 397 mode = 'output'
404 398 if out_match:
405 399 idx = out_match.end()
406 400 else:
407 401 idx = out_match_rstrip.end()
408 402 code = line[idx:]
409 403 # Use the 'heading' token for output. We cannot use Generic.Error
410 404 # since it would conflict with exceptions.
411 405 insertion = (0, Generic.Heading, line[:idx])
412 406 return mode, code, insertion
413 407
414 408
415 409 # Check for input or continuation prompt (non stripped version)
416 410 in1_match = self.in1_regex.match(line)
417 411 if in1_match or (in2_match and self.mode != 'tb'):
418 412 # New input or when not in tb, continued input.
419 413 # We do not check for continued input when in tb since it is
420 414 # allowable to replace a long stack with an ellipsis.
421 415 mode = 'input'
422 416 if in1_match:
423 417 idx = in1_match.end()
424 418 else: # in2_match
425 419 idx = in2_match.end()
426 420 code = line[idx:]
427 421 insertion = (0, Generic.Prompt, line[:idx])
428 422 return mode, code, insertion
429 423
430 424 # Check for input or continuation prompt (stripped version)
431 425 in1_match_rstrip = self.in1_regex_rstrip.match(line)
432 426 if in1_match_rstrip or (in2_match_rstrip and self.mode != 'tb'):
433 427 # New input or when not in tb, continued input.
434 428 # We do not check for continued input when in tb since it is
435 429 # allowable to replace a long stack with an ellipsis.
436 430 mode = 'input'
437 431 if in1_match_rstrip:
438 432 idx = in1_match_rstrip.end()
439 433 else: # in2_match
440 434 idx = in2_match_rstrip.end()
441 435 code = line[idx:]
442 436 insertion = (0, Generic.Prompt, line[:idx])
443 437 return mode, code, insertion
444 438
445 439 # Check for traceback
446 440 if self.ipytb_start.match(line):
447 441 mode = 'tb'
448 442 code = line
449 443 insertion = None
450 444 return mode, code, insertion
451 445
452 446 # All other stuff...
453 447 if self.mode in ('input', 'output'):
454 448 # We assume all other text is output. Multiline input that
455 449 # does not use the continuation marker cannot be detected.
456 450 # For example, the 3 in the following is clearly output:
457 451 #
458 452 # In [1]: print 3
459 453 # 3
460 454 #
461 455 # But the following second line is part of the input:
462 456 #
463 457 # In [2]: while True:
464 458 # print True
465 459 #
466 460 # In both cases, the 2nd line will be 'output'.
467 461 #
468 462 mode = 'output'
469 463 else:
470 464 mode = 'tb'
471 465
472 466 code = line
473 467 insertion = None
474 468
475 469 return mode, code, insertion
476 470
477 471 def get_tokens_unprocessed(self, text):
478 472 self.reset()
479 473 for match in line_re.finditer(text):
480 474 line = match.group()
481 475 mode, code, insertion = self.get_mci(line)
482 476
483 477 if mode != self.mode:
484 478 # Yield buffered tokens before transitioning to new mode.
485 479 for token in self.buffered_tokens():
486 480 yield token
487 481 self.mode = mode
488 482
489 483 if insertion:
490 484 self.insertions.append((len(self.buffer), [insertion]))
491 485 self.buffer += code
492 486
493 487 for token in self.buffered_tokens():
494 488 yield token
495 489
496 490 class IPyLexer(Lexer):
497 491 r"""
498 492 Primary lexer for all IPython-like code.
499 493
500 494 This is a simple helper lexer. If the first line of the text begins with
501 495 "In \[[0-9]+\]:", then the entire text is parsed with an IPython console
502 496 lexer. If not, then the entire text is parsed with an IPython lexer.
503 497
504 498 The goal is to reduce the number of lexers that are registered
505 499 with Pygments.
506 500
507 501 """
508 502 name = 'IPy session'
509 503 aliases = ['ipy']
510 504
511 505 def __init__(self, **options):
512 506 self.python3 = get_bool_opt(options, 'python3', False)
513 507 if self.python3:
514 508 self.aliases = ['ipy3']
515 509 else:
516 510 self.aliases = ['ipy2', 'ipy']
517 511
518 512 Lexer.__init__(self, **options)
519 513
520 514 self.IPythonLexer = IPythonLexer(**options)
521 515 self.IPythonConsoleLexer = IPythonConsoleLexer(**options)
522 516
523 517 def get_tokens_unprocessed(self, text):
524 518 # Search for the input prompt anywhere...this allows code blocks to
525 519 # begin with comments as well.
526 520 if re.match(r'.*(In \[[0-9]+\]:)', text.strip(), re.DOTALL):
527 521 lex = self.IPythonConsoleLexer
528 522 else:
529 523 lex = self.IPythonLexer
530 524 for token in lex.get_tokens_unprocessed(text):
531 525 yield token
532 526
General Comments 0
You need to be logged in to leave comments. Login now