##// END OF EJS Templates
Misc fix for Python 3.13 testing (unicode range and skip tests) (#14278)...
Misc fix for Python 3.13 testing (unicode range and skip tests) (#14278) Update unicode range for python 3.13, and skip jedi test while in alpha as there are no grammar available.

File last commit:

r28494:56f48b87 revert-14190-1402...
r28562:90276e81 merge
Show More
lexers.py
540 lines | 20.5 KiB | text/x-python | PythonLexer
Thomas Kluyver
Move IPython lexers module to lib...
r20625 # -*- coding: utf-8 -*-
"""
Defines a variety of Pygments lexers for highlighting IPython code.
This includes:
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 IPythonLexer, IPython3Lexer
Lexers for pure IPython (python + magic/shell commands)
Thomas Kluyver
Move IPython lexers module to lib...
r20625
IPythonPartialTracebackLexer, IPythonTracebackLexer
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 Supports 2.x and 3.x via keyword `python3`. The partial traceback
lexer reads everything but the Python code appearing in a traceback.
The full lexer combines the partial lexer with an IPython lexer.
Thomas Kluyver
Move IPython lexers module to lib...
r20625
IPythonConsoleLexer
A lexer for IPython console sessions, with support for tracebacks.
IPyLexer
A friendly lexer which examines the first line of text and from it,
decides whether to use an IPython lexer or an IPython console lexer.
This is probably the only lexer that needs to be explicitly added
to Pygments.
"""
#-----------------------------------------------------------------------------
# Copyright (c) 2013, the IPython Development Team.
#
# Distributed under the terms of the Modified BSD License.
#
# The full license is in the file COPYING.txt, distributed with this software.
#-----------------------------------------------------------------------------
# Standard library
import re
# Third party
Jörg Dietrich
Add ipython_tokens for syntax highlighting following cell magic
r24344 from pygments.lexers import (
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 BashLexer, HtmlLexer, JavascriptLexer, RubyLexer, PerlLexer, PythonLexer,
Python3Lexer, TexLexer)
Thomas Kluyver
Move IPython lexers module to lib...
r20625 from pygments.lexer import (
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 Lexer, DelegatingLexer, RegexLexer, do_insertions, bygroups, using,
Thomas Kluyver
Move IPython lexers module to lib...
r20625 )
from pygments.token import (
Matthias Bussonnier
A few style fixes:...
r22354 Generic, Keyword, Literal, Name, Operator, Other, Text, Error,
Thomas Kluyver
Move IPython lexers module to lib...
r20625 )
from pygments.util import get_bool_opt
# Local
line_re = re.compile('.*?\n')
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 __all__ = ['build_ipy_lexer', 'IPython3Lexer', 'IPythonLexer',
'IPythonPartialTracebackLexer', 'IPythonTracebackLexer',
'IPythonConsoleLexer', 'IPyLexer']
Thomas Kluyver
Move IPython lexers module to lib...
r20625
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 def build_ipy_lexer(python3):
"""Builds IPython lexers depending on the value of `python3`.
Thomas Kluyver
Move IPython lexers module to lib...
r20625
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 The lexer inherits from an appropriate Python lexer and then adds
information about IPython specific keywords (i.e. magic commands,
shell commands, etc.)
Thomas Kluyver
Move IPython lexers module to lib...
r20625
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 Parameters
----------
python3 : bool
If `True`, then build an IPython lexer from a Python 3 lexer.
"""
# It would be nice to have a single IPython lexer class which takes
# a boolean `python3`. But since there are two Python lexer classes,
# we will also have two IPython lexer classes.
if python3:
PyLexer = Python3Lexer
name = 'IPython3'
aliases = ['ipython3']
doc = """IPython3 Lexer"""
else:
PyLexer = PythonLexer
name = 'IPython'
aliases = ['ipython2', 'ipython']
doc = """IPython Lexer"""
ipython_tokens = [
(r'(?s)(\s*)(%%capture)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
(r'(?s)(\s*)(%%debug)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
(r'(?is)(\s*)(%%html)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(HtmlLexer))),
(r'(?s)(\s*)(%%javascript)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(JavascriptLexer))),
(r'(?s)(\s*)(%%js)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(JavascriptLexer))),
(r'(?s)(\s*)(%%latex)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(TexLexer))),
(r'(?s)(\s*)(%%perl)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PerlLexer))),
(r'(?s)(\s*)(%%prun)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
(r'(?s)(\s*)(%%pypy)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
(r'(?s)(\s*)(%%python)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
(r'(?s)(\s*)(%%python2)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PythonLexer))),
(r'(?s)(\s*)(%%python3)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(Python3Lexer))),
(r'(?s)(\s*)(%%ruby)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(RubyLexer))),
(r'(?s)(\s*)(%%time)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
(r'(?s)(\s*)(%%timeit)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
(r'(?s)(\s*)(%%writefile)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
(r'(?s)(\s*)(%%file)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(PyLexer))),
(r"(?s)(\s*)(%%)(\w+)(.*)", bygroups(Text, Operator, Keyword, Text)),
(r'(?s)(^\s*)(%%!)([^\n]*\n)(.*)', bygroups(Text, Operator, Text, using(BashLexer))),
(r"(%%?)(\w+)(\?\??)$", bygroups(Operator, Keyword, Operator)),
(r"\b(\?\??)(\s*)$", bygroups(Operator, Text)),
(r'(%)(sx|sc|system)(.*)(\n)', bygroups(Operator, Keyword,
using(BashLexer), Text)),
(r'(%)(\w+)(.*\n)', bygroups(Operator, Keyword, Text)),
(r'^(!!)(.+)(\n)', bygroups(Operator, using(BashLexer), Text)),
(r'(!)(?!=)(.+)(\n)', bygroups(Operator, using(BashLexer), Text)),
(r'^(\s*)(\?\??)(\s*%{0,2}[\w\.\*]*)', bygroups(Text, Operator, Text)),
(r'(\s*%{0,2}[\w\.\*]*)(\?\??)(\s*)$', bygroups(Text, Operator, Text)),
]
tokens = PyLexer.tokens.copy()
tokens['root'] = ipython_tokens + tokens['root']
attrs = {'name': name, 'aliases': aliases, 'filenames': [],
'__doc__': doc, 'tokens': tokens}
return type(name, (PyLexer,), attrs)
IPython3Lexer = build_ipy_lexer(python3=True)
IPythonLexer = build_ipy_lexer(python3=False)
Thomas Kluyver
Move IPython lexers module to lib...
r20625
class IPythonPartialTracebackLexer(RegexLexer):
"""
Partial lexer for IPython tracebacks.
Terry Davis
Remove python 2 references from docstrings, where appropriate.
r25450 Handles all the non-python output.
Thomas Kluyver
Move IPython lexers module to lib...
r20625
"""
name = 'IPython Partial Traceback'
tokens = {
'root': [
# Tracebacks for syntax errors have a different style.
# For both types of tracebacks, we mark the first line with
# Generic.Traceback. For syntax errors, we mark the filename
# as we mark the filenames for non-syntax tracebacks.
#
# These two regexps define how IPythonConsoleLexer finds a
# traceback.
#
## Non-syntax traceback
(r'^(\^C)?(-+\n)', bygroups(Error, Generic.Traceback)),
## Syntax traceback
(r'^( File)(.*)(, line )(\d+\n)',
bygroups(Generic.Traceback, Name.Namespace,
Generic.Traceback, Literal.Number.Integer)),
# (Exception Identifier)(Whitespace)(Traceback Message)
(r'(?u)(^[^\d\W]\w*)(\s*)(Traceback.*?\n)',
bygroups(Name.Exception, Generic.Whitespace, Text)),
# (Module/Filename)(Text)(Callee)(Function Signature)
# Better options for callee and function signature?
(r'(.*)( in )(.*)(\(.*\)\n)',
bygroups(Name.Namespace, Text, Name.Entity, Name.Tag)),
# Regular line: (Whitespace)(Line Number)(Python Code)
(r'(\s*?)(\d+)(.*?\n)',
bygroups(Generic.Whitespace, Literal.Number.Integer, Other)),
# Emphasized line: (Arrow)(Line Number)(Python Code)
# Using Exception token so arrow color matches the Exception.
(r'(-*>?\s?)(\d+)(.*?\n)',
bygroups(Name.Exception, Literal.Number.Integer, Other)),
# (Exception Identifier)(Message)
(r'(?u)(^[^\d\W]\w*)(:.*?\n)',
bygroups(Name.Exception, Text)),
# Tag everything else as Other, will be handled later.
(r'.*\n', Other),
],
}
class IPythonTracebackLexer(DelegatingLexer):
"""
IPython traceback lexer.
For doctests, the tracebacks can be snipped as much as desired with the
exception to the lines that designate a traceback. For non-syntax error
tracebacks, this is the line of hyphens. For syntax error tracebacks,
this is the line which lists the File and line number.
"""
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 # The lexer inherits from DelegatingLexer. The "root" lexer is an
# appropriate IPython lexer, which depends on the value of the boolean
# `python3`. First, we parse with the partial IPython traceback lexer.
Thomas Kluyver
Move IPython lexers module to lib...
r20625 # Then, any code marked with the "Other" token is delegated to the root
# lexer.
#
name = 'IPython Traceback'
aliases = ['ipythontb']
def __init__(self, **options):
Matthias Bussonnier
Fix failing docs build....
r28215 """
A subclass of `DelegatingLexer` which delegates to the appropriate to either IPyLexer,
IPythonPartialTracebackLexer.
"""
# note we need a __init__ doc, as otherwise it inherits the doc from the super class
# which will fail the documentation build as it references section of the pygments docs that
# do not exists when building IPython's docs.
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 self.python3 = get_bool_opt(options, 'python3', False)
if self.python3:
self.aliases = ['ipython3tb']
else:
self.aliases = ['ipython2tb', 'ipythontb']
Thomas Kluyver
Move IPython lexers module to lib...
r20625
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 if self.python3:
IPyLexer = IPython3Lexer
else:
IPyLexer = IPythonLexer
Thomas Kluyver
Move IPython lexers module to lib...
r20625
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 DelegatingLexer.__init__(self, IPyLexer,
IPythonPartialTracebackLexer, **options)
Thomas Kluyver
Move IPython lexers module to lib...
r20625
class IPythonConsoleLexer(Lexer):
"""
An IPython console lexer for IPython code-blocks and doctests, such as:
.. code-block:: rst
.. code-block:: ipythonconsole
In [1]: a = 'foo'
In [2]: a
Out[2]: 'foo'
Nikita Kniazev
Fix IPythonConsoleLexer doctest
r26935 In [3]: print(a)
Thomas Kluyver
Move IPython lexers module to lib...
r20625 foo
Support is also provided for IPython exceptions:
.. code-block:: rst
.. code-block:: ipythonconsole
In [1]: raise Exception
Nikita Kniazev
Fix IPythonConsoleLexer doctest
r26935 Traceback (most recent call last):
...
Exception
Thomas Kluyver
Move IPython lexers module to lib...
r20625
"""
name = 'IPython console session'
aliases = ['ipythonconsole']
mimetypes = ['text/x-ipython-console']
# The regexps used to determine what is input and what is output.
# The default prompts for IPython are:
#
Matthias Bussonnier
Remove a few usage of PromptManager in example...
r22447 # in = 'In [#]: '
# continuation = ' .D.: '
# template = 'Out[#]: '
#
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 # Where '#' is the 'prompt number' or 'execution count' and 'D'
# D is a number of dots matching the width of the execution count
Thomas Kluyver
Move IPython lexers module to lib...
r20625 #
in1_regex = r'In \[[0-9]+\]: '
in2_regex = r' \.\.+\.: '
out_regex = r'Out\[[0-9]+\]: '
#: The regex to determine when a traceback starts.
ipytb_start = re.compile(r'^(\^C)?(-+\n)|^( File)(.*)(, line )(\d+\n)')
def __init__(self, **options):
"""Initialize the IPython console lexer.
Parameters
----------
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 python3 : bool
If `True`, then the console inputs are parsed using a Python 3
lexer. Otherwise, they are parsed using a Python 2 lexer.
Thomas Kluyver
Move IPython lexers module to lib...
r20625 in1_regex : RegexObject
The compiled regular expression used to detect the start
of inputs. Although the IPython configuration setting may have a
trailing whitespace, do not include it in the regex. If `None`,
then the default input prompt is assumed.
in2_regex : RegexObject
The compiled regular expression used to detect the continuation
of inputs. Although the IPython configuration setting may have a
trailing whitespace, do not include it in the regex. If `None`,
then the default input prompt is assumed.
out_regex : RegexObject
The compiled regular expression used to detect outputs. If `None`,
then the default output prompt is assumed.
"""
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 self.python3 = get_bool_opt(options, 'python3', False)
if self.python3:
self.aliases = ['ipython3console']
else:
self.aliases = ['ipython2console', 'ipythonconsole']
Thomas Kluyver
Move IPython lexers module to lib...
r20625
in1_regex = options.get('in1_regex', self.in1_regex)
in2_regex = options.get('in2_regex', self.in2_regex)
out_regex = options.get('out_regex', self.out_regex)
# So that we can work with input and output prompts which have been
# rstrip'd (possibly by editors) we also need rstrip'd variants. If
# we do not do this, then such prompts will be tagged as 'output'.
# The reason can't just use the rstrip'd variants instead is because
# we want any whitespace associated with the prompt to be inserted
# with the token. This allows formatted code to be modified so as hide
# the appearance of prompts, with the whitespace included. One example
# use of this is in copybutton.js from the standard lib Python docs.
in1_regex_rstrip = in1_regex.rstrip() + '\n'
in2_regex_rstrip = in2_regex.rstrip() + '\n'
out_regex_rstrip = out_regex.rstrip() + '\n'
# Compile and save them all.
attrs = ['in1_regex', 'in2_regex', 'out_regex',
'in1_regex_rstrip', 'in2_regex_rstrip', 'out_regex_rstrip']
for attr in attrs:
self.__setattr__(attr, re.compile(locals()[attr]))
Lexer.__init__(self, **options)
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 if self.python3:
pylexer = IPython3Lexer
tblexer = IPythonTracebackLexer
else:
pylexer = IPythonLexer
tblexer = IPythonTracebackLexer
self.pylexer = pylexer(**options)
self.tblexer = tblexer(**options)
Thomas Kluyver
Move IPython lexers module to lib...
r20625
self.reset()
def reset(self):
self.mode = 'output'
self.index = 0
self.buffer = u''
self.insertions = []
def buffered_tokens(self):
"""
Generator of unprocessed tokens after doing insertions and before
changing to a new state.
"""
if self.mode == 'output':
tokens = [(0, Generic.Output, self.buffer)]
elif self.mode == 'input':
tokens = self.pylexer.get_tokens_unprocessed(self.buffer)
else: # traceback
tokens = self.tblexer.get_tokens_unprocessed(self.buffer)
for i, t, v in do_insertions(self.insertions, tokens):
# All token indexes are relative to the buffer.
yield self.index + i, t, v
# Clear it all
self.index += len(self.buffer)
self.buffer = u''
self.insertions = []
def get_mci(self, line):
"""
Parses the line and returns a 3-tuple: (mode, code, insertion).
`mode` is the next mode (or state) of the lexer, and is always equal
to 'input', 'output', or 'tb'.
`code` is a portion of the line that should be added to the buffer
corresponding to the next mode and eventually lexed by another lexer.
For example, `code` could be Python code if `mode` were 'input'.
`insertion` is a 3-tuple (index, token, text) representing an
unprocessed "token" that will be inserted into the stream of tokens
that are created from the buffer once we change modes. This is usually
the input or output prompt.
In general, the next mode depends on current mode and on the contents
of `line`.
"""
# To reduce the number of regex match checks, we have multiple
# 'if' blocks instead of 'if-elif' blocks.
# Check for possible end of input
in2_match = self.in2_regex.match(line)
in2_match_rstrip = self.in2_regex_rstrip.match(line)
if (in2_match and in2_match.group().rstrip() == line.rstrip()) or \
in2_match_rstrip:
end_input = True
else:
end_input = False
if end_input and self.mode != 'tb':
# Only look for an end of input when not in tb mode.
# An ellipsis could appear within the traceback.
mode = 'output'
code = u''
insertion = (0, Generic.Prompt, line)
return mode, code, insertion
# Check for output prompt
out_match = self.out_regex.match(line)
out_match_rstrip = self.out_regex_rstrip.match(line)
if out_match or out_match_rstrip:
mode = 'output'
if out_match:
idx = out_match.end()
else:
idx = out_match_rstrip.end()
code = line[idx:]
# Use the 'heading' token for output. We cannot use Generic.Error
# since it would conflict with exceptions.
insertion = (0, Generic.Heading, line[:idx])
return mode, code, insertion
# Check for input or continuation prompt (non stripped version)
in1_match = self.in1_regex.match(line)
if in1_match or (in2_match and self.mode != 'tb'):
# New input or when not in tb, continued input.
# We do not check for continued input when in tb since it is
# allowable to replace a long stack with an ellipsis.
mode = 'input'
if in1_match:
idx = in1_match.end()
else: # in2_match
idx = in2_match.end()
code = line[idx:]
insertion = (0, Generic.Prompt, line[:idx])
return mode, code, insertion
# Check for input or continuation prompt (stripped version)
in1_match_rstrip = self.in1_regex_rstrip.match(line)
if in1_match_rstrip or (in2_match_rstrip and self.mode != 'tb'):
# New input or when not in tb, continued input.
# We do not check for continued input when in tb since it is
# allowable to replace a long stack with an ellipsis.
mode = 'input'
if in1_match_rstrip:
idx = in1_match_rstrip.end()
else: # in2_match
idx = in2_match_rstrip.end()
code = line[idx:]
insertion = (0, Generic.Prompt, line[:idx])
return mode, code, insertion
# Check for traceback
if self.ipytb_start.match(line):
mode = 'tb'
code = line
insertion = None
return mode, code, insertion
# All other stuff...
if self.mode in ('input', 'output'):
# We assume all other text is output. Multiline input that
# does not use the continuation marker cannot be detected.
# For example, the 3 in the following is clearly output:
#
# In [1]: print 3
# 3
#
# But the following second line is part of the input:
#
# In [2]: while True:
# print True
#
# In both cases, the 2nd line will be 'output'.
#
mode = 'output'
else:
mode = 'tb'
code = line
insertion = None
return mode, code, insertion
def get_tokens_unprocessed(self, text):
self.reset()
for match in line_re.finditer(text):
line = match.group()
mode, code, insertion = self.get_mci(line)
if mode != self.mode:
# Yield buffered tokens before transitioning to new mode.
for token in self.buffered_tokens():
yield token
self.mode = mode
if insertion:
self.insertions.append((len(self.buffer), [insertion]))
self.buffer += code
Matthias Bussonnier
A few style fixes:...
r22354
for token in self.buffered_tokens():
yield token
Thomas Kluyver
Move IPython lexers module to lib...
r20625
class IPyLexer(Lexer):
Matthias Bussonnier
fix some other syntax warnings
r24780 r"""
Thomas Kluyver
Move IPython lexers module to lib...
r20625 Primary lexer for all IPython-like code.
This is a simple helper lexer. If the first line of the text begins with
"In \[[0-9]+\]:", then the entire text is parsed with an IPython console
lexer. If not, then the entire text is parsed with an IPython lexer.
The goal is to reduce the number of lexers that are registered
with Pygments.
"""
name = 'IPy session'
aliases = ['ipy']
def __init__(self, **options):
Matthias Bussonnier
Fix failing docs build....
r28215 """
Create a new IPyLexer instance which dispatch to either an
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 IPythonCOnsoleLexer (if In prompts are present) or and IPythonLexer (if
Matthias Bussonnier
Fix failing docs build....
r28215 In prompts are not present).
"""
# init docstring is necessary for docs not to fail to build do to parent
# docs referenceing a section in pygments docs.
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 self.python3 = get_bool_opt(options, 'python3', False)
if self.python3:
self.aliases = ['ipy3']
else:
self.aliases = ['ipy2', 'ipy']
Thomas Kluyver
Move IPython lexers module to lib...
r20625
Lexer.__init__(self, **options)
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494 self.IPythonLexer = IPythonLexer(**options)
Thomas Kluyver
Move IPython lexers module to lib...
r20625 self.IPythonConsoleLexer = IPythonConsoleLexer(**options)
def get_tokens_unprocessed(self, text):
# Search for the input prompt anywhere...this allows code blocks to
# begin with comments as well.
if re.match(r'.*(In \[[0-9]+\]:)', text.strip(), re.DOTALL):
lex = self.IPythonConsoleLexer
else:
lex = self.IPythonLexer
for token in lex.get_tokens_unprocessed(text):
yield token
Matthias Bussonnier
Revert "Cleanup Python 2 compact from Lexers"
r28494