##// END OF EJS Templates
Fix typos
Fix typos

File last commit:

r28888:f98cd150
r28888:f98cd150
Show More
test_completer.py
1769 lines | 60.7 KiB | text/x-python | PythonLexer
# encoding: utf-8
"""Tests for the IPython tab-completion machinery."""
# Copyright (c) IPython Development Team.
# Distributed under the terms of the Modified BSD License.
import os
import pytest
import sys
import textwrap
import unittest
from importlib.metadata import version
from contextlib import contextmanager
from traitlets.config.loader import Config
from IPython import get_ipython
from IPython.core import completer
from IPython.utils.tempdir import TemporaryDirectory, TemporaryWorkingDirectory
from IPython.utils.generics import complete_object
from IPython.testing import decorators as dec
from IPython.core.completer import (
Completion,
provisionalcompleter,
match_dict_keys,
_deduplicate_completions,
_match_number_in_dict_key_prefix,
completion_matcher,
SimpleCompletion,
CompletionContext,
)
from packaging.version import parse
# -----------------------------------------------------------------------------
# Test functions
# -----------------------------------------------------------------------------
def recompute_unicode_ranges():
"""
utility to recompute the largest unicode range without any characters
use to recompute the gap in the global _UNICODE_RANGES of completer.py
"""
import itertools
import unicodedata
valid = []
for c in range(0, 0x10FFFF + 1):
try:
unicodedata.name(chr(c))
except ValueError:
continue
valid.append(c)
def ranges(i):
for a, b in itertools.groupby(enumerate(i), lambda pair: pair[1] - pair[0]):
b = list(b)
yield b[0][1], b[-1][1]
rg = list(ranges(valid))
lens = []
gap_lens = []
pstart, pstop = 0, 0
for start, stop in rg:
lens.append(stop - start)
gap_lens.append(
(
start - pstop,
hex(pstop + 1),
hex(start),
f"{round((start - pstop)/0xe01f0*100)}%",
)
)
pstart, pstop = start, stop
return sorted(gap_lens)[-1]
def test_unicode_range():
"""
Test that the ranges we test for unicode names give the same number of
results than testing the full length.
"""
from IPython.core.completer import _unicode_name_compute, _UNICODE_RANGES
expected_list = _unicode_name_compute([(0, 0x110000)])
test = _unicode_name_compute(_UNICODE_RANGES)
len_exp = len(expected_list)
len_test = len(test)
# do not inline the len() or on error pytest will try to print the 130 000 +
# elements.
message = None
if len_exp != len_test or len_exp > 131808:
size, start, stop, prct = recompute_unicode_ranges()
message = f"""_UNICODE_RANGES likely wrong and need updating. This is
likely due to a new release of Python. We've find that the biggest gap
in unicode characters has reduces in size to be {size} characters
({prct}), from {start}, to {stop}. In completer.py likely update to
_UNICODE_RANGES = [(32, {start}), ({stop}, 0xe01f0)]
And update the assertion below to use
len_exp <= {len_exp}
"""
assert len_exp == len_test, message
# fail if new unicode symbols have been added.
assert len_exp <= 143668, message
@contextmanager
def greedy_completion():
ip = get_ipython()
greedy_original = ip.Completer.greedy
try:
ip.Completer.greedy = True
yield
finally:
ip.Completer.greedy = greedy_original
@contextmanager
def evaluation_policy(evaluation: str):
ip = get_ipython()
evaluation_original = ip.Completer.evaluation
try:
ip.Completer.evaluation = evaluation
yield
finally:
ip.Completer.evaluation = evaluation_original
@contextmanager
def custom_matchers(matchers):
ip = get_ipython()
try:
ip.Completer.custom_matchers.extend(matchers)
yield
finally:
ip.Completer.custom_matchers.clear()
def test_protect_filename():
if sys.platform == "win32":
pairs = [
("abc", "abc"),
(" abc", '" abc"'),
("a bc", '"a bc"'),
("a bc", '"a bc"'),
(" bc", '" bc"'),
]
else:
pairs = [
("abc", "abc"),
(" abc", r"\ abc"),
("a bc", r"a\ bc"),
("a bc", r"a\ \ bc"),
(" bc", r"\ \ bc"),
# On posix, we also protect parens and other special characters.
("a(bc", r"a\(bc"),
("a)bc", r"a\)bc"),
("a( )bc", r"a\(\ \)bc"),
("a[1]bc", r"a\[1\]bc"),
("a{1}bc", r"a\{1\}bc"),
("a#bc", r"a\#bc"),
("a?bc", r"a\?bc"),
("a=bc", r"a\=bc"),
("a\\bc", r"a\\bc"),
("a|bc", r"a\|bc"),
("a;bc", r"a\;bc"),
("a:bc", r"a\:bc"),
("a'bc", r"a\'bc"),
("a*bc", r"a\*bc"),
('a"bc', r"a\"bc"),
("a^bc", r"a\^bc"),
("a&bc", r"a\&bc"),
]
# run the actual tests
for s1, s2 in pairs:
s1p = completer.protect_filename(s1)
assert s1p == s2
def check_line_split(splitter, test_specs):
for part1, part2, split in test_specs:
cursor_pos = len(part1)
line = part1 + part2
out = splitter.split_line(line, cursor_pos)
assert out == split
def test_line_split():
"""Basic line splitter test with default specs."""
sp = completer.CompletionSplitter()
# The format of the test specs is: part1, part2, expected answer. Parts 1
# and 2 are joined into the 'line' sent to the splitter, as if the cursor
# was at the end of part1. So an empty part2 represents someone hitting
# tab at the end of the line, the most common case.
t = [
("run some/script", "", "some/script"),
("run scripts/er", "ror.py foo", "scripts/er"),
("echo $HOM", "", "HOM"),
("print sys.pa", "", "sys.pa"),
("print(sys.pa", "", "sys.pa"),
("execfile('scripts/er", "", "scripts/er"),
("a[x.", "", "x."),
("a[x.", "y", "x."),
('cd "some_file/', "", "some_file/"),
]
check_line_split(sp, t)
# Ensure splitting works OK with unicode by re-running the tests with
# all inputs turned into unicode
check_line_split(sp, [map(str, p) for p in t])
class NamedInstanceClass:
instances = {}
def __init__(self, name):
self.instances[name] = self
@classmethod
def _ipython_key_completions_(cls):
return cls.instances.keys()
class KeyCompletable:
def __init__(self, things=()):
self.things = things
def _ipython_key_completions_(self):
return list(self.things)
class TestCompleter(unittest.TestCase):
def setUp(self):
"""
We want to silence all PendingDeprecationWarning when testing the completer
"""
self._assertwarns = self.assertWarns(PendingDeprecationWarning)
self._assertwarns.__enter__()
def tearDown(self):
try:
self._assertwarns.__exit__(None, None, None)
except AssertionError:
pass
def test_custom_completion_error(self):
"""Test that errors from custom attribute completers are silenced."""
ip = get_ipython()
class A:
pass
ip.user_ns["x"] = A()
@complete_object.register(A)
def complete_A(a, existing_completions):
raise TypeError("this should be silenced")
ip.complete("x.")
def test_custom_completion_ordering(self):
"""Test that errors from custom attribute completers are silenced."""
ip = get_ipython()
_, matches = ip.complete('in')
assert matches.index('input') < matches.index('int')
def complete_example(a):
return ['example2', 'example1']
ip.Completer.custom_completers.add_re('ex*', complete_example)
_, matches = ip.complete('ex')
assert matches.index('example2') < matches.index('example1')
def test_unicode_completions(self):
ip = get_ipython()
# Some strings that trigger different types of completion. Check them both
# in str and unicode forms
s = ["ru", "%ru", "cd /", "floa", "float(x)/"]
for t in s + list(map(str, s)):
# We don't need to check exact completion values (they may change
# depending on the state of the namespace, but at least no exceptions
# should be thrown and the return value should be a pair of text, list
# values.
text, matches = ip.complete(t)
self.assertIsInstance(text, str)
self.assertIsInstance(matches, list)
def test_latex_completions(self):
from IPython.core.latex_symbols import latex_symbols
import random
ip = get_ipython()
# Test some random unicode symbols
keys = random.sample(sorted(latex_symbols), 10)
for k in keys:
text, matches = ip.complete(k)
self.assertEqual(text, k)
self.assertEqual(matches, [latex_symbols[k]])
# Test a more complex line
text, matches = ip.complete("print(\\alpha")
self.assertEqual(text, "\\alpha")
self.assertEqual(matches[0], latex_symbols["\\alpha"])
# Test multiple matching latex symbols
text, matches = ip.complete("\\al")
self.assertIn("\\alpha", matches)
self.assertIn("\\aleph", matches)
def test_latex_no_results(self):
"""
forward latex should really return nothing in either field if nothing is found.
"""
ip = get_ipython()
text, matches = ip.Completer.latex_matches("\\really_i_should_match_nothing")
self.assertEqual(text, "")
self.assertEqual(matches, ())
def test_back_latex_completion(self):
ip = get_ipython()
# do not return more than 1 matches for \beta, only the latex one.
name, matches = ip.complete("\\β")
self.assertEqual(matches, ["\\beta"])
def test_back_unicode_completion(self):
ip = get_ipython()
name, matches = ip.complete("\\â…¤")
self.assertEqual(matches, ["\\ROMAN NUMERAL FIVE"])
def test_forward_unicode_completion(self):
ip = get_ipython()
name, matches = ip.complete("\\ROMAN NUMERAL FIVE")
self.assertEqual(matches, ["â…¤"]) # This is not a V
self.assertEqual(matches, ["\u2164"]) # same as above but explicit.
def test_delim_setting(self):
sp = completer.CompletionSplitter()
sp.delims = " "
self.assertEqual(sp.delims, " ")
self.assertEqual(sp._delim_expr, r"[\ ]")
def test_spaces(self):
"""Test with only spaces as split chars."""
sp = completer.CompletionSplitter()
sp.delims = " "
t = [("foo", "", "foo"), ("run foo", "", "foo"), ("run foo", "bar", "foo")]
check_line_split(sp, t)
def test_has_open_quotes1(self):
for s in ["'", "'''", "'hi' '"]:
self.assertEqual(completer.has_open_quotes(s), "'")
def test_has_open_quotes2(self):
for s in ['"', '"""', '"hi" "']:
self.assertEqual(completer.has_open_quotes(s), '"')
def test_has_open_quotes3(self):
for s in ["''", "''' '''", "'hi' 'ipython'"]:
self.assertFalse(completer.has_open_quotes(s))
def test_has_open_quotes4(self):
for s in ['""', '""" """', '"hi" "ipython"']:
self.assertFalse(completer.has_open_quotes(s))
@pytest.mark.xfail(
sys.platform == "win32", reason="abspath completions fail on Windows"
)
def test_abspath_file_completions(self):
ip = get_ipython()
with TemporaryDirectory() as tmpdir:
prefix = os.path.join(tmpdir, "foo")
suffixes = ["1", "2"]
names = [prefix + s for s in suffixes]
for n in names:
open(n, "w", encoding="utf-8").close()
# Check simple completion
c = ip.complete(prefix)[1]
self.assertEqual(c, names)
# Now check with a function call
cmd = 'a = f("%s' % prefix
c = ip.complete(prefix, cmd)[1]
comp = [prefix + s for s in suffixes]
self.assertEqual(c, comp)
def test_local_file_completions(self):
ip = get_ipython()
with TemporaryWorkingDirectory():
prefix = "./foo"
suffixes = ["1", "2"]
names = [prefix + s for s in suffixes]
for n in names:
open(n, "w", encoding="utf-8").close()
# Check simple completion
c = ip.complete(prefix)[1]
self.assertEqual(c, names)
# Now check with a function call
cmd = 'a = f("%s' % prefix
c = ip.complete(prefix, cmd)[1]
comp = {prefix + s for s in suffixes}
self.assertTrue(comp.issubset(set(c)))
def test_quoted_file_completions(self):
ip = get_ipython()
def _(text):
return ip.Completer._complete(
cursor_line=0, cursor_pos=len(text), full_text=text
)["IPCompleter.file_matcher"]["completions"]
with TemporaryWorkingDirectory():
name = "foo'bar"
open(name, "w", encoding="utf-8").close()
# Don't escape Windows
escaped = name if sys.platform == "win32" else "foo\\'bar"
# Single quote matches embedded single quote
c = _("open('foo")[0]
self.assertEqual(c.text, escaped)
# Double quote requires no escape
c = _('open("foo')[0]
self.assertEqual(c.text, name)
# No quote requires an escape
c = _("%ls foo")[0]
self.assertEqual(c.text, escaped)
@pytest.mark.xfail(
sys.version_info.releaselevel in ("alpha",),
reason="Parso does not yet parse 3.13",
)
def test_all_completions_dups(self):
"""
Make sure the output of `IPCompleter.all_completions` does not have
duplicated prefixes.
"""
ip = get_ipython()
c = ip.Completer
ip.ex("class TestClass():\n\ta=1\n\ta1=2")
for jedi_status in [True, False]:
with provisionalcompleter():
ip.Completer.use_jedi = jedi_status
matches = c.all_completions("TestCl")
assert matches == ["TestClass"], (jedi_status, matches)
matches = c.all_completions("TestClass.")
assert len(matches) > 2, (jedi_status, matches)
matches = c.all_completions("TestClass.a")
if jedi_status:
assert matches == ["TestClass.a", "TestClass.a1"], jedi_status
else:
assert matches == [".a", ".a1"], jedi_status
@pytest.mark.xfail(
sys.version_info.releaselevel in ("alpha",),
reason="Parso does not yet parse 3.13",
)
def test_jedi(self):
"""
A couple of issue we had with Jedi
"""
ip = get_ipython()
def _test_complete(reason, s, comp, start=None, end=None):
l = len(s)
start = start if start is not None else l
end = end if end is not None else l
with provisionalcompleter():
ip.Completer.use_jedi = True
completions = set(ip.Completer.completions(s, l))
ip.Completer.use_jedi = False
assert Completion(start, end, comp) in completions, reason
def _test_not_complete(reason, s, comp):
l = len(s)
with provisionalcompleter():
ip.Completer.use_jedi = True
completions = set(ip.Completer.completions(s, l))
ip.Completer.use_jedi = False
assert Completion(l, l, comp) not in completions, reason
import jedi
jedi_version = tuple(int(i) for i in jedi.__version__.split(".")[:3])
if jedi_version > (0, 10):
_test_complete("jedi >0.9 should complete and not crash", "a=1;a.", "real")
_test_complete("can infer first argument", 'a=(1,"foo");a[0].', "real")
_test_complete("can infer second argument", 'a=(1,"foo");a[1].', "capitalize")
_test_complete("cover duplicate completions", "im", "import", 0, 2)
_test_not_complete("does not mix types", 'a=(1,"foo");a[0].', "capitalize")
@pytest.mark.xfail(
sys.version_info.releaselevel in ("alpha",),
reason="Parso does not yet parse 3.13",
)
def test_completion_have_signature(self):
"""
Lets make sure jedi is capable of pulling out the signature of the function we are completing.
"""
ip = get_ipython()
with provisionalcompleter():
ip.Completer.use_jedi = True
completions = ip.Completer.completions("ope", 3)
c = next(completions) # should be `open`
ip.Completer.use_jedi = False
assert "file" in c.signature, "Signature of function was not found by completer"
assert (
"encoding" in c.signature
), "Signature of function was not found by completer"
@pytest.mark.xfail(
sys.version_info.releaselevel in ("alpha",),
reason="Parso does not yet parse 3.13",
)
def test_completions_have_type(self):
"""
Lets make sure matchers provide completion type.
"""
ip = get_ipython()
with provisionalcompleter():
ip.Completer.use_jedi = False
completions = ip.Completer.completions("%tim", 3)
c = next(completions) # should be `%time` or similar
assert c.type == "magic", "Type of magic was not assigned by completer"
@pytest.mark.xfail(
parse(version("jedi")) <= parse("0.18.0"),
reason="Known failure on jedi<=0.18.0",
strict=True,
)
def test_deduplicate_completions(self):
"""
Test that completions are correctly deduplicated (even if ranges are not the same)
"""
ip = get_ipython()
ip.ex(
textwrap.dedent(
"""
class Z:
zoo = 1
"""
)
)
with provisionalcompleter():
ip.Completer.use_jedi = True
l = list(
_deduplicate_completions("Z.z", ip.Completer.completions("Z.z", 3))
)
ip.Completer.use_jedi = False
assert len(l) == 1, "Completions (Z.z<tab>) correctly deduplicate: %s " % l
assert l[0].text == "zoo" # and not `it.accumulate`
@pytest.mark.xfail(
sys.version_info.releaselevel in ("alpha",),
reason="Parso does not yet parse 3.13",
)
def test_greedy_completions(self):
"""
Test the capability of the Greedy completer.
Most of the test here does not really show off the greedy completer, for proof
each of the text below now pass with Jedi. The greedy completer is capable of more.
See the :any:`test_dict_key_completion_contexts`
"""
ip = get_ipython()
ip.ex("a=list(range(5))")
ip.ex("d = {'a b': str}")
_, c = ip.complete(".", line="a[0].")
self.assertFalse(".real" in c, "Shouldn't have completed on a[0]: %s" % c)
def _(line, cursor_pos, expect, message, completion):
with greedy_completion(), provisionalcompleter():
ip.Completer.use_jedi = False
_, c = ip.complete(".", line=line, cursor_pos=cursor_pos)
self.assertIn(expect, c, message % c)
ip.Completer.use_jedi = True
with provisionalcompleter():
completions = ip.Completer.completions(line, cursor_pos)
self.assertIn(completion, list(completions))
with provisionalcompleter():
_(
"a[0].",
5,
".real",
"Should have completed on a[0].: %s",
Completion(5, 5, "real"),
)
_(
"a[0].r",
6,
".real",
"Should have completed on a[0].r: %s",
Completion(5, 6, "real"),
)
_(
"a[0].from_",
10,
".from_bytes",
"Should have completed on a[0].from_: %s",
Completion(5, 10, "from_bytes"),
)
_(
"assert str.star",
14,
".startswith",
"Should have completed on `assert str.star`: %s",
Completion(11, 14, "startswith"),
)
_(
"d['a b'].str",
12,
".strip",
"Should have completed on `d['a b'].str`: %s",
Completion(9, 12, "strip"),
)
_(
"a.app",
4,
".append",
"Should have completed on `a.app`: %s",
Completion(2, 4, "append"),
)
def test_omit__names(self):
# also happens to test IPCompleter as a configurable
ip = get_ipython()
ip._hidden_attr = 1
ip._x = {}
c = ip.Completer
ip.ex("ip=get_ipython()")
cfg = Config()
cfg.IPCompleter.omit__names = 0
c.update_config(cfg)
with provisionalcompleter():
c.use_jedi = False
s, matches = c.complete("ip.")
self.assertIn(".__str__", matches)
self.assertIn("._hidden_attr", matches)
# c.use_jedi = True
# completions = set(c.completions('ip.', 3))
# self.assertIn(Completion(3, 3, '__str__'), completions)
# self.assertIn(Completion(3,3, "_hidden_attr"), completions)
cfg = Config()
cfg.IPCompleter.omit__names = 1
c.update_config(cfg)
with provisionalcompleter():
c.use_jedi = False
s, matches = c.complete("ip.")
self.assertNotIn(".__str__", matches)
# self.assertIn('ip._hidden_attr', matches)
# c.use_jedi = True
# completions = set(c.completions('ip.', 3))
# self.assertNotIn(Completion(3,3,'__str__'), completions)
# self.assertIn(Completion(3,3, "_hidden_attr"), completions)
cfg = Config()
cfg.IPCompleter.omit__names = 2
c.update_config(cfg)
with provisionalcompleter():
c.use_jedi = False
s, matches = c.complete("ip.")
self.assertNotIn(".__str__", matches)
self.assertNotIn("._hidden_attr", matches)
# c.use_jedi = True
# completions = set(c.completions('ip.', 3))
# self.assertNotIn(Completion(3,3,'__str__'), completions)
# self.assertNotIn(Completion(3,3, "_hidden_attr"), completions)
with provisionalcompleter():
c.use_jedi = False
s, matches = c.complete("ip._x.")
self.assertIn(".keys", matches)
# c.use_jedi = True
# completions = set(c.completions('ip._x.', 6))
# self.assertIn(Completion(6,6, "keys"), completions)
del ip._hidden_attr
del ip._x
def test_limit_to__all__False_ok(self):
"""
Limit to all is deprecated, once we remove it this test can go away.
"""
ip = get_ipython()
c = ip.Completer
c.use_jedi = False
ip.ex("class D: x=24")
ip.ex("d=D()")
cfg = Config()
cfg.IPCompleter.limit_to__all__ = False
c.update_config(cfg)
s, matches = c.complete("d.")
self.assertIn(".x", matches)
def test_get__all__entries_ok(self):
class A:
__all__ = ["x", 1]
words = completer.get__all__entries(A())
self.assertEqual(words, ["x"])
def test_get__all__entries_no__all__ok(self):
class A:
pass
words = completer.get__all__entries(A())
self.assertEqual(words, [])
def test_func_kw_completions(self):
ip = get_ipython()
c = ip.Completer
c.use_jedi = False
ip.ex("def myfunc(a=1,b=2): return a+b")
s, matches = c.complete(None, "myfunc(1,b")
self.assertIn("b=", matches)
# Simulate completing with cursor right after b (pos==10):
s, matches = c.complete(None, "myfunc(1,b)", 10)
self.assertIn("b=", matches)
s, matches = c.complete(None, 'myfunc(a="escaped\\")string",b')
self.assertIn("b=", matches)
# builtin function
s, matches = c.complete(None, "min(k, k")
self.assertIn("key=", matches)
def test_default_arguments_from_docstring(self):
ip = get_ipython()
c = ip.Completer
kwd = c._default_arguments_from_docstring("min(iterable[, key=func]) -> value")
self.assertEqual(kwd, ["key"])
# with cython type etc
kwd = c._default_arguments_from_docstring(
"Minuit.migrad(self, int ncall=10000, resume=True, int nsplit=1)\n"
)
self.assertEqual(kwd, ["ncall", "resume", "nsplit"])
# white spaces
kwd = c._default_arguments_from_docstring(
"\n Minuit.migrad(self, int ncall=10000, resume=True, int nsplit=1)\n"
)
self.assertEqual(kwd, ["ncall", "resume", "nsplit"])
def test_line_magics(self):
ip = get_ipython()
c = ip.Completer
s, matches = c.complete(None, "lsmag")
self.assertIn("%lsmagic", matches)
s, matches = c.complete(None, "%lsmag")
self.assertIn("%lsmagic", matches)
def test_cell_magics(self):
from IPython.core.magic import register_cell_magic
@register_cell_magic
def _foo_cellm(line, cell):
pass
ip = get_ipython()
c = ip.Completer
s, matches = c.complete(None, "_foo_ce")
self.assertIn("%%_foo_cellm", matches)
s, matches = c.complete(None, "%%_foo_ce")
self.assertIn("%%_foo_cellm", matches)
def test_line_cell_magics(self):
from IPython.core.magic import register_line_cell_magic
@register_line_cell_magic
def _bar_cellm(line, cell):
pass
ip = get_ipython()
c = ip.Completer
# The policy here is trickier, see comments in completion code. The
# returned values depend on whether the user passes %% or not explicitly,
# and this will show a difference if the same name is both a line and cell
# magic.
s, matches = c.complete(None, "_bar_ce")
self.assertIn("%_bar_cellm", matches)
self.assertIn("%%_bar_cellm", matches)
s, matches = c.complete(None, "%_bar_ce")
self.assertIn("%_bar_cellm", matches)
self.assertIn("%%_bar_cellm", matches)
s, matches = c.complete(None, "%%_bar_ce")
self.assertNotIn("%_bar_cellm", matches)
self.assertIn("%%_bar_cellm", matches)
def test_magic_completion_order(self):
ip = get_ipython()
c = ip.Completer
# Test ordering of line and cell magics.
text, matches = c.complete("timeit")
self.assertEqual(matches, ["%timeit", "%%timeit"])
def test_magic_completion_shadowing(self):
ip = get_ipython()
c = ip.Completer
c.use_jedi = False
# Before importing matplotlib, %matplotlib magic should be the only option.
text, matches = c.complete("mat")
self.assertEqual(matches, ["%matplotlib"])
# The newly introduced name should shadow the magic.
ip.run_cell("matplotlib = 1")
text, matches = c.complete("mat")
self.assertEqual(matches, ["matplotlib"])
# After removing matplotlib from namespace, the magic should again be
# the only option.
del ip.user_ns["matplotlib"]
text, matches = c.complete("mat")
self.assertEqual(matches, ["%matplotlib"])
def test_magic_completion_shadowing_explicit(self):
"""
If the user try to complete a shadowed magic, and explicit % start should
still return the completions.
"""
ip = get_ipython()
c = ip.Completer
# Before importing matplotlib, %matplotlib magic should be the only option.
text, matches = c.complete("%mat")
self.assertEqual(matches, ["%matplotlib"])
ip.run_cell("matplotlib = 1")
# After removing matplotlib from namespace, the magic should still be
# the only option.
text, matches = c.complete("%mat")
self.assertEqual(matches, ["%matplotlib"])
def test_magic_config(self):
ip = get_ipython()
c = ip.Completer
s, matches = c.complete(None, "conf")
self.assertIn("%config", matches)
s, matches = c.complete(None, "conf")
self.assertNotIn("AliasManager", matches)
s, matches = c.complete(None, "config ")
self.assertIn("AliasManager", matches)
s, matches = c.complete(None, "%config ")
self.assertIn("AliasManager", matches)
s, matches = c.complete(None, "config Ali")
self.assertListEqual(["AliasManager"], matches)
s, matches = c.complete(None, "%config Ali")
self.assertListEqual(["AliasManager"], matches)
s, matches = c.complete(None, "config AliasManager")
self.assertListEqual(["AliasManager"], matches)
s, matches = c.complete(None, "%config AliasManager")
self.assertListEqual(["AliasManager"], matches)
s, matches = c.complete(None, "config AliasManager.")
self.assertIn("AliasManager.default_aliases", matches)
s, matches = c.complete(None, "%config AliasManager.")
self.assertIn("AliasManager.default_aliases", matches)
s, matches = c.complete(None, "config AliasManager.de")
self.assertListEqual(["AliasManager.default_aliases"], matches)
s, matches = c.complete(None, "config AliasManager.de")
self.assertListEqual(["AliasManager.default_aliases"], matches)
def test_magic_color(self):
ip = get_ipython()
c = ip.Completer
s, matches = c.complete(None, "colo")
self.assertIn("%colors", matches)
s, matches = c.complete(None, "colo")
self.assertNotIn("NoColor", matches)
s, matches = c.complete(None, "%colors") # No trailing space
self.assertNotIn("NoColor", matches)
s, matches = c.complete(None, "colors ")
self.assertIn("NoColor", matches)
s, matches = c.complete(None, "%colors ")
self.assertIn("NoColor", matches)
s, matches = c.complete(None, "colors NoCo")
self.assertListEqual(["NoColor"], matches)
s, matches = c.complete(None, "%colors NoCo")
self.assertListEqual(["NoColor"], matches)
def test_match_dict_keys(self):
"""
Test that match_dict_keys works on a couple of use case does return what
expected, and does not crash
"""
delims = " \t\n`!@#$^&*()=+[{]}\\|;:'\",<>?"
def match(*args, **kwargs):
quote, offset, matches = match_dict_keys(*args, delims=delims, **kwargs)
return quote, offset, list(matches)
keys = ["foo", b"far"]
assert match(keys, "b'") == ("'", 2, ["far"])
assert match(keys, "b'f") == ("'", 2, ["far"])
assert match(keys, 'b"') == ('"', 2, ["far"])
assert match(keys, 'b"f') == ('"', 2, ["far"])
assert match(keys, "'") == ("'", 1, ["foo"])
assert match(keys, "'f") == ("'", 1, ["foo"])
assert match(keys, '"') == ('"', 1, ["foo"])
assert match(keys, '"f') == ('"', 1, ["foo"])
# Completion on first item of tuple
keys = [("foo", 1111), ("foo", 2222), (3333, "bar"), (3333, "test")]
assert match(keys, "'f") == ("'", 1, ["foo"])
assert match(keys, "33") == ("", 0, ["3333"])
# Completion on numbers
keys = [
0xDEADBEEF,
1111,
1234,
"1999",
0b10101,
22,
] # 0xDEADBEEF = 3735928559; 0b10101 = 21
assert match(keys, "0xdead") == ("", 0, ["0xdeadbeef"])
assert match(keys, "1") == ("", 0, ["1111", "1234"])
assert match(keys, "2") == ("", 0, ["21", "22"])
assert match(keys, "0b101") == ("", 0, ["0b10101", "0b10110"])
# Should yield on variables
assert match(keys, "a_variable") == ("", 0, [])
# Should pass over invalid literals
assert match(keys, "'' ''") == ("", 0, [])
def test_match_dict_keys_tuple(self):
"""
Test that match_dict_keys called with extra prefix works on a couple of use case,
does return what expected, and does not crash.
"""
delims = " \t\n`!@#$^&*()=+[{]}\\|;:'\",<>?"
keys = [("foo", "bar"), ("foo", "oof"), ("foo", b"bar"), ('other', 'test')]
def match(*args, extra=None, **kwargs):
quote, offset, matches = match_dict_keys(
*args, delims=delims, extra_prefix=extra, **kwargs
)
return quote, offset, list(matches)
# Completion on first key == "foo"
assert match(keys, "'", extra=("foo",)) == ("'", 1, ["bar", "oof"])
assert match(keys, '"', extra=("foo",)) == ('"', 1, ["bar", "oof"])
assert match(keys, "'o", extra=("foo",)) == ("'", 1, ["oof"])
assert match(keys, '"o', extra=("foo",)) == ('"', 1, ["oof"])
assert match(keys, "b'", extra=("foo",)) == ("'", 2, ["bar"])
assert match(keys, 'b"', extra=("foo",)) == ('"', 2, ["bar"])
assert match(keys, "b'b", extra=("foo",)) == ("'", 2, ["bar"])
assert match(keys, 'b"b', extra=("foo",)) == ('"', 2, ["bar"])
# No Completion
assert match(keys, "'", extra=("no_foo",)) == ("'", 1, [])
assert match(keys, "'", extra=("fo",)) == ("'", 1, [])
keys = [("foo1", "foo2", "foo3", "foo4"), ("foo1", "foo2", "bar", "foo4")]
assert match(keys, "'foo", extra=("foo1",)) == ("'", 1, ["foo2"])
assert match(keys, "'foo", extra=("foo1", "foo2")) == ("'", 1, ["foo3"])
assert match(keys, "'foo", extra=("foo1", "foo2", "foo3")) == ("'", 1, ["foo4"])
assert match(keys, "'foo", extra=("foo1", "foo2", "foo3", "foo4")) == (
"'",
1,
[],
)
keys = [("foo", 1111), ("foo", "2222"), (3333, "bar"), (3333, 4444)]
assert match(keys, "'", extra=("foo",)) == ("'", 1, ["2222"])
assert match(keys, "", extra=("foo",)) == ("", 0, ["1111", "'2222'"])
assert match(keys, "'", extra=(3333,)) == ("'", 1, ["bar"])
assert match(keys, "", extra=(3333,)) == ("", 0, ["'bar'", "4444"])
assert match(keys, "'", extra=("3333",)) == ("'", 1, [])
assert match(keys, "33") == ("", 0, ["3333"])
def test_dict_key_completion_closures(self):
ip = get_ipython()
complete = ip.Completer.complete
ip.Completer.auto_close_dict_keys = True
ip.user_ns["d"] = {
# tuple only
("aa", 11): None,
# tuple and non-tuple
("bb", 22): None,
"bb": None,
# non-tuple only
"cc": None,
# numeric tuple only
(77, "x"): None,
# numeric tuple and non-tuple
(88, "y"): None,
88: None,
# numeric non-tuple only
99: None,
}
_, matches = complete(line_buffer="d[")
# should append `, ` if matches a tuple only
self.assertIn("'aa', ", matches)
# should not append anything if matches a tuple and an item
self.assertIn("'bb'", matches)
# should append `]` if matches and item only
self.assertIn("'cc']", matches)
# should append `, ` if matches a tuple only
self.assertIn("77, ", matches)
# should not append anything if matches a tuple and an item
self.assertIn("88", matches)
# should append `]` if matches and item only
self.assertIn("99]", matches)
_, matches = complete(line_buffer="d['aa', ")
# should restrict matches to those matching tuple prefix
self.assertIn("11]", matches)
self.assertNotIn("'bb'", matches)
self.assertNotIn("'bb', ", matches)
self.assertNotIn("'bb']", matches)
self.assertNotIn("'cc'", matches)
self.assertNotIn("'cc', ", matches)
self.assertNotIn("'cc']", matches)
ip.Completer.auto_close_dict_keys = False
def test_dict_key_completion_string(self):
"""Test dictionary key completion for string keys"""
ip = get_ipython()
complete = ip.Completer.complete
ip.user_ns["d"] = {"abc": None}
# check completion at different stages
_, matches = complete(line_buffer="d[")
self.assertIn("'abc'", matches)
self.assertNotIn("'abc']", matches)
_, matches = complete(line_buffer="d['")
self.assertIn("abc", matches)
self.assertNotIn("abc']", matches)
_, matches = complete(line_buffer="d['a")
self.assertIn("abc", matches)
self.assertNotIn("abc']", matches)
# check use of different quoting
_, matches = complete(line_buffer='d["')
self.assertIn("abc", matches)
self.assertNotIn('abc"]', matches)
_, matches = complete(line_buffer='d["a')
self.assertIn("abc", matches)
self.assertNotIn('abc"]', matches)
# check sensitivity to following context
_, matches = complete(line_buffer="d[]", cursor_pos=2)
self.assertIn("'abc'", matches)
_, matches = complete(line_buffer="d['']", cursor_pos=3)
self.assertIn("abc", matches)
self.assertNotIn("abc'", matches)
self.assertNotIn("abc']", matches)
# check multiple solutions are correctly returned and that noise is not
ip.user_ns["d"] = {
"abc": None,
"abd": None,
"bad": None,
object(): None,
5: None,
("abe", None): None,
(None, "abf"): None
}
_, matches = complete(line_buffer="d['a")
self.assertIn("abc", matches)
self.assertIn("abd", matches)
self.assertNotIn("bad", matches)
self.assertNotIn("abe", matches)
self.assertNotIn("abf", matches)
assert not any(m.endswith(("]", '"', "'")) for m in matches), matches
# check escaping and whitespace
ip.user_ns["d"] = {"a\nb": None, "a'b": None, 'a"b': None, "a word": None}
_, matches = complete(line_buffer="d['a")
self.assertIn("a\\nb", matches)
self.assertIn("a\\'b", matches)
self.assertIn('a"b', matches)
self.assertIn("a word", matches)
assert not any(m.endswith(("]", '"', "'")) for m in matches), matches
# - can complete on non-initial word of the string
_, matches = complete(line_buffer="d['a w")
self.assertIn("word", matches)
# - understands quote escaping
_, matches = complete(line_buffer="d['a\\'")
self.assertIn("b", matches)
# - default quoting should work like repr
_, matches = complete(line_buffer="d[")
self.assertIn('"a\'b"', matches)
# - when opening quote with ", possible to match with unescaped apostrophe
_, matches = complete(line_buffer="d[\"a'")
self.assertIn("b", matches)
# need to not split at delims that readline won't split at
if "-" not in ip.Completer.splitter.delims:
ip.user_ns["d"] = {"before-after": None}
_, matches = complete(line_buffer="d['before-af")
self.assertIn("before-after", matches)
# check completion on tuple-of-string keys at different stage - on first key
ip.user_ns["d"] = {('foo', 'bar'): None}
_, matches = complete(line_buffer="d[")
self.assertIn("'foo'", matches)
self.assertNotIn("'foo']", matches)
self.assertNotIn("'bar'", matches)
self.assertNotIn("foo", matches)
self.assertNotIn("bar", matches)
# - match the prefix
_, matches = complete(line_buffer="d['f")
self.assertIn("foo", matches)
self.assertNotIn("foo']", matches)
self.assertNotIn('foo"]', matches)
_, matches = complete(line_buffer="d['foo")
self.assertIn("foo", matches)
# - can complete on second key
_, matches = complete(line_buffer="d['foo', ")
self.assertIn("'bar'", matches)
_, matches = complete(line_buffer="d['foo', 'b")
self.assertIn("bar", matches)
self.assertNotIn("foo", matches)
# - does not propose missing keys
_, matches = complete(line_buffer="d['foo', 'f")
self.assertNotIn("bar", matches)
self.assertNotIn("foo", matches)
# check sensitivity to following context
_, matches = complete(line_buffer="d['foo',]", cursor_pos=8)
self.assertIn("'bar'", matches)
self.assertNotIn("bar", matches)
self.assertNotIn("'foo'", matches)
self.assertNotIn("foo", matches)
_, matches = complete(line_buffer="d['']", cursor_pos=3)
self.assertIn("foo", matches)
assert not any(m.endswith(("]", '"', "'")) for m in matches), matches
_, matches = complete(line_buffer='d[""]', cursor_pos=3)
self.assertIn("foo", matches)
assert not any(m.endswith(("]", '"', "'")) for m in matches), matches
_, matches = complete(line_buffer='d["foo","]', cursor_pos=9)
self.assertIn("bar", matches)
assert not any(m.endswith(("]", '"', "'")) for m in matches), matches
_, matches = complete(line_buffer='d["foo",]', cursor_pos=8)
self.assertIn("'bar'", matches)
self.assertNotIn("bar", matches)
# Can complete with longer tuple keys
ip.user_ns["d"] = {('foo', 'bar', 'foobar'): None}
# - can complete second key
_, matches = complete(line_buffer="d['foo', 'b")
self.assertIn("bar", matches)
self.assertNotIn("foo", matches)
self.assertNotIn("foobar", matches)
# - can complete third key
_, matches = complete(line_buffer="d['foo', 'bar', 'fo")
self.assertIn("foobar", matches)
self.assertNotIn("foo", matches)
self.assertNotIn("bar", matches)
def test_dict_key_completion_numbers(self):
ip = get_ipython()
complete = ip.Completer.complete
ip.user_ns["d"] = {
0xDEADBEEF: None, # 3735928559
1111: None,
1234: None,
"1999": None,
0b10101: None, # 21
22: None,
}
_, matches = complete(line_buffer="d[1")
self.assertIn("1111", matches)
self.assertIn("1234", matches)
self.assertNotIn("1999", matches)
self.assertNotIn("'1999'", matches)
_, matches = complete(line_buffer="d[0xdead")
self.assertIn("0xdeadbeef", matches)
_, matches = complete(line_buffer="d[2")
self.assertIn("21", matches)
self.assertIn("22", matches)
_, matches = complete(line_buffer="d[0b101")
self.assertIn("0b10101", matches)
self.assertIn("0b10110", matches)
def test_dict_key_completion_contexts(self):
"""Test expression contexts in which dict key completion occurs"""
ip = get_ipython()
complete = ip.Completer.complete
d = {"abc": None}
ip.user_ns["d"] = d
class C:
data = d
ip.user_ns["C"] = C
ip.user_ns["get"] = lambda: d
ip.user_ns["nested"] = {"x": d}
def assert_no_completion(**kwargs):
_, matches = complete(**kwargs)
self.assertNotIn("abc", matches)
self.assertNotIn("abc'", matches)
self.assertNotIn("abc']", matches)
self.assertNotIn("'abc'", matches)
self.assertNotIn("'abc']", matches)
def assert_completion(**kwargs):
_, matches = complete(**kwargs)
self.assertIn("'abc'", matches)
self.assertNotIn("'abc']", matches)
# no completion after string closed, even if reopened
assert_no_completion(line_buffer="d['a'")
assert_no_completion(line_buffer='d["a"')
assert_no_completion(line_buffer="d['a' + ")
assert_no_completion(line_buffer="d['a' + '")
# completion in non-trivial expressions
assert_completion(line_buffer="+ d[")
assert_completion(line_buffer="(d[")
assert_completion(line_buffer="C.data[")
# nested dict completion
assert_completion(line_buffer="nested['x'][")
with evaluation_policy("minimal"):
with pytest.raises(AssertionError):
assert_completion(line_buffer="nested['x'][")
# greedy flag
def assert_completion(**kwargs):
_, matches = complete(**kwargs)
self.assertIn("get()['abc']", matches)
assert_no_completion(line_buffer="get()[")
with greedy_completion():
assert_completion(line_buffer="get()[")
assert_completion(line_buffer="get()['")
assert_completion(line_buffer="get()['a")
assert_completion(line_buffer="get()['ab")
assert_completion(line_buffer="get()['abc")
def test_dict_key_completion_bytes(self):
"""Test handling of bytes in dict key completion"""
ip = get_ipython()
complete = ip.Completer.complete
ip.user_ns["d"] = {"abc": None, b"abd": None}
_, matches = complete(line_buffer="d[")
self.assertIn("'abc'", matches)
self.assertIn("b'abd'", matches)
if False: # not currently implemented
_, matches = complete(line_buffer="d[b")
self.assertIn("b'abd'", matches)
self.assertNotIn("b'abc'", matches)
_, matches = complete(line_buffer="d[b'")
self.assertIn("abd", matches)
self.assertNotIn("abc", matches)
_, matches = complete(line_buffer="d[B'")
self.assertIn("abd", matches)
self.assertNotIn("abc", matches)
_, matches = complete(line_buffer="d['")
self.assertIn("abc", matches)
self.assertNotIn("abd", matches)
def test_dict_key_completion_unicode_py3(self):
"""Test handling of unicode in dict key completion"""
ip = get_ipython()
complete = ip.Completer.complete
ip.user_ns["d"] = {"a\u05d0": None}
# query using escape
if sys.platform != "win32":
# Known failure on Windows
_, matches = complete(line_buffer="d['a\\u05d0")
self.assertIn("u05d0", matches) # tokenized after \\
# query using character
_, matches = complete(line_buffer="d['a\u05d0")
self.assertIn("a\u05d0", matches)
with greedy_completion():
# query using escape
_, matches = complete(line_buffer="d['a\\u05d0")
self.assertIn("d['a\\u05d0']", matches) # tokenized after \\
# query using character
_, matches = complete(line_buffer="d['a\u05d0")
self.assertIn("d['a\u05d0']", matches)
@dec.skip_without("numpy")
def test_struct_array_key_completion(self):
"""Test dict key completion applies to numpy struct arrays"""
import numpy
ip = get_ipython()
complete = ip.Completer.complete
ip.user_ns["d"] = numpy.array([], dtype=[("hello", "f"), ("world", "f")])
_, matches = complete(line_buffer="d['")
self.assertIn("hello", matches)
self.assertIn("world", matches)
# complete on the numpy struct itself
dt = numpy.dtype(
[("my_head", [("my_dt", ">u4"), ("my_df", ">u4")]), ("my_data", ">f4", 5)]
)
x = numpy.zeros(2, dtype=dt)
ip.user_ns["d"] = x[1]
_, matches = complete(line_buffer="d['")
self.assertIn("my_head", matches)
self.assertIn("my_data", matches)
def completes_on_nested():
ip.user_ns["d"] = numpy.zeros(2, dtype=dt)
_, matches = complete(line_buffer="d[1]['my_head']['")
self.assertTrue(any(["my_dt" in m for m in matches]))
self.assertTrue(any(["my_df" in m for m in matches]))
# complete on a nested level
with greedy_completion():
completes_on_nested()
with evaluation_policy("limited"):
completes_on_nested()
with evaluation_policy("minimal"):
with pytest.raises(AssertionError):
completes_on_nested()
@dec.skip_without("pandas")
def test_dataframe_key_completion(self):
"""Test dict key completion applies to pandas DataFrames"""
import pandas
ip = get_ipython()
complete = ip.Completer.complete
ip.user_ns["d"] = pandas.DataFrame({"hello": [1], "world": [2]})
_, matches = complete(line_buffer="d['")
self.assertIn("hello", matches)
self.assertIn("world", matches)
_, matches = complete(line_buffer="d.loc[:, '")
self.assertIn("hello", matches)
self.assertIn("world", matches)
_, matches = complete(line_buffer="d.loc[1:, '")
self.assertIn("hello", matches)
_, matches = complete(line_buffer="d.loc[1:1, '")
self.assertIn("hello", matches)
_, matches = complete(line_buffer="d.loc[1:1:-1, '")
self.assertIn("hello", matches)
_, matches = complete(line_buffer="d.loc[::, '")
self.assertIn("hello", matches)
def test_dict_key_completion_invalids(self):
"""Smoke test cases dict key completion can't handle"""
ip = get_ipython()
complete = ip.Completer.complete
ip.user_ns["no_getitem"] = None
ip.user_ns["no_keys"] = []
ip.user_ns["cant_call_keys"] = dict
ip.user_ns["empty"] = {}
ip.user_ns["d"] = {"abc": 5}
_, matches = complete(line_buffer="no_getitem['")
_, matches = complete(line_buffer="no_keys['")
_, matches = complete(line_buffer="cant_call_keys['")
_, matches = complete(line_buffer="empty['")
_, matches = complete(line_buffer="name_error['")
_, matches = complete(line_buffer="d['\\") # incomplete escape
def test_object_key_completion(self):
ip = get_ipython()
ip.user_ns["key_completable"] = KeyCompletable(["qwerty", "qwick"])
_, matches = ip.Completer.complete(line_buffer="key_completable['qw")
self.assertIn("qwerty", matches)
self.assertIn("qwick", matches)
def test_class_key_completion(self):
ip = get_ipython()
NamedInstanceClass("qwerty")
NamedInstanceClass("qwick")
ip.user_ns["named_instance_class"] = NamedInstanceClass
_, matches = ip.Completer.complete(line_buffer="named_instance_class['qw")
self.assertIn("qwerty", matches)
self.assertIn("qwick", matches)
def test_tryimport(self):
"""
Test that try-import don't crash on trailing dot, and import modules before
"""
from IPython.core.completerlib import try_import
assert try_import("IPython.")
def test_aimport_module_completer(self):
ip = get_ipython()
_, matches = ip.complete("i", "%aimport i")
self.assertIn("io", matches)
self.assertNotIn("int", matches)
def test_nested_import_module_completer(self):
ip = get_ipython()
_, matches = ip.complete(None, "import IPython.co", 17)
self.assertIn("IPython.core", matches)
self.assertNotIn("import IPython.core", matches)
self.assertNotIn("IPython.display", matches)
def test_import_module_completer(self):
ip = get_ipython()
_, matches = ip.complete("i", "import i")
self.assertIn("io", matches)
self.assertNotIn("int", matches)
def test_from_module_completer(self):
ip = get_ipython()
_, matches = ip.complete("B", "from io import B", 16)
self.assertIn("BytesIO", matches)
self.assertNotIn("BaseException", matches)
def test_snake_case_completion(self):
ip = get_ipython()
ip.Completer.use_jedi = False
ip.user_ns["some_three"] = 3
ip.user_ns["some_four"] = 4
_, matches = ip.complete("s_", "print(s_f")
self.assertIn("some_three", matches)
self.assertIn("some_four", matches)
def test_mix_terms(self):
ip = get_ipython()
from textwrap import dedent
ip.Completer.use_jedi = False
ip.ex(
dedent(
"""
class Test:
def meth(self, meth_arg1):
print("meth")
def meth_1(self, meth1_arg1, meth1_arg2):
print("meth1")
def meth_2(self, meth2_arg1, meth2_arg2):
print("meth2")
test = Test()
"""
)
)
_, matches = ip.complete(None, "test.meth(")
self.assertIn("meth_arg1=", matches)
self.assertNotIn("meth2_arg1=", matches)
def test_percent_symbol_restrict_to_magic_completions(self):
ip = get_ipython()
completer = ip.Completer
text = "%a"
with provisionalcompleter():
completer.use_jedi = True
completions = completer.completions(text, len(text))
for c in completions:
self.assertEqual(c.text[0], "%")
def test_fwd_unicode_restricts(self):
ip = get_ipython()
completer = ip.Completer
text = "\\ROMAN NUMERAL FIVE"
with provisionalcompleter():
completer.use_jedi = True
completions = [
completion.text for completion in completer.completions(text, len(text))
]
self.assertEqual(completions, ["\u2164"])
def test_dict_key_restrict_to_dicts(self):
"""Test that dict key suppresses non-dict completion items"""
ip = get_ipython()
c = ip.Completer
d = {"abc": None}
ip.user_ns["d"] = d
text = 'd["a'
def _():
with provisionalcompleter():
c.use_jedi = True
return [
completion.text for completion in c.completions(text, len(text))
]
completions = _()
self.assertEqual(completions, ["abc"])
# check that it can be disabled in granular manner:
cfg = Config()
cfg.IPCompleter.suppress_competing_matchers = {
"IPCompleter.dict_key_matcher": False
}
c.update_config(cfg)
completions = _()
self.assertIn("abc", completions)
self.assertGreater(len(completions), 1)
def test_matcher_suppression(self):
@completion_matcher(identifier="a_matcher")
def a_matcher(text):
return ["completion_a"]
@completion_matcher(identifier="b_matcher", api_version=2)
def b_matcher(context: CompletionContext):
text = context.token
result = {"completions": [SimpleCompletion("completion_b")]}
if text == "suppress c":
result["suppress"] = {"c_matcher"}
if text.startswith("suppress all"):
result["suppress"] = True
if text == "suppress all but c":
result["do_not_suppress"] = {"c_matcher"}
if text == "suppress all but a":
result["do_not_suppress"] = {"a_matcher"}
return result
@completion_matcher(identifier="c_matcher")
def c_matcher(text):
return ["completion_c"]
with custom_matchers([a_matcher, b_matcher, c_matcher]):
ip = get_ipython()
c = ip.Completer
def _(text, expected):
c.use_jedi = False
s, matches = c.complete(text)
self.assertEqual(expected, matches)
_("do not suppress", ["completion_a", "completion_b", "completion_c"])
_("suppress all", ["completion_b"])
_("suppress all but a", ["completion_a", "completion_b"])
_("suppress all but c", ["completion_b", "completion_c"])
def configure(suppression_config):
cfg = Config()
cfg.IPCompleter.suppress_competing_matchers = suppression_config
c.update_config(cfg)
# test that configuration takes priority over the run-time decisions
configure(False)
_("suppress all", ["completion_a", "completion_b", "completion_c"])
configure({"b_matcher": False})
_("suppress all", ["completion_a", "completion_b", "completion_c"])
configure({"a_matcher": False})
_("suppress all", ["completion_b"])
configure({"b_matcher": True})
_("do not suppress", ["completion_b"])
configure(True)
_("do not suppress", ["completion_a"])
def test_matcher_suppression_with_iterator(self):
@completion_matcher(identifier="matcher_returning_iterator")
def matcher_returning_iterator(text):
return iter(["completion_iter"])
@completion_matcher(identifier="matcher_returning_list")
def matcher_returning_list(text):
return ["completion_list"]
with custom_matchers([matcher_returning_iterator, matcher_returning_list]):
ip = get_ipython()
c = ip.Completer
def _(text, expected):
c.use_jedi = False
s, matches = c.complete(text)
self.assertEqual(expected, matches)
def configure(suppression_config):
cfg = Config()
cfg.IPCompleter.suppress_competing_matchers = suppression_config
c.update_config(cfg)
configure(False)
_("---", ["completion_iter", "completion_list"])
configure(True)
_("---", ["completion_iter"])
configure(None)
_("--", ["completion_iter", "completion_list"])
@pytest.mark.xfail(
sys.version_info.releaselevel in ("alpha",),
reason="Parso does not yet parse 3.13",
)
def test_matcher_suppression_with_jedi(self):
ip = get_ipython()
c = ip.Completer
c.use_jedi = True
def configure(suppression_config):
cfg = Config()
cfg.IPCompleter.suppress_competing_matchers = suppression_config
c.update_config(cfg)
def _():
with provisionalcompleter():
matches = [completion.text for completion in c.completions("dict.", 5)]
self.assertIn("keys", matches)
configure(False)
_()
configure(True)
_()
configure(None)
_()
def test_matcher_disabling(self):
@completion_matcher(identifier="a_matcher")
def a_matcher(text):
return ["completion_a"]
@completion_matcher(identifier="b_matcher")
def b_matcher(text):
return ["completion_b"]
def _(expected):
s, matches = c.complete("completion_")
self.assertEqual(expected, matches)
with custom_matchers([a_matcher, b_matcher]):
ip = get_ipython()
c = ip.Completer
_(["completion_a", "completion_b"])
cfg = Config()
cfg.IPCompleter.disable_matchers = ["b_matcher"]
c.update_config(cfg)
_(["completion_a"])
cfg.IPCompleter.disable_matchers = []
c.update_config(cfg)
def test_matcher_priority(self):
@completion_matcher(identifier="a_matcher", priority=0, api_version=2)
def a_matcher(text):
return {"completions": [SimpleCompletion("completion_a")], "suppress": True}
@completion_matcher(identifier="b_matcher", priority=2, api_version=2)
def b_matcher(text):
return {"completions": [SimpleCompletion("completion_b")], "suppress": True}
def _(expected):
s, matches = c.complete("completion_")
self.assertEqual(expected, matches)
with custom_matchers([a_matcher, b_matcher]):
ip = get_ipython()
c = ip.Completer
_(["completion_b"])
a_matcher.matcher_priority = 3
_(["completion_a"])
@pytest.mark.parametrize(
"input, expected",
[
["1.234", "1.234"],
# should match signed numbers
["+1", "+1"],
["-1", "-1"],
["-1.0", "-1.0"],
["-1.", "-1."],
["+1.", "+1."],
[".1", ".1"],
# should not match non-numbers
["1..", None],
["..", None],
[".1.", None],
# should match after comma
[",1", "1"],
[", 1", "1"],
[", .1", ".1"],
[", +.1", "+.1"],
# should not match after trailing spaces
[".1 ", None],
# some complex cases
["0b_0011_1111_0100_1110", "0b_0011_1111_0100_1110"],
["0xdeadbeef", "0xdeadbeef"],
["0b_1110_0101", "0b_1110_0101"],
# should not match if in an operation
["1 + 1", None],
[", 1 + 1", None],
],
)
def test_match_numeric_literal_for_dict_key(input, expected):
assert _match_number_in_dict_key_prefix(input) == expected