Show More
@@ -198,9 +198,9 b' def fastannotate(ui, repo, *pats, **opts' | |||||
198 | formatter.write(result, lines, existinglines=existinglines) |
|
198 | formatter.write(result, lines, existinglines=existinglines) | |
199 | formatter.end() |
|
199 | formatter.end() | |
200 |
|
200 | |||
201 |
_newopts = set( |
|
201 | _newopts = set() | |
202 |
_knownopts = |
|
202 | _knownopts = {opt[1].replace('-', '_') for opt in | |
203 |
|
|
203 | (fastannotatecommandargs[r'options'] + commands.globalopts)} | |
204 |
|
204 | |||
205 | def _annotatewrapper(orig, ui, repo, *pats, **opts): |
|
205 | def _annotatewrapper(orig, ui, repo, *pats, **opts): | |
206 | """used by wrapdefault""" |
|
206 | """used by wrapdefault""" |
@@ -136,7 +136,7 b' class basestore(object):' | |||||
136 | failed = self._verifyfiles(contents, filestocheck) |
|
136 | failed = self._verifyfiles(contents, filestocheck) | |
137 |
|
137 | |||
138 | numrevs = len(verified) |
|
138 | numrevs = len(verified) | |
139 |
numlfiles = len( |
|
139 | numlfiles = len({fname for (fname, fnode) in verified}) | |
140 | if contents: |
|
140 | if contents: | |
141 | self.ui.status( |
|
141 | self.ui.status( | |
142 | _('verified contents of %d revisions of %d largefiles\n') |
|
142 | _('verified contents of %d revisions of %d largefiles\n') |
@@ -1940,7 +1940,7 b' class queue(object):' | |||||
1940 | self.ui.write(patchname, label='qseries.' + state) |
|
1940 | self.ui.write(patchname, label='qseries.' + state) | |
1941 | self.ui.write('\n') |
|
1941 | self.ui.write('\n') | |
1942 |
|
1942 | |||
1943 |
applied = |
|
1943 | applied = {p.name for p in self.applied} | |
1944 | if length is None: |
|
1944 | if length is None: | |
1945 | length = len(self.series) - start |
|
1945 | length = len(self.series) - start | |
1946 | if not missing: |
|
1946 | if not missing: | |
@@ -3658,7 +3658,7 b' def revsetmq(repo, subset, x):' | |||||
3658 | """Changesets managed by MQ. |
|
3658 | """Changesets managed by MQ. | |
3659 | """ |
|
3659 | """ | |
3660 | revsetlang.getargs(x, 0, 0, _("mq takes no arguments")) |
|
3660 | revsetlang.getargs(x, 0, 0, _("mq takes no arguments")) | |
3661 |
applied = |
|
3661 | applied = {repo[r.node].rev() for r in repo.mq.applied} | |
3662 | return smartset.baseset([r for r in subset if r in applied]) |
|
3662 | return smartset.baseset([r for r in subset if r in applied]) | |
3663 |
|
3663 | |||
3664 | # tell hggettext to extract docstrings from these functions: |
|
3664 | # tell hggettext to extract docstrings from these functions: |
@@ -1878,8 +1878,8 b' def _computeobsoletenotrebased(repo, reb' | |||||
1878 | obsolete successors. |
|
1878 | obsolete successors. | |
1879 | """ |
|
1879 | """ | |
1880 | obsoletenotrebased = {} |
|
1880 | obsoletenotrebased = {} | |
1881 |
obsoletewithoutsuccessorindestination = set( |
|
1881 | obsoletewithoutsuccessorindestination = set() | |
1882 |
obsoleteextinctsuccessors = set( |
|
1882 | obsoleteextinctsuccessors = set() | |
1883 |
|
1883 | |||
1884 | assert repo.filtername is None |
|
1884 | assert repo.filtername is None | |
1885 | cl = repo.changelog |
|
1885 | cl = repo.changelog |
@@ -804,7 +804,7 b' def gcclient(ui, cachepath):' | |||||
804 | return |
|
804 | return | |
805 |
|
805 | |||
806 | reposfile = open(repospath, 'rb') |
|
806 | reposfile = open(repospath, 'rb') | |
807 |
repos = |
|
807 | repos = {r[:-1] for r in reposfile.readlines()} | |
808 | reposfile.close() |
|
808 | reposfile.close() | |
809 |
|
809 | |||
810 | # build list of useful files |
|
810 | # build list of useful files |
@@ -154,9 +154,9 b' def _deletebigpacks(repo, folder, files)' | |||||
154 |
|
154 | |||
155 | # Either an oversize index or datapack will trigger cleanup of the whole |
|
155 | # Either an oversize index or datapack will trigger cleanup of the whole | |
156 | # pack: |
|
156 | # pack: | |
157 |
oversized = |
|
157 | oversized = {os.path.splitext(path)[0] for path, ftype, stat in files | |
158 | if (stat.st_size > maxsize and (os.path.splitext(path)[1] |
|
158 | if (stat.st_size > maxsize and (os.path.splitext(path)[1] | |
159 |
in VALIDEXTS)) |
|
159 | in VALIDEXTS))} | |
160 |
|
160 | |||
161 | for rootfname in oversized: |
|
161 | for rootfname in oversized: | |
162 | rootpath = os.path.join(folder, rootfname) |
|
162 | rootpath = os.path.join(folder, rootfname) |
@@ -243,7 +243,7 b' def showstack(ui, repo, displayer):' | |||||
243 | else: |
|
243 | else: | |
244 | newheads = set() |
|
244 | newheads = set() | |
245 |
|
245 | |||
246 |
allrevs = set(stackrevs) | newheads | |
|
246 | allrevs = set(stackrevs) | newheads | {baserev} | |
247 | nodelen = longestshortest(repo, allrevs) |
|
247 | nodelen = longestshortest(repo, allrevs) | |
248 |
|
248 | |||
249 | try: |
|
249 | try: |
@@ -147,7 +147,7 b' def uncommit(ui, repo, *pats, **opts):' | |||||
147 | # if not everything tracked in that directory can be |
|
147 | # if not everything tracked in that directory can be | |
148 | # uncommitted. |
|
148 | # uncommitted. | |
149 | if badfiles: |
|
149 | if badfiles: | |
150 |
badfiles -= |
|
150 | badfiles -= {f for f in util.dirs(eligible)} | |
151 |
|
151 | |||
152 | for f in sorted(badfiles): |
|
152 | for f in sorted(badfiles): | |
153 | if f in s.clean: |
|
153 | if f in s.clean: |
@@ -2315,7 +2315,7 b' def widen_bundle(repo, oldmatcher, newma' | |||||
2315 | oldmatcher=oldmatcher, |
|
2315 | oldmatcher=oldmatcher, | |
2316 | matcher=newmatcher, |
|
2316 | matcher=newmatcher, | |
2317 | fullnodes=commonnodes) |
|
2317 | fullnodes=commonnodes) | |
2318 |
cgdata = packer.generate( |
|
2318 | cgdata = packer.generate({nodemod.nullid}, list(commonnodes), | |
2319 | False, 'narrow_widen', changelog=False) |
|
2319 | False, 'narrow_widen', changelog=False) | |
2320 |
|
2320 | |||
2321 | part = bundler.newpart('changegroup', data=cgdata) |
|
2321 | part = bundler.newpart('changegroup', data=cgdata) |
@@ -458,7 +458,7 b' class dirnode(object):' | |||||
458 |
|
458 | |||
459 | def __init__(self, dirpath): |
|
459 | def __init__(self, dirpath): | |
460 | self.path = dirpath |
|
460 | self.path = dirpath | |
461 |
self.statuses = set( |
|
461 | self.statuses = set() | |
462 | self.files = [] |
|
462 | self.files = [] | |
463 | self.subdirs = {} |
|
463 | self.subdirs = {} | |
464 |
|
464 | |||
@@ -2495,8 +2495,7 b' def amend(ui, repo, old, extra, pats, op' | |||||
2495 | if len(old.parents()) > 1: |
|
2495 | if len(old.parents()) > 1: | |
2496 | # ctx.files() isn't reliable for merges, so fall back to the |
|
2496 | # ctx.files() isn't reliable for merges, so fall back to the | |
2497 | # slower repo.status() method |
|
2497 | # slower repo.status() method | |
2498 |
files = |
|
2498 | files = {fn for st in base.status(old)[:3] for fn in st} | |
2499 | for fn in st]) |
|
|||
2500 | else: |
|
2499 | else: | |
2501 | files = set(old.files()) |
|
2500 | files = set(old.files()) | |
2502 |
|
2501 |
@@ -764,7 +764,7 b' def headrevs(revs, parentsfn):' | |||||
764 | the input set. |
|
764 | the input set. | |
765 | """ |
|
765 | """ | |
766 | headrevs = set(revs) |
|
766 | headrevs = set(revs) | |
767 |
parents = |
|
767 | parents = {node.nullrev} | |
768 | up = parents.update |
|
768 | up = parents.update | |
769 |
|
769 | |||
770 | for rev in revs: |
|
770 | for rev in revs: |
@@ -707,8 +707,8 b' def _pushdiscoverybookmarks(pushop):' | |||||
707 |
|
707 | |||
708 | remotebookmark = listkeys(remote, 'bookmarks') |
|
708 | remotebookmark = listkeys(remote, 'bookmarks') | |
709 |
|
709 | |||
710 |
explicit = |
|
710 | explicit = {repo._bookmarks.expandname(bookmark) | |
711 |
|
|
711 | for bookmark in pushop.bookmarks} | |
712 |
|
712 | |||
713 | remotebookmark = bookmod.unhexlifybookmarks(remotebookmark) |
|
713 | remotebookmark = bookmod.unhexlifybookmarks(remotebookmark) | |
714 | comp = bookmod.comparebookmarks(repo, repo._bookmarks, remotebookmark) |
|
714 | comp = bookmod.comparebookmarks(repo, repo._bookmarks, remotebookmark) |
@@ -34,7 +34,7 b' def bisect(repo, state):' | |||||
34 |
|
34 | |||
35 | changelog = repo.changelog |
|
35 | changelog = repo.changelog | |
36 | clparents = changelog.parentrevs |
|
36 | clparents = changelog.parentrevs | |
37 |
skip = |
|
37 | skip = {changelog.rev(n) for n in state['skip']} | |
38 |
|
38 | |||
39 | def buildancestors(bad, good): |
|
39 | def buildancestors(bad, good): | |
40 | badrev = min([changelog.rev(n) for n in bad]) |
|
40 | badrev = min([changelog.rev(n) for n in bad]) |
@@ -748,7 +748,7 b' def help_(ui, commands, name, unknowncmd' | |||||
748 | ct = mod.cmdtable |
|
748 | ct = mod.cmdtable | |
749 | except AttributeError: |
|
749 | except AttributeError: | |
750 | ct = {} |
|
750 | ct = {} | |
751 |
modcmds = |
|
751 | modcmds = {c.partition('|')[0] for c in ct} | |
752 | rst.extend(helplist(modcmds.__contains__)) |
|
752 | rst.extend(helplist(modcmds.__contains__)) | |
753 | else: |
|
753 | else: | |
754 | rst.append(_("(use 'hg help extensions' for information on enabling" |
|
754 | rst.append(_("(use 'hg help extensions' for information on enabling" |
@@ -815,8 +815,8 b' def _checkunknownfiles(repo, wctx, mctx,' | |||||
815 | fileconflicts.add(f) |
|
815 | fileconflicts.add(f) | |
816 |
|
816 | |||
817 | allconflicts = fileconflicts | pathconflicts |
|
817 | allconflicts = fileconflicts | pathconflicts | |
818 |
ignoredconflicts = |
|
818 | ignoredconflicts = {c for c in allconflicts | |
819 |
|
|
819 | if repo.dirstate._ignore(c)} | |
820 | unknownconflicts = allconflicts - ignoredconflicts |
|
820 | unknownconflicts = allconflicts - ignoredconflicts | |
821 | collectconflicts(ignoredconflicts, ignoredconfig) |
|
821 | collectconflicts(ignoredconflicts, ignoredconfig) | |
822 | collectconflicts(unknownconflicts, unknownconfig) |
|
822 | collectconflicts(unknownconflicts, unknownconfig) | |
@@ -1104,7 +1104,7 b' def _filternarrowactions(narrowmatch, br' | |||||
1104 | Raise an exception if the merge cannot be completed because the repo is |
|
1104 | Raise an exception if the merge cannot be completed because the repo is | |
1105 | narrowed. |
|
1105 | narrowed. | |
1106 | """ |
|
1106 | """ | |
1107 |
nooptypes = |
|
1107 | nooptypes = {'k'} # TODO: handle with nonconflicttypes | |
1108 | nonconflicttypes = set('a am c cm f g r e'.split()) |
|
1108 | nonconflicttypes = set('a am c cm f g r e'.split()) | |
1109 | # We mutate the items in the dict during iteration, so iterate |
|
1109 | # We mutate the items in the dict during iteration, so iterate | |
1110 | # over a copy. |
|
1110 | # over a copy. |
@@ -743,7 +743,7 b' class obsstore(object):' | |||||
743 | pruned = [m for m in succsmarkers.get(current, ()) if not m[1]] |
|
743 | pruned = [m for m in succsmarkers.get(current, ()) if not m[1]] | |
744 | direct.update(pruned) |
|
744 | direct.update(pruned) | |
745 | direct -= seenmarkers |
|
745 | direct -= seenmarkers | |
746 |
pendingnodes = |
|
746 | pendingnodes = {m[0] for m in direct} | |
747 | seenmarkers |= direct |
|
747 | seenmarkers |= direct | |
748 | pendingnodes -= seennodes |
|
748 | pendingnodes -= seennodes | |
749 | seennodes |= pendingnodes |
|
749 | seennodes |= pendingnodes |
@@ -637,7 +637,7 b' def _candidategroups(revlog, textlen, p1' | |||||
637 |
|
637 | |||
638 | deltas_limit = textlen * LIMIT_DELTA2TEXT |
|
638 | deltas_limit = textlen * LIMIT_DELTA2TEXT | |
639 |
|
639 | |||
640 |
tested = |
|
640 | tested = {nullrev} | |
641 | candidates = _refinedgroups(revlog, p1, p2, cachedelta) |
|
641 | candidates = _refinedgroups(revlog, p1, p2, cachedelta) | |
642 | while True: |
|
642 | while True: | |
643 | temptative = candidates.send(good) |
|
643 | temptative = candidates.send(good) |
@@ -345,7 +345,7 b' def find_pullbundle(repo, proto, opts, c' | |||||
345 | one specific branch of many. |
|
345 | one specific branch of many. | |
346 | """ |
|
346 | """ | |
347 | def decodehexstring(s): |
|
347 | def decodehexstring(s): | |
348 |
return |
|
348 | return {binascii.unhexlify(h) for h in s.split(';')} | |
349 |
|
349 | |||
350 | manifest = repo.vfs.tryread('pullbundles.manifest') |
|
350 | manifest = repo.vfs.tryread('pullbundles.manifest') | |
351 | if not manifest: |
|
351 | if not manifest: |
General Comments 0
You need to be logged in to leave comments.
Login now