##// END OF EJS Templates
revset: point out wrong behavior in fullreposet...
Pierre-Yves David -
r25547:99a1f73a default
parent child Browse files
Show More
@@ -1,3562 +1,3573 b''
1 1 # revset.py - revision set queries for mercurial
2 2 #
3 3 # Copyright 2010 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 import re
9 9 import parser, util, error, hbisect, phases
10 10 import node
11 11 import heapq
12 12 import match as matchmod
13 13 from i18n import _
14 14 import encoding
15 15 import obsolete as obsmod
16 16 import pathutil
17 17 import repoview
18 18
19 19 def _revancestors(repo, revs, followfirst):
20 20 """Like revlog.ancestors(), but supports followfirst."""
21 21 if followfirst:
22 22 cut = 1
23 23 else:
24 24 cut = None
25 25 cl = repo.changelog
26 26
27 27 def iterate():
28 28 revs.sort(reverse=True)
29 29 irevs = iter(revs)
30 30 h = []
31 31
32 32 inputrev = next(irevs, None)
33 33 if inputrev is not None:
34 34 heapq.heappush(h, -inputrev)
35 35
36 36 seen = set()
37 37 while h:
38 38 current = -heapq.heappop(h)
39 39 if current == inputrev:
40 40 inputrev = next(irevs, None)
41 41 if inputrev is not None:
42 42 heapq.heappush(h, -inputrev)
43 43 if current not in seen:
44 44 seen.add(current)
45 45 yield current
46 46 for parent in cl.parentrevs(current)[:cut]:
47 47 if parent != node.nullrev:
48 48 heapq.heappush(h, -parent)
49 49
50 50 return generatorset(iterate(), iterasc=False)
51 51
52 52 def _revdescendants(repo, revs, followfirst):
53 53 """Like revlog.descendants() but supports followfirst."""
54 54 if followfirst:
55 55 cut = 1
56 56 else:
57 57 cut = None
58 58
59 59 def iterate():
60 60 cl = repo.changelog
61 61 first = min(revs)
62 62 nullrev = node.nullrev
63 63 if first == nullrev:
64 64 # Are there nodes with a null first parent and a non-null
65 65 # second one? Maybe. Do we care? Probably not.
66 66 for i in cl:
67 67 yield i
68 68 else:
69 69 seen = set(revs)
70 70 for i in cl.revs(first + 1):
71 71 for x in cl.parentrevs(i)[:cut]:
72 72 if x != nullrev and x in seen:
73 73 seen.add(i)
74 74 yield i
75 75 break
76 76
77 77 return generatorset(iterate(), iterasc=True)
78 78
79 79 def _revsbetween(repo, roots, heads):
80 80 """Return all paths between roots and heads, inclusive of both endpoint
81 81 sets."""
82 82 if not roots:
83 83 return baseset()
84 84 parentrevs = repo.changelog.parentrevs
85 85 visit = list(heads)
86 86 reachable = set()
87 87 seen = {}
88 88 minroot = min(roots)
89 89 roots = set(roots)
90 90 # open-code the post-order traversal due to the tiny size of
91 91 # sys.getrecursionlimit()
92 92 while visit:
93 93 rev = visit.pop()
94 94 if rev in roots:
95 95 reachable.add(rev)
96 96 parents = parentrevs(rev)
97 97 seen[rev] = parents
98 98 for parent in parents:
99 99 if parent >= minroot and parent not in seen:
100 100 visit.append(parent)
101 101 if not reachable:
102 102 return baseset()
103 103 for rev in sorted(seen):
104 104 for parent in seen[rev]:
105 105 if parent in reachable:
106 106 reachable.add(rev)
107 107 return baseset(sorted(reachable))
108 108
109 109 elements = {
110 110 "(": (21, ("group", 1, ")"), ("func", 1, ")")),
111 111 "##": (20, None, ("_concat", 20)),
112 112 "~": (18, None, ("ancestor", 18)),
113 113 "^": (18, None, ("parent", 18), ("parentpost", 18)),
114 114 "-": (5, ("negate", 19), ("minus", 5)),
115 115 "::": (17, ("dagrangepre", 17), ("dagrange", 17),
116 116 ("dagrangepost", 17)),
117 117 "..": (17, ("dagrangepre", 17), ("dagrange", 17),
118 118 ("dagrangepost", 17)),
119 119 ":": (15, ("rangepre", 15), ("range", 15), ("rangepost", 15)),
120 120 "not": (10, ("not", 10)),
121 121 "!": (10, ("not", 10)),
122 122 "and": (5, None, ("and", 5)),
123 123 "&": (5, None, ("and", 5)),
124 124 "%": (5, None, ("only", 5), ("onlypost", 5)),
125 125 "or": (4, None, ("or", 4)),
126 126 "|": (4, None, ("or", 4)),
127 127 "+": (4, None, ("or", 4)),
128 128 ",": (2, None, ("list", 2)),
129 129 ")": (0, None, None),
130 130 "symbol": (0, ("symbol",), None),
131 131 "string": (0, ("string",), None),
132 132 "end": (0, None, None),
133 133 }
134 134
135 135 keywords = set(['and', 'or', 'not'])
136 136
137 137 # default set of valid characters for the initial letter of symbols
138 138 _syminitletters = set(c for c in [chr(i) for i in xrange(256)]
139 139 if c.isalnum() or c in '._@' or ord(c) > 127)
140 140
141 141 # default set of valid characters for non-initial letters of symbols
142 142 _symletters = set(c for c in [chr(i) for i in xrange(256)]
143 143 if c.isalnum() or c in '-._/@' or ord(c) > 127)
144 144
145 145 def tokenize(program, lookup=None, syminitletters=None, symletters=None):
146 146 '''
147 147 Parse a revset statement into a stream of tokens
148 148
149 149 ``syminitletters`` is the set of valid characters for the initial
150 150 letter of symbols.
151 151
152 152 By default, character ``c`` is recognized as valid for initial
153 153 letter of symbols, if ``c.isalnum() or c in '._@' or ord(c) > 127``.
154 154
155 155 ``symletters`` is the set of valid characters for non-initial
156 156 letters of symbols.
157 157
158 158 By default, character ``c`` is recognized as valid for non-initial
159 159 letters of symbols, if ``c.isalnum() or c in '-._/@' or ord(c) > 127``.
160 160
161 161 Check that @ is a valid unquoted token character (issue3686):
162 162 >>> list(tokenize("@::"))
163 163 [('symbol', '@', 0), ('::', None, 1), ('end', None, 3)]
164 164
165 165 '''
166 166 if syminitletters is None:
167 167 syminitletters = _syminitletters
168 168 if symletters is None:
169 169 symletters = _symletters
170 170
171 171 pos, l = 0, len(program)
172 172 while pos < l:
173 173 c = program[pos]
174 174 if c.isspace(): # skip inter-token whitespace
175 175 pass
176 176 elif c == ':' and program[pos:pos + 2] == '::': # look ahead carefully
177 177 yield ('::', None, pos)
178 178 pos += 1 # skip ahead
179 179 elif c == '.' and program[pos:pos + 2] == '..': # look ahead carefully
180 180 yield ('..', None, pos)
181 181 pos += 1 # skip ahead
182 182 elif c == '#' and program[pos:pos + 2] == '##': # look ahead carefully
183 183 yield ('##', None, pos)
184 184 pos += 1 # skip ahead
185 185 elif c in "():,-|&+!~^%": # handle simple operators
186 186 yield (c, None, pos)
187 187 elif (c in '"\'' or c == 'r' and
188 188 program[pos:pos + 2] in ("r'", 'r"')): # handle quoted strings
189 189 if c == 'r':
190 190 pos += 1
191 191 c = program[pos]
192 192 decode = lambda x: x
193 193 else:
194 194 decode = lambda x: x.decode('string-escape')
195 195 pos += 1
196 196 s = pos
197 197 while pos < l: # find closing quote
198 198 d = program[pos]
199 199 if d == '\\': # skip over escaped characters
200 200 pos += 2
201 201 continue
202 202 if d == c:
203 203 yield ('string', decode(program[s:pos]), s)
204 204 break
205 205 pos += 1
206 206 else:
207 207 raise error.ParseError(_("unterminated string"), s)
208 208 # gather up a symbol/keyword
209 209 elif c in syminitletters:
210 210 s = pos
211 211 pos += 1
212 212 while pos < l: # find end of symbol
213 213 d = program[pos]
214 214 if d not in symletters:
215 215 break
216 216 if d == '.' and program[pos - 1] == '.': # special case for ..
217 217 pos -= 1
218 218 break
219 219 pos += 1
220 220 sym = program[s:pos]
221 221 if sym in keywords: # operator keywords
222 222 yield (sym, None, s)
223 223 elif '-' in sym:
224 224 # some jerk gave us foo-bar-baz, try to check if it's a symbol
225 225 if lookup and lookup(sym):
226 226 # looks like a real symbol
227 227 yield ('symbol', sym, s)
228 228 else:
229 229 # looks like an expression
230 230 parts = sym.split('-')
231 231 for p in parts[:-1]:
232 232 if p: # possible consecutive -
233 233 yield ('symbol', p, s)
234 234 s += len(p)
235 235 yield ('-', None, pos)
236 236 s += 1
237 237 if parts[-1]: # possible trailing -
238 238 yield ('symbol', parts[-1], s)
239 239 else:
240 240 yield ('symbol', sym, s)
241 241 pos -= 1
242 242 else:
243 243 raise error.ParseError(_("syntax error in revset '%s'") %
244 244 program, pos)
245 245 pos += 1
246 246 yield ('end', None, pos)
247 247
248 248 def parseerrordetail(inst):
249 249 """Compose error message from specified ParseError object
250 250 """
251 251 if len(inst.args) > 1:
252 252 return _('at %s: %s') % (inst.args[1], inst.args[0])
253 253 else:
254 254 return inst.args[0]
255 255
256 256 # helpers
257 257
258 258 def getstring(x, err):
259 259 if x and (x[0] == 'string' or x[0] == 'symbol'):
260 260 return x[1]
261 261 raise error.ParseError(err)
262 262
263 263 def getlist(x):
264 264 if not x:
265 265 return []
266 266 if x[0] == 'list':
267 267 return getlist(x[1]) + [x[2]]
268 268 return [x]
269 269
270 270 def getargs(x, min, max, err):
271 271 l = getlist(x)
272 272 if len(l) < min or (max >= 0 and len(l) > max):
273 273 raise error.ParseError(err)
274 274 return l
275 275
276 276 def isvalidsymbol(tree):
277 277 """Examine whether specified ``tree`` is valid ``symbol`` or not
278 278 """
279 279 return tree[0] == 'symbol' and len(tree) > 1
280 280
281 281 def getsymbol(tree):
282 282 """Get symbol name from valid ``symbol`` in ``tree``
283 283
284 284 This assumes that ``tree`` is already examined by ``isvalidsymbol``.
285 285 """
286 286 return tree[1]
287 287
288 288 def isvalidfunc(tree):
289 289 """Examine whether specified ``tree`` is valid ``func`` or not
290 290 """
291 291 return tree[0] == 'func' and len(tree) > 1 and isvalidsymbol(tree[1])
292 292
293 293 def getfuncname(tree):
294 294 """Get function name from valid ``func`` in ``tree``
295 295
296 296 This assumes that ``tree`` is already examined by ``isvalidfunc``.
297 297 """
298 298 return getsymbol(tree[1])
299 299
300 300 def getfuncargs(tree):
301 301 """Get list of function arguments from valid ``func`` in ``tree``
302 302
303 303 This assumes that ``tree`` is already examined by ``isvalidfunc``.
304 304 """
305 305 if len(tree) > 2:
306 306 return getlist(tree[2])
307 307 else:
308 308 return []
309 309
310 310 def getset(repo, subset, x):
311 311 if not x:
312 312 raise error.ParseError(_("missing argument"))
313 313 s = methods[x[0]](repo, subset, *x[1:])
314 314 if util.safehasattr(s, 'isascending'):
315 315 return s
316 316 return baseset(s)
317 317
318 318 def _getrevsource(repo, r):
319 319 extra = repo[r].extra()
320 320 for label in ('source', 'transplant_source', 'rebase_source'):
321 321 if label in extra:
322 322 try:
323 323 return repo[extra[label]].rev()
324 324 except error.RepoLookupError:
325 325 pass
326 326 return None
327 327
328 328 # operator methods
329 329
330 330 def stringset(repo, subset, x):
331 331 x = repo[x].rev()
332 332 if (x in subset
333 333 or x == node.nullrev and isinstance(subset, fullreposet)):
334 334 return baseset([x])
335 335 return baseset()
336 336
337 337 def rangeset(repo, subset, x, y):
338 338 m = getset(repo, fullreposet(repo), x)
339 339 n = getset(repo, fullreposet(repo), y)
340 340
341 341 if not m or not n:
342 342 return baseset()
343 343 m, n = m.first(), n.last()
344 344
345 345 if m < n:
346 346 r = spanset(repo, m, n + 1)
347 347 else:
348 348 r = spanset(repo, m, n - 1)
349 349 return r & subset
350 350
351 351 def dagrange(repo, subset, x, y):
352 352 r = fullreposet(repo)
353 353 xs = _revsbetween(repo, getset(repo, r, x), getset(repo, r, y))
354 354 return xs & subset
355 355
356 356 def andset(repo, subset, x, y):
357 357 return getset(repo, getset(repo, subset, x), y)
358 358
359 359 def orset(repo, subset, *xs):
360 360 rs = [getset(repo, subset, x) for x in xs]
361 361 return _combinesets(rs)
362 362
363 363 def notset(repo, subset, x):
364 364 return subset - getset(repo, subset, x)
365 365
366 366 def listset(repo, subset, a, b):
367 367 raise error.ParseError(_("can't use a list in this context"))
368 368
369 369 def func(repo, subset, a, b):
370 370 if a[0] == 'symbol' and a[1] in symbols:
371 371 return symbols[a[1]](repo, subset, b)
372 372 raise error.UnknownIdentifier(a[1], symbols.keys())
373 373
374 374 # functions
375 375
376 376 def adds(repo, subset, x):
377 377 """``adds(pattern)``
378 378 Changesets that add a file matching pattern.
379 379
380 380 The pattern without explicit kind like ``glob:`` is expected to be
381 381 relative to the current directory and match against a file or a
382 382 directory.
383 383 """
384 384 # i18n: "adds" is a keyword
385 385 pat = getstring(x, _("adds requires a pattern"))
386 386 return checkstatus(repo, subset, pat, 1)
387 387
388 388 def ancestor(repo, subset, x):
389 389 """``ancestor(*changeset)``
390 390 A greatest common ancestor of the changesets.
391 391
392 392 Accepts 0 or more changesets.
393 393 Will return empty list when passed no args.
394 394 Greatest common ancestor of a single changeset is that changeset.
395 395 """
396 396 # i18n: "ancestor" is a keyword
397 397 l = getlist(x)
398 398 rl = fullreposet(repo)
399 399 anc = None
400 400
401 401 # (getset(repo, rl, i) for i in l) generates a list of lists
402 402 for revs in (getset(repo, rl, i) for i in l):
403 403 for r in revs:
404 404 if anc is None:
405 405 anc = repo[r]
406 406 else:
407 407 anc = anc.ancestor(repo[r])
408 408
409 409 if anc is not None and anc.rev() in subset:
410 410 return baseset([anc.rev()])
411 411 return baseset()
412 412
413 413 def _ancestors(repo, subset, x, followfirst=False):
414 414 heads = getset(repo, fullreposet(repo), x)
415 415 if not heads:
416 416 return baseset()
417 417 s = _revancestors(repo, heads, followfirst)
418 418 return subset & s
419 419
420 420 def ancestors(repo, subset, x):
421 421 """``ancestors(set)``
422 422 Changesets that are ancestors of a changeset in set.
423 423 """
424 424 return _ancestors(repo, subset, x)
425 425
426 426 def _firstancestors(repo, subset, x):
427 427 # ``_firstancestors(set)``
428 428 # Like ``ancestors(set)`` but follows only the first parents.
429 429 return _ancestors(repo, subset, x, followfirst=True)
430 430
431 431 def ancestorspec(repo, subset, x, n):
432 432 """``set~n``
433 433 Changesets that are the Nth ancestor (first parents only) of a changeset
434 434 in set.
435 435 """
436 436 try:
437 437 n = int(n[1])
438 438 except (TypeError, ValueError):
439 439 raise error.ParseError(_("~ expects a number"))
440 440 ps = set()
441 441 cl = repo.changelog
442 442 for r in getset(repo, fullreposet(repo), x):
443 443 for i in range(n):
444 444 r = cl.parentrevs(r)[0]
445 445 ps.add(r)
446 446 return subset & ps
447 447
448 448 def author(repo, subset, x):
449 449 """``author(string)``
450 450 Alias for ``user(string)``.
451 451 """
452 452 # i18n: "author" is a keyword
453 453 n = encoding.lower(getstring(x, _("author requires a string")))
454 454 kind, pattern, matcher = _substringmatcher(n)
455 455 return subset.filter(lambda x: matcher(encoding.lower(repo[x].user())))
456 456
457 457 def bisect(repo, subset, x):
458 458 """``bisect(string)``
459 459 Changesets marked in the specified bisect status:
460 460
461 461 - ``good``, ``bad``, ``skip``: csets explicitly marked as good/bad/skip
462 462 - ``goods``, ``bads`` : csets topologically good/bad
463 463 - ``range`` : csets taking part in the bisection
464 464 - ``pruned`` : csets that are goods, bads or skipped
465 465 - ``untested`` : csets whose fate is yet unknown
466 466 - ``ignored`` : csets ignored due to DAG topology
467 467 - ``current`` : the cset currently being bisected
468 468 """
469 469 # i18n: "bisect" is a keyword
470 470 status = getstring(x, _("bisect requires a string")).lower()
471 471 state = set(hbisect.get(repo, status))
472 472 return subset & state
473 473
474 474 # Backward-compatibility
475 475 # - no help entry so that we do not advertise it any more
476 476 def bisected(repo, subset, x):
477 477 return bisect(repo, subset, x)
478 478
479 479 def bookmark(repo, subset, x):
480 480 """``bookmark([name])``
481 481 The named bookmark or all bookmarks.
482 482
483 483 If `name` starts with `re:`, the remainder of the name is treated as
484 484 a regular expression. To match a bookmark that actually starts with `re:`,
485 485 use the prefix `literal:`.
486 486 """
487 487 # i18n: "bookmark" is a keyword
488 488 args = getargs(x, 0, 1, _('bookmark takes one or no arguments'))
489 489 if args:
490 490 bm = getstring(args[0],
491 491 # i18n: "bookmark" is a keyword
492 492 _('the argument to bookmark must be a string'))
493 493 kind, pattern, matcher = _stringmatcher(bm)
494 494 bms = set()
495 495 if kind == 'literal':
496 496 bmrev = repo._bookmarks.get(pattern, None)
497 497 if not bmrev:
498 498 raise error.RepoLookupError(_("bookmark '%s' does not exist")
499 499 % bm)
500 500 bms.add(repo[bmrev].rev())
501 501 else:
502 502 matchrevs = set()
503 503 for name, bmrev in repo._bookmarks.iteritems():
504 504 if matcher(name):
505 505 matchrevs.add(bmrev)
506 506 if not matchrevs:
507 507 raise error.RepoLookupError(_("no bookmarks exist"
508 508 " that match '%s'") % pattern)
509 509 for bmrev in matchrevs:
510 510 bms.add(repo[bmrev].rev())
511 511 else:
512 512 bms = set([repo[r].rev()
513 513 for r in repo._bookmarks.values()])
514 514 bms -= set([node.nullrev])
515 515 return subset & bms
516 516
517 517 def branch(repo, subset, x):
518 518 """``branch(string or set)``
519 519 All changesets belonging to the given branch or the branches of the given
520 520 changesets.
521 521
522 522 If `string` starts with `re:`, the remainder of the name is treated as
523 523 a regular expression. To match a branch that actually starts with `re:`,
524 524 use the prefix `literal:`.
525 525 """
526 526 getbi = repo.revbranchcache().branchinfo
527 527
528 528 try:
529 529 b = getstring(x, '')
530 530 except error.ParseError:
531 531 # not a string, but another revspec, e.g. tip()
532 532 pass
533 533 else:
534 534 kind, pattern, matcher = _stringmatcher(b)
535 535 if kind == 'literal':
536 536 # note: falls through to the revspec case if no branch with
537 537 # this name exists
538 538 if pattern in repo.branchmap():
539 539 return subset.filter(lambda r: matcher(getbi(r)[0]))
540 540 else:
541 541 return subset.filter(lambda r: matcher(getbi(r)[0]))
542 542
543 543 s = getset(repo, fullreposet(repo), x)
544 544 b = set()
545 545 for r in s:
546 546 b.add(getbi(r)[0])
547 547 c = s.__contains__
548 548 return subset.filter(lambda r: c(r) or getbi(r)[0] in b)
549 549
550 550 def bumped(repo, subset, x):
551 551 """``bumped()``
552 552 Mutable changesets marked as successors of public changesets.
553 553
554 554 Only non-public and non-obsolete changesets can be `bumped`.
555 555 """
556 556 # i18n: "bumped" is a keyword
557 557 getargs(x, 0, 0, _("bumped takes no arguments"))
558 558 bumped = obsmod.getrevs(repo, 'bumped')
559 559 return subset & bumped
560 560
561 561 def bundle(repo, subset, x):
562 562 """``bundle()``
563 563 Changesets in the bundle.
564 564
565 565 Bundle must be specified by the -R option."""
566 566
567 567 try:
568 568 bundlerevs = repo.changelog.bundlerevs
569 569 except AttributeError:
570 570 raise util.Abort(_("no bundle provided - specify with -R"))
571 571 return subset & bundlerevs
572 572
573 573 def checkstatus(repo, subset, pat, field):
574 574 hasset = matchmod.patkind(pat) == 'set'
575 575
576 576 mcache = [None]
577 577 def matches(x):
578 578 c = repo[x]
579 579 if not mcache[0] or hasset:
580 580 mcache[0] = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
581 581 m = mcache[0]
582 582 fname = None
583 583 if not m.anypats() and len(m.files()) == 1:
584 584 fname = m.files()[0]
585 585 if fname is not None:
586 586 if fname not in c.files():
587 587 return False
588 588 else:
589 589 for f in c.files():
590 590 if m(f):
591 591 break
592 592 else:
593 593 return False
594 594 files = repo.status(c.p1().node(), c.node())[field]
595 595 if fname is not None:
596 596 if fname in files:
597 597 return True
598 598 else:
599 599 for f in files:
600 600 if m(f):
601 601 return True
602 602
603 603 return subset.filter(matches)
604 604
605 605 def _children(repo, narrow, parentset):
606 606 cs = set()
607 607 if not parentset:
608 608 return baseset(cs)
609 609 pr = repo.changelog.parentrevs
610 610 minrev = min(parentset)
611 611 for r in narrow:
612 612 if r <= minrev:
613 613 continue
614 614 for p in pr(r):
615 615 if p in parentset:
616 616 cs.add(r)
617 617 return baseset(cs)
618 618
619 619 def children(repo, subset, x):
620 620 """``children(set)``
621 621 Child changesets of changesets in set.
622 622 """
623 623 s = getset(repo, fullreposet(repo), x)
624 624 cs = _children(repo, subset, s)
625 625 return subset & cs
626 626
627 627 def closed(repo, subset, x):
628 628 """``closed()``
629 629 Changeset is closed.
630 630 """
631 631 # i18n: "closed" is a keyword
632 632 getargs(x, 0, 0, _("closed takes no arguments"))
633 633 return subset.filter(lambda r: repo[r].closesbranch())
634 634
635 635 def contains(repo, subset, x):
636 636 """``contains(pattern)``
637 637 The revision's manifest contains a file matching pattern (but might not
638 638 modify it). See :hg:`help patterns` for information about file patterns.
639 639
640 640 The pattern without explicit kind like ``glob:`` is expected to be
641 641 relative to the current directory and match against a file exactly
642 642 for efficiency.
643 643 """
644 644 # i18n: "contains" is a keyword
645 645 pat = getstring(x, _("contains requires a pattern"))
646 646
647 647 def matches(x):
648 648 if not matchmod.patkind(pat):
649 649 pats = pathutil.canonpath(repo.root, repo.getcwd(), pat)
650 650 if pats in repo[x]:
651 651 return True
652 652 else:
653 653 c = repo[x]
654 654 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
655 655 for f in c.manifest():
656 656 if m(f):
657 657 return True
658 658 return False
659 659
660 660 return subset.filter(matches)
661 661
662 662 def converted(repo, subset, x):
663 663 """``converted([id])``
664 664 Changesets converted from the given identifier in the old repository if
665 665 present, or all converted changesets if no identifier is specified.
666 666 """
667 667
668 668 # There is exactly no chance of resolving the revision, so do a simple
669 669 # string compare and hope for the best
670 670
671 671 rev = None
672 672 # i18n: "converted" is a keyword
673 673 l = getargs(x, 0, 1, _('converted takes one or no arguments'))
674 674 if l:
675 675 # i18n: "converted" is a keyword
676 676 rev = getstring(l[0], _('converted requires a revision'))
677 677
678 678 def _matchvalue(r):
679 679 source = repo[r].extra().get('convert_revision', None)
680 680 return source is not None and (rev is None or source.startswith(rev))
681 681
682 682 return subset.filter(lambda r: _matchvalue(r))
683 683
684 684 def date(repo, subset, x):
685 685 """``date(interval)``
686 686 Changesets within the interval, see :hg:`help dates`.
687 687 """
688 688 # i18n: "date" is a keyword
689 689 ds = getstring(x, _("date requires a string"))
690 690 dm = util.matchdate(ds)
691 691 return subset.filter(lambda x: dm(repo[x].date()[0]))
692 692
693 693 def desc(repo, subset, x):
694 694 """``desc(string)``
695 695 Search commit message for string. The match is case-insensitive.
696 696 """
697 697 # i18n: "desc" is a keyword
698 698 ds = encoding.lower(getstring(x, _("desc requires a string")))
699 699
700 700 def matches(x):
701 701 c = repo[x]
702 702 return ds in encoding.lower(c.description())
703 703
704 704 return subset.filter(matches)
705 705
706 706 def _descendants(repo, subset, x, followfirst=False):
707 707 roots = getset(repo, fullreposet(repo), x)
708 708 if not roots:
709 709 return baseset()
710 710 s = _revdescendants(repo, roots, followfirst)
711 711
712 712 # Both sets need to be ascending in order to lazily return the union
713 713 # in the correct order.
714 714 base = subset & roots
715 715 desc = subset & s
716 716 result = base + desc
717 717 if subset.isascending():
718 718 result.sort()
719 719 elif subset.isdescending():
720 720 result.sort(reverse=True)
721 721 else:
722 722 result = subset & result
723 723 return result
724 724
725 725 def descendants(repo, subset, x):
726 726 """``descendants(set)``
727 727 Changesets which are descendants of changesets in set.
728 728 """
729 729 return _descendants(repo, subset, x)
730 730
731 731 def _firstdescendants(repo, subset, x):
732 732 # ``_firstdescendants(set)``
733 733 # Like ``descendants(set)`` but follows only the first parents.
734 734 return _descendants(repo, subset, x, followfirst=True)
735 735
736 736 def destination(repo, subset, x):
737 737 """``destination([set])``
738 738 Changesets that were created by a graft, transplant or rebase operation,
739 739 with the given revisions specified as the source. Omitting the optional set
740 740 is the same as passing all().
741 741 """
742 742 if x is not None:
743 743 sources = getset(repo, fullreposet(repo), x)
744 744 else:
745 745 sources = fullreposet(repo)
746 746
747 747 dests = set()
748 748
749 749 # subset contains all of the possible destinations that can be returned, so
750 750 # iterate over them and see if their source(s) were provided in the arg set.
751 751 # Even if the immediate src of r is not in the arg set, src's source (or
752 752 # further back) may be. Scanning back further than the immediate src allows
753 753 # transitive transplants and rebases to yield the same results as transitive
754 754 # grafts.
755 755 for r in subset:
756 756 src = _getrevsource(repo, r)
757 757 lineage = None
758 758
759 759 while src is not None:
760 760 if lineage is None:
761 761 lineage = list()
762 762
763 763 lineage.append(r)
764 764
765 765 # The visited lineage is a match if the current source is in the arg
766 766 # set. Since every candidate dest is visited by way of iterating
767 767 # subset, any dests further back in the lineage will be tested by a
768 768 # different iteration over subset. Likewise, if the src was already
769 769 # selected, the current lineage can be selected without going back
770 770 # further.
771 771 if src in sources or src in dests:
772 772 dests.update(lineage)
773 773 break
774 774
775 775 r = src
776 776 src = _getrevsource(repo, r)
777 777
778 778 return subset.filter(dests.__contains__)
779 779
780 780 def divergent(repo, subset, x):
781 781 """``divergent()``
782 782 Final successors of changesets with an alternative set of final successors.
783 783 """
784 784 # i18n: "divergent" is a keyword
785 785 getargs(x, 0, 0, _("divergent takes no arguments"))
786 786 divergent = obsmod.getrevs(repo, 'divergent')
787 787 return subset & divergent
788 788
789 789 def draft(repo, subset, x):
790 790 """``draft()``
791 791 Changeset in draft phase."""
792 792 # i18n: "draft" is a keyword
793 793 getargs(x, 0, 0, _("draft takes no arguments"))
794 794 phase = repo._phasecache.phase
795 795 target = phases.draft
796 796 condition = lambda r: phase(repo, r) == target
797 797 return subset.filter(condition, cache=False)
798 798
799 799 def extinct(repo, subset, x):
800 800 """``extinct()``
801 801 Obsolete changesets with obsolete descendants only.
802 802 """
803 803 # i18n: "extinct" is a keyword
804 804 getargs(x, 0, 0, _("extinct takes no arguments"))
805 805 extincts = obsmod.getrevs(repo, 'extinct')
806 806 return subset & extincts
807 807
808 808 def extra(repo, subset, x):
809 809 """``extra(label, [value])``
810 810 Changesets with the given label in the extra metadata, with the given
811 811 optional value.
812 812
813 813 If `value` starts with `re:`, the remainder of the value is treated as
814 814 a regular expression. To match a value that actually starts with `re:`,
815 815 use the prefix `literal:`.
816 816 """
817 817
818 818 # i18n: "extra" is a keyword
819 819 l = getargs(x, 1, 2, _('extra takes at least 1 and at most 2 arguments'))
820 820 # i18n: "extra" is a keyword
821 821 label = getstring(l[0], _('first argument to extra must be a string'))
822 822 value = None
823 823
824 824 if len(l) > 1:
825 825 # i18n: "extra" is a keyword
826 826 value = getstring(l[1], _('second argument to extra must be a string'))
827 827 kind, value, matcher = _stringmatcher(value)
828 828
829 829 def _matchvalue(r):
830 830 extra = repo[r].extra()
831 831 return label in extra and (value is None or matcher(extra[label]))
832 832
833 833 return subset.filter(lambda r: _matchvalue(r))
834 834
835 835 def filelog(repo, subset, x):
836 836 """``filelog(pattern)``
837 837 Changesets connected to the specified filelog.
838 838
839 839 For performance reasons, visits only revisions mentioned in the file-level
840 840 filelog, rather than filtering through all changesets (much faster, but
841 841 doesn't include deletes or duplicate changes). For a slower, more accurate
842 842 result, use ``file()``.
843 843
844 844 The pattern without explicit kind like ``glob:`` is expected to be
845 845 relative to the current directory and match against a file exactly
846 846 for efficiency.
847 847
848 848 If some linkrev points to revisions filtered by the current repoview, we'll
849 849 work around it to return a non-filtered value.
850 850 """
851 851
852 852 # i18n: "filelog" is a keyword
853 853 pat = getstring(x, _("filelog requires a pattern"))
854 854 s = set()
855 855 cl = repo.changelog
856 856
857 857 if not matchmod.patkind(pat):
858 858 f = pathutil.canonpath(repo.root, repo.getcwd(), pat)
859 859 files = [f]
860 860 else:
861 861 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=repo[None])
862 862 files = (f for f in repo[None] if m(f))
863 863
864 864 for f in files:
865 865 backrevref = {} # final value for: filerev -> changerev
866 866 lowestchild = {} # lowest known filerev child of a filerev
867 867 delayed = [] # filerev with filtered linkrev, for post-processing
868 868 lowesthead = None # cache for manifest content of all head revisions
869 869 fl = repo.file(f)
870 870 for fr in list(fl):
871 871 rev = fl.linkrev(fr)
872 872 if rev not in cl:
873 873 # changerev pointed in linkrev is filtered
874 874 # record it for post processing.
875 875 delayed.append((fr, rev))
876 876 continue
877 877 for p in fl.parentrevs(fr):
878 878 if 0 <= p and p not in lowestchild:
879 879 lowestchild[p] = fr
880 880 backrevref[fr] = rev
881 881 s.add(rev)
882 882
883 883 # Post-processing of all filerevs we skipped because they were
884 884 # filtered. If such filerevs have known and unfiltered children, this
885 885 # means they have an unfiltered appearance out there. We'll use linkrev
886 886 # adjustment to find one of these appearances. The lowest known child
887 887 # will be used as a starting point because it is the best upper-bound we
888 888 # have.
889 889 #
890 890 # This approach will fail when an unfiltered but linkrev-shadowed
891 891 # appearance exists in a head changeset without unfiltered filerev
892 892 # children anywhere.
893 893 while delayed:
894 894 # must be a descending iteration. To slowly fill lowest child
895 895 # information that is of potential use by the next item.
896 896 fr, rev = delayed.pop()
897 897 lkr = rev
898 898
899 899 child = lowestchild.get(fr)
900 900
901 901 if child is None:
902 902 # search for existence of this file revision in a head revision.
903 903 # There are three possibilities:
904 904 # - the revision exists in a head and we can find an
905 905 # introduction from there,
906 906 # - the revision does not exist in a head because it has been
907 907 # changed since its introduction: we would have found a child
908 908 # and be in the other 'else' clause,
909 909 # - all versions of the revision are hidden.
910 910 if lowesthead is None:
911 911 lowesthead = {}
912 912 for h in repo.heads():
913 913 fnode = repo[h].manifest().get(f)
914 914 if fnode is not None:
915 915 lowesthead[fl.rev(fnode)] = h
916 916 headrev = lowesthead.get(fr)
917 917 if headrev is None:
918 918 # content is nowhere unfiltered
919 919 continue
920 920 rev = repo[headrev][f].introrev()
921 921 else:
922 922 # the lowest known child is a good upper bound
923 923 childcrev = backrevref[child]
924 924 # XXX this does not guarantee returning the lowest
925 925 # introduction of this revision, but this gives a
926 926 # result which is a good start and will fit in most
927 927 # cases. We probably need to fix the multiple
928 928 # introductions case properly (report each
929 929 # introduction, even for identical file revisions)
930 930 # once and for all at some point anyway.
931 931 for p in repo[childcrev][f].parents():
932 932 if p.filerev() == fr:
933 933 rev = p.rev()
934 934 break
935 935 if rev == lkr: # no shadowed entry found
936 936 # XXX This should never happen unless some manifest points
937 937 # to biggish file revisions (like a revision that uses a
938 938 # parent that never appears in the manifest ancestors)
939 939 continue
940 940
941 941 # Fill the data for the next iteration.
942 942 for p in fl.parentrevs(fr):
943 943 if 0 <= p and p not in lowestchild:
944 944 lowestchild[p] = fr
945 945 backrevref[fr] = rev
946 946 s.add(rev)
947 947
948 948 return subset & s
949 949
950 950 def first(repo, subset, x):
951 951 """``first(set, [n])``
952 952 An alias for limit().
953 953 """
954 954 return limit(repo, subset, x)
955 955
956 956 def _follow(repo, subset, x, name, followfirst=False):
957 957 l = getargs(x, 0, 1, _("%s takes no arguments or a filename") % name)
958 958 c = repo['.']
959 959 if l:
960 960 x = getstring(l[0], _("%s expected a filename") % name)
961 961 if x in c:
962 962 cx = c[x]
963 963 s = set(ctx.rev() for ctx in cx.ancestors(followfirst=followfirst))
964 964 # include the revision responsible for the most recent version
965 965 s.add(cx.introrev())
966 966 else:
967 967 return baseset()
968 968 else:
969 969 s = _revancestors(repo, baseset([c.rev()]), followfirst)
970 970
971 971 return subset & s
972 972
973 973 def follow(repo, subset, x):
974 974 """``follow([file])``
975 975 An alias for ``::.`` (ancestors of the working directory's first parent).
976 976 If a filename is specified, the history of the given file is followed,
977 977 including copies.
978 978 """
979 979 return _follow(repo, subset, x, 'follow')
980 980
981 981 def _followfirst(repo, subset, x):
982 982 # ``followfirst([file])``
983 983 # Like ``follow([file])`` but follows only the first parent of
984 984 # every revision or file revision.
985 985 return _follow(repo, subset, x, '_followfirst', followfirst=True)
986 986
987 987 def getall(repo, subset, x):
988 988 """``all()``
989 989 All changesets, the same as ``0:tip``.
990 990 """
991 991 # i18n: "all" is a keyword
992 992 getargs(x, 0, 0, _("all takes no arguments"))
993 993 return subset & spanset(repo) # drop "null" if any
994 994
995 995 def grep(repo, subset, x):
996 996 """``grep(regex)``
997 997 Like ``keyword(string)`` but accepts a regex. Use ``grep(r'...')``
998 998 to ensure special escape characters are handled correctly. Unlike
999 999 ``keyword(string)``, the match is case-sensitive.
1000 1000 """
1001 1001 try:
1002 1002 # i18n: "grep" is a keyword
1003 1003 gr = re.compile(getstring(x, _("grep requires a string")))
1004 1004 except re.error, e:
1005 1005 raise error.ParseError(_('invalid match pattern: %s') % e)
1006 1006
1007 1007 def matches(x):
1008 1008 c = repo[x]
1009 1009 for e in c.files() + [c.user(), c.description()]:
1010 1010 if gr.search(e):
1011 1011 return True
1012 1012 return False
1013 1013
1014 1014 return subset.filter(matches)
1015 1015
1016 1016 def _matchfiles(repo, subset, x):
1017 1017 # _matchfiles takes a revset list of prefixed arguments:
1018 1018 #
1019 1019 # [p:foo, i:bar, x:baz]
1020 1020 #
1021 1021 # builds a match object from them and filters subset. Allowed
1022 1022 # prefixes are 'p:' for regular patterns, 'i:' for include
1023 1023 # patterns and 'x:' for exclude patterns. Use 'r:' prefix to pass
1024 1024 # a revision identifier, or the empty string to reference the
1025 1025 # working directory, from which the match object is
1026 1026 # initialized. Use 'd:' to set the default matching mode, default
1027 1027 # to 'glob'. At most one 'r:' and 'd:' argument can be passed.
1028 1028
1029 1029 # i18n: "_matchfiles" is a keyword
1030 1030 l = getargs(x, 1, -1, _("_matchfiles requires at least one argument"))
1031 1031 pats, inc, exc = [], [], []
1032 1032 rev, default = None, None
1033 1033 for arg in l:
1034 1034 # i18n: "_matchfiles" is a keyword
1035 1035 s = getstring(arg, _("_matchfiles requires string arguments"))
1036 1036 prefix, value = s[:2], s[2:]
1037 1037 if prefix == 'p:':
1038 1038 pats.append(value)
1039 1039 elif prefix == 'i:':
1040 1040 inc.append(value)
1041 1041 elif prefix == 'x:':
1042 1042 exc.append(value)
1043 1043 elif prefix == 'r:':
1044 1044 if rev is not None:
1045 1045 # i18n: "_matchfiles" is a keyword
1046 1046 raise error.ParseError(_('_matchfiles expected at most one '
1047 1047 'revision'))
1048 1048 if value != '': # empty means working directory; leave rev as None
1049 1049 rev = value
1050 1050 elif prefix == 'd:':
1051 1051 if default is not None:
1052 1052 # i18n: "_matchfiles" is a keyword
1053 1053 raise error.ParseError(_('_matchfiles expected at most one '
1054 1054 'default mode'))
1055 1055 default = value
1056 1056 else:
1057 1057 # i18n: "_matchfiles" is a keyword
1058 1058 raise error.ParseError(_('invalid _matchfiles prefix: %s') % prefix)
1059 1059 if not default:
1060 1060 default = 'glob'
1061 1061
1062 1062 m = matchmod.match(repo.root, repo.getcwd(), pats, include=inc,
1063 1063 exclude=exc, ctx=repo[rev], default=default)
1064 1064
1065 1065 def matches(x):
1066 1066 for f in repo[x].files():
1067 1067 if m(f):
1068 1068 return True
1069 1069 return False
1070 1070
1071 1071 return subset.filter(matches)
1072 1072
1073 1073 def hasfile(repo, subset, x):
1074 1074 """``file(pattern)``
1075 1075 Changesets affecting files matched by pattern.
1076 1076
1077 1077 For a faster but less accurate result, consider using ``filelog()``
1078 1078 instead.
1079 1079
1080 1080 This predicate uses ``glob:`` as the default kind of pattern.
1081 1081 """
1082 1082 # i18n: "file" is a keyword
1083 1083 pat = getstring(x, _("file requires a pattern"))
1084 1084 return _matchfiles(repo, subset, ('string', 'p:' + pat))
1085 1085
1086 1086 def head(repo, subset, x):
1087 1087 """``head()``
1088 1088 Changeset is a named branch head.
1089 1089 """
1090 1090 # i18n: "head" is a keyword
1091 1091 getargs(x, 0, 0, _("head takes no arguments"))
1092 1092 hs = set()
1093 1093 for b, ls in repo.branchmap().iteritems():
1094 1094 hs.update(repo[h].rev() for h in ls)
1095 1095 return baseset(hs).filter(subset.__contains__)
1096 1096
1097 1097 def heads(repo, subset, x):
1098 1098 """``heads(set)``
1099 1099 Members of set with no children in set.
1100 1100 """
1101 1101 s = getset(repo, subset, x)
1102 1102 ps = parents(repo, subset, x)
1103 1103 return s - ps
1104 1104
1105 1105 def hidden(repo, subset, x):
1106 1106 """``hidden()``
1107 1107 Hidden changesets.
1108 1108 """
1109 1109 # i18n: "hidden" is a keyword
1110 1110 getargs(x, 0, 0, _("hidden takes no arguments"))
1111 1111 hiddenrevs = repoview.filterrevs(repo, 'visible')
1112 1112 return subset & hiddenrevs
1113 1113
1114 1114 def keyword(repo, subset, x):
1115 1115 """``keyword(string)``
1116 1116 Search commit message, user name, and names of changed files for
1117 1117 string. The match is case-insensitive.
1118 1118 """
1119 1119 # i18n: "keyword" is a keyword
1120 1120 kw = encoding.lower(getstring(x, _("keyword requires a string")))
1121 1121
1122 1122 def matches(r):
1123 1123 c = repo[r]
1124 1124 return any(kw in encoding.lower(t) for t in c.files() + [c.user(),
1125 1125 c.description()])
1126 1126
1127 1127 return subset.filter(matches)
1128 1128
1129 1129 def limit(repo, subset, x):
1130 1130 """``limit(set, [n])``
1131 1131 First n members of set, defaulting to 1.
1132 1132 """
1133 1133 # i18n: "limit" is a keyword
1134 1134 l = getargs(x, 1, 2, _("limit requires one or two arguments"))
1135 1135 try:
1136 1136 lim = 1
1137 1137 if len(l) == 2:
1138 1138 # i18n: "limit" is a keyword
1139 1139 lim = int(getstring(l[1], _("limit requires a number")))
1140 1140 except (TypeError, ValueError):
1141 1141 # i18n: "limit" is a keyword
1142 1142 raise error.ParseError(_("limit expects a number"))
1143 1143 ss = subset
1144 1144 os = getset(repo, fullreposet(repo), l[0])
1145 1145 result = []
1146 1146 it = iter(os)
1147 1147 for x in xrange(lim):
1148 1148 y = next(it, None)
1149 1149 if y is None:
1150 1150 break
1151 1151 elif y in ss:
1152 1152 result.append(y)
1153 1153 return baseset(result)
1154 1154
1155 1155 def last(repo, subset, x):
1156 1156 """``last(set, [n])``
1157 1157 Last n members of set, defaulting to 1.
1158 1158 """
1159 1159 # i18n: "last" is a keyword
1160 1160 l = getargs(x, 1, 2, _("last requires one or two arguments"))
1161 1161 try:
1162 1162 lim = 1
1163 1163 if len(l) == 2:
1164 1164 # i18n: "last" is a keyword
1165 1165 lim = int(getstring(l[1], _("last requires a number")))
1166 1166 except (TypeError, ValueError):
1167 1167 # i18n: "last" is a keyword
1168 1168 raise error.ParseError(_("last expects a number"))
1169 1169 ss = subset
1170 1170 os = getset(repo, fullreposet(repo), l[0])
1171 1171 os.reverse()
1172 1172 result = []
1173 1173 it = iter(os)
1174 1174 for x in xrange(lim):
1175 1175 y = next(it, None)
1176 1176 if y is None:
1177 1177 break
1178 1178 elif y in ss:
1179 1179 result.append(y)
1180 1180 return baseset(result)
1181 1181
1182 1182 def maxrev(repo, subset, x):
1183 1183 """``max(set)``
1184 1184 Changeset with highest revision number in set.
1185 1185 """
1186 1186 os = getset(repo, fullreposet(repo), x)
1187 1187 if os:
1188 1188 m = os.max()
1189 1189 if m in subset:
1190 1190 return baseset([m])
1191 1191 return baseset()
1192 1192
1193 1193 def merge(repo, subset, x):
1194 1194 """``merge()``
1195 1195 Changeset is a merge changeset.
1196 1196 """
1197 1197 # i18n: "merge" is a keyword
1198 1198 getargs(x, 0, 0, _("merge takes no arguments"))
1199 1199 cl = repo.changelog
1200 1200 return subset.filter(lambda r: cl.parentrevs(r)[1] != -1)
1201 1201
1202 1202 def branchpoint(repo, subset, x):
1203 1203 """``branchpoint()``
1204 1204 Changesets with more than one child.
1205 1205 """
1206 1206 # i18n: "branchpoint" is a keyword
1207 1207 getargs(x, 0, 0, _("branchpoint takes no arguments"))
1208 1208 cl = repo.changelog
1209 1209 if not subset:
1210 1210 return baseset()
1211 1211 baserev = min(subset)
1212 1212 parentscount = [0]*(len(repo) - baserev)
1213 1213 for r in cl.revs(start=baserev + 1):
1214 1214 for p in cl.parentrevs(r):
1215 1215 if p >= baserev:
1216 1216 parentscount[p - baserev] += 1
1217 1217 return subset.filter(lambda r: parentscount[r - baserev] > 1)
1218 1218
1219 1219 def minrev(repo, subset, x):
1220 1220 """``min(set)``
1221 1221 Changeset with lowest revision number in set.
1222 1222 """
1223 1223 os = getset(repo, fullreposet(repo), x)
1224 1224 if os:
1225 1225 m = os.min()
1226 1226 if m in subset:
1227 1227 return baseset([m])
1228 1228 return baseset()
1229 1229
1230 1230 def modifies(repo, subset, x):
1231 1231 """``modifies(pattern)``
1232 1232 Changesets modifying files matched by pattern.
1233 1233
1234 1234 The pattern without explicit kind like ``glob:`` is expected to be
1235 1235 relative to the current directory and match against a file or a
1236 1236 directory.
1237 1237 """
1238 1238 # i18n: "modifies" is a keyword
1239 1239 pat = getstring(x, _("modifies requires a pattern"))
1240 1240 return checkstatus(repo, subset, pat, 0)
1241 1241
1242 1242 def named(repo, subset, x):
1243 1243 """``named(namespace)``
1244 1244 The changesets in a given namespace.
1245 1245
1246 1246 If `namespace` starts with `re:`, the remainder of the string is treated as
1247 1247 a regular expression. To match a namespace that actually starts with `re:`,
1248 1248 use the prefix `literal:`.
1249 1249 """
1250 1250 # i18n: "named" is a keyword
1251 1251 args = getargs(x, 1, 1, _('named requires a namespace argument'))
1252 1252
1253 1253 ns = getstring(args[0],
1254 1254 # i18n: "named" is a keyword
1255 1255 _('the argument to named must be a string'))
1256 1256 kind, pattern, matcher = _stringmatcher(ns)
1257 1257 namespaces = set()
1258 1258 if kind == 'literal':
1259 1259 if pattern not in repo.names:
1260 1260 raise error.RepoLookupError(_("namespace '%s' does not exist")
1261 1261 % ns)
1262 1262 namespaces.add(repo.names[pattern])
1263 1263 else:
1264 1264 for name, ns in repo.names.iteritems():
1265 1265 if matcher(name):
1266 1266 namespaces.add(ns)
1267 1267 if not namespaces:
1268 1268 raise error.RepoLookupError(_("no namespace exists"
1269 1269 " that match '%s'") % pattern)
1270 1270
1271 1271 names = set()
1272 1272 for ns in namespaces:
1273 1273 for name in ns.listnames(repo):
1274 1274 if name not in ns.deprecated:
1275 1275 names.update(repo[n].rev() for n in ns.nodes(repo, name))
1276 1276
1277 1277 names -= set([node.nullrev])
1278 1278 return subset & names
1279 1279
1280 1280 def node_(repo, subset, x):
1281 1281 """``id(string)``
1282 1282 Revision non-ambiguously specified by the given hex string prefix.
1283 1283 """
1284 1284 # i18n: "id" is a keyword
1285 1285 l = getargs(x, 1, 1, _("id requires one argument"))
1286 1286 # i18n: "id" is a keyword
1287 1287 n = getstring(l[0], _("id requires a string"))
1288 1288 if len(n) == 40:
1289 1289 try:
1290 1290 rn = repo.changelog.rev(node.bin(n))
1291 1291 except (LookupError, TypeError):
1292 1292 rn = None
1293 1293 else:
1294 1294 rn = None
1295 1295 pm = repo.changelog._partialmatch(n)
1296 1296 if pm is not None:
1297 1297 rn = repo.changelog.rev(pm)
1298 1298
1299 1299 if rn is None:
1300 1300 return baseset()
1301 1301 result = baseset([rn])
1302 1302 return result & subset
1303 1303
1304 1304 def obsolete(repo, subset, x):
1305 1305 """``obsolete()``
1306 1306 Mutable changeset with a newer version."""
1307 1307 # i18n: "obsolete" is a keyword
1308 1308 getargs(x, 0, 0, _("obsolete takes no arguments"))
1309 1309 obsoletes = obsmod.getrevs(repo, 'obsolete')
1310 1310 return subset & obsoletes
1311 1311
1312 1312 def only(repo, subset, x):
1313 1313 """``only(set, [set])``
1314 1314 Changesets that are ancestors of the first set that are not ancestors
1315 1315 of any other head in the repo. If a second set is specified, the result
1316 1316 is ancestors of the first set that are not ancestors of the second set
1317 1317 (i.e. ::<set1> - ::<set2>).
1318 1318 """
1319 1319 cl = repo.changelog
1320 1320 # i18n: "only" is a keyword
1321 1321 args = getargs(x, 1, 2, _('only takes one or two arguments'))
1322 1322 include = getset(repo, fullreposet(repo), args[0])
1323 1323 if len(args) == 1:
1324 1324 if not include:
1325 1325 return baseset()
1326 1326
1327 1327 descendants = set(_revdescendants(repo, include, False))
1328 1328 exclude = [rev for rev in cl.headrevs()
1329 1329 if not rev in descendants and not rev in include]
1330 1330 else:
1331 1331 exclude = getset(repo, fullreposet(repo), args[1])
1332 1332
1333 1333 results = set(cl.findmissingrevs(common=exclude, heads=include))
1334 1334 return subset & results
1335 1335
1336 1336 def origin(repo, subset, x):
1337 1337 """``origin([set])``
1338 1338 Changesets that were specified as a source for the grafts, transplants or
1339 1339 rebases that created the given revisions. Omitting the optional set is the
1340 1340 same as passing all(). If a changeset created by these operations is itself
1341 1341 specified as a source for one of these operations, only the source changeset
1342 1342 for the first operation is selected.
1343 1343 """
1344 1344 if x is not None:
1345 1345 dests = getset(repo, fullreposet(repo), x)
1346 1346 else:
1347 1347 dests = fullreposet(repo)
1348 1348
1349 1349 def _firstsrc(rev):
1350 1350 src = _getrevsource(repo, rev)
1351 1351 if src is None:
1352 1352 return None
1353 1353
1354 1354 while True:
1355 1355 prev = _getrevsource(repo, src)
1356 1356
1357 1357 if prev is None:
1358 1358 return src
1359 1359 src = prev
1360 1360
1361 1361 o = set([_firstsrc(r) for r in dests])
1362 1362 o -= set([None])
1363 1363 return subset & o
1364 1364
1365 1365 def outgoing(repo, subset, x):
1366 1366 """``outgoing([path])``
1367 1367 Changesets not found in the specified destination repository, or the
1368 1368 default push location.
1369 1369 """
1370 1370 # Avoid cycles.
1371 1371 import discovery
1372 1372 import hg
1373 1373 # i18n: "outgoing" is a keyword
1374 1374 l = getargs(x, 0, 1, _("outgoing takes one or no arguments"))
1375 1375 # i18n: "outgoing" is a keyword
1376 1376 dest = l and getstring(l[0], _("outgoing requires a repository path")) or ''
1377 1377 dest = repo.ui.expandpath(dest or 'default-push', dest or 'default')
1378 1378 dest, branches = hg.parseurl(dest)
1379 1379 revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
1380 1380 if revs:
1381 1381 revs = [repo.lookup(rev) for rev in revs]
1382 1382 other = hg.peer(repo, {}, dest)
1383 1383 repo.ui.pushbuffer()
1384 1384 outgoing = discovery.findcommonoutgoing(repo, other, onlyheads=revs)
1385 1385 repo.ui.popbuffer()
1386 1386 cl = repo.changelog
1387 1387 o = set([cl.rev(r) for r in outgoing.missing])
1388 1388 return subset & o
1389 1389
1390 1390 def p1(repo, subset, x):
1391 1391 """``p1([set])``
1392 1392 First parent of changesets in set, or the working directory.
1393 1393 """
1394 1394 if x is None:
1395 1395 p = repo[x].p1().rev()
1396 1396 if p >= 0:
1397 1397 return subset & baseset([p])
1398 1398 return baseset()
1399 1399
1400 1400 ps = set()
1401 1401 cl = repo.changelog
1402 1402 for r in getset(repo, fullreposet(repo), x):
1403 1403 ps.add(cl.parentrevs(r)[0])
1404 1404 ps -= set([node.nullrev])
1405 1405 return subset & ps
1406 1406
1407 1407 def p2(repo, subset, x):
1408 1408 """``p2([set])``
1409 1409 Second parent of changesets in set, or the working directory.
1410 1410 """
1411 1411 if x is None:
1412 1412 ps = repo[x].parents()
1413 1413 try:
1414 1414 p = ps[1].rev()
1415 1415 if p >= 0:
1416 1416 return subset & baseset([p])
1417 1417 return baseset()
1418 1418 except IndexError:
1419 1419 return baseset()
1420 1420
1421 1421 ps = set()
1422 1422 cl = repo.changelog
1423 1423 for r in getset(repo, fullreposet(repo), x):
1424 1424 ps.add(cl.parentrevs(r)[1])
1425 1425 ps -= set([node.nullrev])
1426 1426 return subset & ps
1427 1427
1428 1428 def parents(repo, subset, x):
1429 1429 """``parents([set])``
1430 1430 The set of all parents for all changesets in set, or the working directory.
1431 1431 """
1432 1432 if x is None:
1433 1433 ps = set(p.rev() for p in repo[x].parents())
1434 1434 else:
1435 1435 ps = set()
1436 1436 cl = repo.changelog
1437 1437 for r in getset(repo, fullreposet(repo), x):
1438 1438 ps.update(cl.parentrevs(r))
1439 1439 ps -= set([node.nullrev])
1440 1440 return subset & ps
1441 1441
1442 1442 def parentspec(repo, subset, x, n):
1443 1443 """``set^0``
1444 1444 The set.
1445 1445 ``set^1`` (or ``set^``), ``set^2``
1446 1446 First or second parent, respectively, of all changesets in set.
1447 1447 """
1448 1448 try:
1449 1449 n = int(n[1])
1450 1450 if n not in (0, 1, 2):
1451 1451 raise ValueError
1452 1452 except (TypeError, ValueError):
1453 1453 raise error.ParseError(_("^ expects a number 0, 1, or 2"))
1454 1454 ps = set()
1455 1455 cl = repo.changelog
1456 1456 for r in getset(repo, fullreposet(repo), x):
1457 1457 if n == 0:
1458 1458 ps.add(r)
1459 1459 elif n == 1:
1460 1460 ps.add(cl.parentrevs(r)[0])
1461 1461 elif n == 2:
1462 1462 parents = cl.parentrevs(r)
1463 1463 if len(parents) > 1:
1464 1464 ps.add(parents[1])
1465 1465 return subset & ps
1466 1466
1467 1467 def present(repo, subset, x):
1468 1468 """``present(set)``
1469 1469 An empty set, if any revision in set isn't found; otherwise,
1470 1470 all revisions in set.
1471 1471
1472 1472 If any of specified revisions is not present in the local repository,
1473 1473 the query is normally aborted. But this predicate allows the query
1474 1474 to continue even in such cases.
1475 1475 """
1476 1476 try:
1477 1477 return getset(repo, subset, x)
1478 1478 except error.RepoLookupError:
1479 1479 return baseset()
1480 1480
1481 1481 # for internal use
1482 1482 def _notpublic(repo, subset, x):
1483 1483 getargs(x, 0, 0, "_notpublic takes no arguments")
1484 1484 if repo._phasecache._phasesets:
1485 1485 s = set()
1486 1486 for u in repo._phasecache._phasesets[1:]:
1487 1487 s.update(u)
1488 1488 return subset & s
1489 1489 else:
1490 1490 phase = repo._phasecache.phase
1491 1491 target = phases.public
1492 1492 condition = lambda r: phase(repo, r) != target
1493 1493 return subset.filter(condition, cache=False)
1494 1494
1495 1495 def public(repo, subset, x):
1496 1496 """``public()``
1497 1497 Changeset in public phase."""
1498 1498 # i18n: "public" is a keyword
1499 1499 getargs(x, 0, 0, _("public takes no arguments"))
1500 1500 phase = repo._phasecache.phase
1501 1501 target = phases.public
1502 1502 condition = lambda r: phase(repo, r) == target
1503 1503 return subset.filter(condition, cache=False)
1504 1504
1505 1505 def remote(repo, subset, x):
1506 1506 """``remote([id [,path]])``
1507 1507 Local revision that corresponds to the given identifier in a
1508 1508 remote repository, if present. Here, the '.' identifier is a
1509 1509 synonym for the current local branch.
1510 1510 """
1511 1511
1512 1512 import hg # avoid start-up nasties
1513 1513 # i18n: "remote" is a keyword
1514 1514 l = getargs(x, 0, 2, _("remote takes one, two or no arguments"))
1515 1515
1516 1516 q = '.'
1517 1517 if len(l) > 0:
1518 1518 # i18n: "remote" is a keyword
1519 1519 q = getstring(l[0], _("remote requires a string id"))
1520 1520 if q == '.':
1521 1521 q = repo['.'].branch()
1522 1522
1523 1523 dest = ''
1524 1524 if len(l) > 1:
1525 1525 # i18n: "remote" is a keyword
1526 1526 dest = getstring(l[1], _("remote requires a repository path"))
1527 1527 dest = repo.ui.expandpath(dest or 'default')
1528 1528 dest, branches = hg.parseurl(dest)
1529 1529 revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
1530 1530 if revs:
1531 1531 revs = [repo.lookup(rev) for rev in revs]
1532 1532 other = hg.peer(repo, {}, dest)
1533 1533 n = other.lookup(q)
1534 1534 if n in repo:
1535 1535 r = repo[n].rev()
1536 1536 if r in subset:
1537 1537 return baseset([r])
1538 1538 return baseset()
1539 1539
1540 1540 def removes(repo, subset, x):
1541 1541 """``removes(pattern)``
1542 1542 Changesets which remove files matching pattern.
1543 1543
1544 1544 The pattern without explicit kind like ``glob:`` is expected to be
1545 1545 relative to the current directory and match against a file or a
1546 1546 directory.
1547 1547 """
1548 1548 # i18n: "removes" is a keyword
1549 1549 pat = getstring(x, _("removes requires a pattern"))
1550 1550 return checkstatus(repo, subset, pat, 2)
1551 1551
1552 1552 def rev(repo, subset, x):
1553 1553 """``rev(number)``
1554 1554 Revision with the given numeric identifier.
1555 1555 """
1556 1556 # i18n: "rev" is a keyword
1557 1557 l = getargs(x, 1, 1, _("rev requires one argument"))
1558 1558 try:
1559 1559 # i18n: "rev" is a keyword
1560 1560 l = int(getstring(l[0], _("rev requires a number")))
1561 1561 except (TypeError, ValueError):
1562 1562 # i18n: "rev" is a keyword
1563 1563 raise error.ParseError(_("rev expects a number"))
1564 1564 if l not in repo.changelog and l != node.nullrev:
1565 1565 return baseset()
1566 1566 return subset & baseset([l])
1567 1567
1568 1568 def matching(repo, subset, x):
1569 1569 """``matching(revision [, field])``
1570 1570 Changesets in which a given set of fields match the set of fields in the
1571 1571 selected revision or set.
1572 1572
1573 1573 To match more than one field pass the list of fields to match separated
1574 1574 by spaces (e.g. ``author description``).
1575 1575
1576 1576 Valid fields are most regular revision fields and some special fields.
1577 1577
1578 1578 Regular revision fields are ``description``, ``author``, ``branch``,
1579 1579 ``date``, ``files``, ``phase``, ``parents``, ``substate``, ``user``
1580 1580 and ``diff``.
1581 1581 Note that ``author`` and ``user`` are synonyms. ``diff`` refers to the
1582 1582 contents of the revision. Two revisions matching their ``diff`` will
1583 1583 also match their ``files``.
1584 1584
1585 1585 Special fields are ``summary`` and ``metadata``:
1586 1586 ``summary`` matches the first line of the description.
1587 1587 ``metadata`` is equivalent to matching ``description user date``
1588 1588 (i.e. it matches the main metadata fields).
1589 1589
1590 1590 ``metadata`` is the default field which is used when no fields are
1591 1591 specified. You can match more than one field at a time.
1592 1592 """
1593 1593 # i18n: "matching" is a keyword
1594 1594 l = getargs(x, 1, 2, _("matching takes 1 or 2 arguments"))
1595 1595
1596 1596 revs = getset(repo, fullreposet(repo), l[0])
1597 1597
1598 1598 fieldlist = ['metadata']
1599 1599 if len(l) > 1:
1600 1600 fieldlist = getstring(l[1],
1601 1601 # i18n: "matching" is a keyword
1602 1602 _("matching requires a string "
1603 1603 "as its second argument")).split()
1604 1604
1605 1605 # Make sure that there are no repeated fields,
1606 1606 # expand the 'special' 'metadata' field type
1607 1607 # and check the 'files' whenever we check the 'diff'
1608 1608 fields = []
1609 1609 for field in fieldlist:
1610 1610 if field == 'metadata':
1611 1611 fields += ['user', 'description', 'date']
1612 1612 elif field == 'diff':
1613 1613 # a revision matching the diff must also match the files
1614 1614 # since matching the diff is very costly, make sure to
1615 1615 # also match the files first
1616 1616 fields += ['files', 'diff']
1617 1617 else:
1618 1618 if field == 'author':
1619 1619 field = 'user'
1620 1620 fields.append(field)
1621 1621 fields = set(fields)
1622 1622 if 'summary' in fields and 'description' in fields:
1623 1623 # If a revision matches its description it also matches its summary
1624 1624 fields.discard('summary')
1625 1625
1626 1626 # We may want to match more than one field
1627 1627 # Not all fields take the same amount of time to be matched
1628 1628 # Sort the selected fields in order of increasing matching cost
1629 1629 fieldorder = ['phase', 'parents', 'user', 'date', 'branch', 'summary',
1630 1630 'files', 'description', 'substate', 'diff']
1631 1631 def fieldkeyfunc(f):
1632 1632 try:
1633 1633 return fieldorder.index(f)
1634 1634 except ValueError:
1635 1635 # assume an unknown field is very costly
1636 1636 return len(fieldorder)
1637 1637 fields = list(fields)
1638 1638 fields.sort(key=fieldkeyfunc)
1639 1639
1640 1640 # Each field will be matched with its own "getfield" function
1641 1641 # which will be added to the getfieldfuncs array of functions
1642 1642 getfieldfuncs = []
1643 1643 _funcs = {
1644 1644 'user': lambda r: repo[r].user(),
1645 1645 'branch': lambda r: repo[r].branch(),
1646 1646 'date': lambda r: repo[r].date(),
1647 1647 'description': lambda r: repo[r].description(),
1648 1648 'files': lambda r: repo[r].files(),
1649 1649 'parents': lambda r: repo[r].parents(),
1650 1650 'phase': lambda r: repo[r].phase(),
1651 1651 'substate': lambda r: repo[r].substate,
1652 1652 'summary': lambda r: repo[r].description().splitlines()[0],
1653 1653 'diff': lambda r: list(repo[r].diff(git=True),)
1654 1654 }
1655 1655 for info in fields:
1656 1656 getfield = _funcs.get(info, None)
1657 1657 if getfield is None:
1658 1658 raise error.ParseError(
1659 1659 # i18n: "matching" is a keyword
1660 1660 _("unexpected field name passed to matching: %s") % info)
1661 1661 getfieldfuncs.append(getfield)
1662 1662 # convert the getfield array of functions into a "getinfo" function
1663 1663 # which returns an array of field values (or a single value if there
1664 1664 # is only one field to match)
1665 1665 getinfo = lambda r: [f(r) for f in getfieldfuncs]
1666 1666
1667 1667 def matches(x):
1668 1668 for rev in revs:
1669 1669 target = getinfo(rev)
1670 1670 match = True
1671 1671 for n, f in enumerate(getfieldfuncs):
1672 1672 if target[n] != f(x):
1673 1673 match = False
1674 1674 if match:
1675 1675 return True
1676 1676 return False
1677 1677
1678 1678 return subset.filter(matches)
1679 1679
1680 1680 def reverse(repo, subset, x):
1681 1681 """``reverse(set)``
1682 1682 Reverse order of set.
1683 1683 """
1684 1684 l = getset(repo, subset, x)
1685 1685 l.reverse()
1686 1686 return l
1687 1687
1688 1688 def roots(repo, subset, x):
1689 1689 """``roots(set)``
1690 1690 Changesets in set with no parent changeset in set.
1691 1691 """
1692 1692 s = getset(repo, fullreposet(repo), x)
1693 1693 subset = subset & s# baseset([r for r in s if r in subset])
1694 1694 cs = _children(repo, subset, s)
1695 1695 return subset - cs
1696 1696
1697 1697 def secret(repo, subset, x):
1698 1698 """``secret()``
1699 1699 Changeset in secret phase."""
1700 1700 # i18n: "secret" is a keyword
1701 1701 getargs(x, 0, 0, _("secret takes no arguments"))
1702 1702 phase = repo._phasecache.phase
1703 1703 target = phases.secret
1704 1704 condition = lambda r: phase(repo, r) == target
1705 1705 return subset.filter(condition, cache=False)
1706 1706
1707 1707 def sort(repo, subset, x):
1708 1708 """``sort(set[, [-]key...])``
1709 1709 Sort set by keys. The default sort order is ascending, specify a key
1710 1710 as ``-key`` to sort in descending order.
1711 1711
1712 1712 The keys can be:
1713 1713
1714 1714 - ``rev`` for the revision number,
1715 1715 - ``branch`` for the branch name,
1716 1716 - ``desc`` for the commit message (description),
1717 1717 - ``user`` for user name (``author`` can be used as an alias),
1718 1718 - ``date`` for the commit date
1719 1719 """
1720 1720 # i18n: "sort" is a keyword
1721 1721 l = getargs(x, 1, 2, _("sort requires one or two arguments"))
1722 1722 keys = "rev"
1723 1723 if len(l) == 2:
1724 1724 # i18n: "sort" is a keyword
1725 1725 keys = getstring(l[1], _("sort spec must be a string"))
1726 1726
1727 1727 s = l[0]
1728 1728 keys = keys.split()
1729 1729 l = []
1730 1730 def invert(s):
1731 1731 return "".join(chr(255 - ord(c)) for c in s)
1732 1732 revs = getset(repo, subset, s)
1733 1733 if keys == ["rev"]:
1734 1734 revs.sort()
1735 1735 return revs
1736 1736 elif keys == ["-rev"]:
1737 1737 revs.sort(reverse=True)
1738 1738 return revs
1739 1739 for r in revs:
1740 1740 c = repo[r]
1741 1741 e = []
1742 1742 for k in keys:
1743 1743 if k == 'rev':
1744 1744 e.append(r)
1745 1745 elif k == '-rev':
1746 1746 e.append(-r)
1747 1747 elif k == 'branch':
1748 1748 e.append(c.branch())
1749 1749 elif k == '-branch':
1750 1750 e.append(invert(c.branch()))
1751 1751 elif k == 'desc':
1752 1752 e.append(c.description())
1753 1753 elif k == '-desc':
1754 1754 e.append(invert(c.description()))
1755 1755 elif k in 'user author':
1756 1756 e.append(c.user())
1757 1757 elif k in '-user -author':
1758 1758 e.append(invert(c.user()))
1759 1759 elif k == 'date':
1760 1760 e.append(c.date()[0])
1761 1761 elif k == '-date':
1762 1762 e.append(-c.date()[0])
1763 1763 else:
1764 1764 raise error.ParseError(_("unknown sort key %r") % k)
1765 1765 e.append(r)
1766 1766 l.append(e)
1767 1767 l.sort()
1768 1768 return baseset([e[-1] for e in l])
1769 1769
1770 1770 def subrepo(repo, subset, x):
1771 1771 """``subrepo([pattern])``
1772 1772 Changesets that add, modify or remove the given subrepo. If no subrepo
1773 1773 pattern is named, any subrepo changes are returned.
1774 1774 """
1775 1775 # i18n: "subrepo" is a keyword
1776 1776 args = getargs(x, 0, 1, _('subrepo takes at most one argument'))
1777 1777 if len(args) != 0:
1778 1778 pat = getstring(args[0], _("subrepo requires a pattern"))
1779 1779
1780 1780 m = matchmod.exact(repo.root, repo.root, ['.hgsubstate'])
1781 1781
1782 1782 def submatches(names):
1783 1783 k, p, m = _stringmatcher(pat)
1784 1784 for name in names:
1785 1785 if m(name):
1786 1786 yield name
1787 1787
1788 1788 def matches(x):
1789 1789 c = repo[x]
1790 1790 s = repo.status(c.p1().node(), c.node(), match=m)
1791 1791
1792 1792 if len(args) == 0:
1793 1793 return s.added or s.modified or s.removed
1794 1794
1795 1795 if s.added:
1796 1796 return any(submatches(c.substate.keys()))
1797 1797
1798 1798 if s.modified:
1799 1799 subs = set(c.p1().substate.keys())
1800 1800 subs.update(c.substate.keys())
1801 1801
1802 1802 for path in submatches(subs):
1803 1803 if c.p1().substate.get(path) != c.substate.get(path):
1804 1804 return True
1805 1805
1806 1806 if s.removed:
1807 1807 return any(submatches(c.p1().substate.keys()))
1808 1808
1809 1809 return False
1810 1810
1811 1811 return subset.filter(matches)
1812 1812
1813 1813 def _stringmatcher(pattern):
1814 1814 """
1815 1815 accepts a string, possibly starting with 're:' or 'literal:' prefix.
1816 1816 returns the matcher name, pattern, and matcher function.
1817 1817 missing or unknown prefixes are treated as literal matches.
1818 1818
1819 1819 helper for tests:
1820 1820 >>> def test(pattern, *tests):
1821 1821 ... kind, pattern, matcher = _stringmatcher(pattern)
1822 1822 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
1823 1823
1824 1824 exact matching (no prefix):
1825 1825 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
1826 1826 ('literal', 'abcdefg', [False, False, True])
1827 1827
1828 1828 regex matching ('re:' prefix)
1829 1829 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
1830 1830 ('re', 'a.+b', [False, False, True])
1831 1831
1832 1832 force exact matches ('literal:' prefix)
1833 1833 >>> test('literal:re:foobar', 'foobar', 're:foobar')
1834 1834 ('literal', 're:foobar', [False, True])
1835 1835
1836 1836 unknown prefixes are ignored and treated as literals
1837 1837 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
1838 1838 ('literal', 'foo:bar', [False, False, True])
1839 1839 """
1840 1840 if pattern.startswith('re:'):
1841 1841 pattern = pattern[3:]
1842 1842 try:
1843 1843 regex = re.compile(pattern)
1844 1844 except re.error, e:
1845 1845 raise error.ParseError(_('invalid regular expression: %s')
1846 1846 % e)
1847 1847 return 're', pattern, regex.search
1848 1848 elif pattern.startswith('literal:'):
1849 1849 pattern = pattern[8:]
1850 1850 return 'literal', pattern, pattern.__eq__
1851 1851
1852 1852 def _substringmatcher(pattern):
1853 1853 kind, pattern, matcher = _stringmatcher(pattern)
1854 1854 if kind == 'literal':
1855 1855 matcher = lambda s: pattern in s
1856 1856 return kind, pattern, matcher
1857 1857
1858 1858 def tag(repo, subset, x):
1859 1859 """``tag([name])``
1860 1860 The specified tag by name, or all tagged revisions if no name is given.
1861 1861
1862 1862 If `name` starts with `re:`, the remainder of the name is treated as
1863 1863 a regular expression. To match a tag that actually starts with `re:`,
1864 1864 use the prefix `literal:`.
1865 1865 """
1866 1866 # i18n: "tag" is a keyword
1867 1867 args = getargs(x, 0, 1, _("tag takes one or no arguments"))
1868 1868 cl = repo.changelog
1869 1869 if args:
1870 1870 pattern = getstring(args[0],
1871 1871 # i18n: "tag" is a keyword
1872 1872 _('the argument to tag must be a string'))
1873 1873 kind, pattern, matcher = _stringmatcher(pattern)
1874 1874 if kind == 'literal':
1875 1875 # avoid resolving all tags
1876 1876 tn = repo._tagscache.tags.get(pattern, None)
1877 1877 if tn is None:
1878 1878 raise error.RepoLookupError(_("tag '%s' does not exist")
1879 1879 % pattern)
1880 1880 s = set([repo[tn].rev()])
1881 1881 else:
1882 1882 s = set([cl.rev(n) for t, n in repo.tagslist() if matcher(t)])
1883 1883 else:
1884 1884 s = set([cl.rev(n) for t, n in repo.tagslist() if t != 'tip'])
1885 1885 return subset & s
1886 1886
1887 1887 def tagged(repo, subset, x):
1888 1888 return tag(repo, subset, x)
1889 1889
1890 1890 def unstable(repo, subset, x):
1891 1891 """``unstable()``
1892 1892 Non-obsolete changesets with obsolete ancestors.
1893 1893 """
1894 1894 # i18n: "unstable" is a keyword
1895 1895 getargs(x, 0, 0, _("unstable takes no arguments"))
1896 1896 unstables = obsmod.getrevs(repo, 'unstable')
1897 1897 return subset & unstables
1898 1898
1899 1899
1900 1900 def user(repo, subset, x):
1901 1901 """``user(string)``
1902 1902 User name contains string. The match is case-insensitive.
1903 1903
1904 1904 If `string` starts with `re:`, the remainder of the string is treated as
1905 1905 a regular expression. To match a user that actually contains `re:`, use
1906 1906 the prefix `literal:`.
1907 1907 """
1908 1908 return author(repo, subset, x)
1909 1909
1910 1910 # experimental
1911 1911 def wdir(repo, subset, x):
1912 1912 # i18n: "wdir" is a keyword
1913 1913 getargs(x, 0, 0, _("wdir takes no arguments"))
1914 1914 if None in subset or isinstance(subset, fullreposet):
1915 1915 return baseset([None])
1916 1916 return baseset()
1917 1917
1918 1918 # for internal use
1919 1919 def _list(repo, subset, x):
1920 1920 s = getstring(x, "internal error")
1921 1921 if not s:
1922 1922 return baseset()
1923 1923 # remove duplicates here. it's difficult for caller to deduplicate sets
1924 1924 # because different symbols can point to the same rev.
1925 1925 cl = repo.changelog
1926 1926 ls = []
1927 1927 seen = set()
1928 1928 for t in s.split('\0'):
1929 1929 try:
1930 1930 # fast path for integer revision
1931 1931 r = int(t)
1932 1932 if str(r) != t or r not in cl:
1933 1933 raise ValueError
1934 1934 except ValueError:
1935 1935 r = repo[t].rev()
1936 1936 if r in seen:
1937 1937 continue
1938 1938 if (r in subset
1939 1939 or r == node.nullrev and isinstance(subset, fullreposet)):
1940 1940 ls.append(r)
1941 1941 seen.add(r)
1942 1942 return baseset(ls)
1943 1943
1944 1944 # for internal use
1945 1945 def _intlist(repo, subset, x):
1946 1946 s = getstring(x, "internal error")
1947 1947 if not s:
1948 1948 return baseset()
1949 1949 ls = [int(r) for r in s.split('\0')]
1950 1950 s = subset
1951 1951 return baseset([r for r in ls if r in s])
1952 1952
1953 1953 # for internal use
1954 1954 def _hexlist(repo, subset, x):
1955 1955 s = getstring(x, "internal error")
1956 1956 if not s:
1957 1957 return baseset()
1958 1958 cl = repo.changelog
1959 1959 ls = [cl.rev(node.bin(r)) for r in s.split('\0')]
1960 1960 s = subset
1961 1961 return baseset([r for r in ls if r in s])
1962 1962
1963 1963 symbols = {
1964 1964 "adds": adds,
1965 1965 "all": getall,
1966 1966 "ancestor": ancestor,
1967 1967 "ancestors": ancestors,
1968 1968 "_firstancestors": _firstancestors,
1969 1969 "author": author,
1970 1970 "bisect": bisect,
1971 1971 "bisected": bisected,
1972 1972 "bookmark": bookmark,
1973 1973 "branch": branch,
1974 1974 "branchpoint": branchpoint,
1975 1975 "bumped": bumped,
1976 1976 "bundle": bundle,
1977 1977 "children": children,
1978 1978 "closed": closed,
1979 1979 "contains": contains,
1980 1980 "converted": converted,
1981 1981 "date": date,
1982 1982 "desc": desc,
1983 1983 "descendants": descendants,
1984 1984 "_firstdescendants": _firstdescendants,
1985 1985 "destination": destination,
1986 1986 "divergent": divergent,
1987 1987 "draft": draft,
1988 1988 "extinct": extinct,
1989 1989 "extra": extra,
1990 1990 "file": hasfile,
1991 1991 "filelog": filelog,
1992 1992 "first": first,
1993 1993 "follow": follow,
1994 1994 "_followfirst": _followfirst,
1995 1995 "grep": grep,
1996 1996 "head": head,
1997 1997 "heads": heads,
1998 1998 "hidden": hidden,
1999 1999 "id": node_,
2000 2000 "keyword": keyword,
2001 2001 "last": last,
2002 2002 "limit": limit,
2003 2003 "_matchfiles": _matchfiles,
2004 2004 "max": maxrev,
2005 2005 "merge": merge,
2006 2006 "min": minrev,
2007 2007 "modifies": modifies,
2008 2008 "named": named,
2009 2009 "obsolete": obsolete,
2010 2010 "only": only,
2011 2011 "origin": origin,
2012 2012 "outgoing": outgoing,
2013 2013 "p1": p1,
2014 2014 "p2": p2,
2015 2015 "parents": parents,
2016 2016 "present": present,
2017 2017 "public": public,
2018 2018 "_notpublic": _notpublic,
2019 2019 "remote": remote,
2020 2020 "removes": removes,
2021 2021 "rev": rev,
2022 2022 "reverse": reverse,
2023 2023 "roots": roots,
2024 2024 "sort": sort,
2025 2025 "secret": secret,
2026 2026 "subrepo": subrepo,
2027 2027 "matching": matching,
2028 2028 "tag": tag,
2029 2029 "tagged": tagged,
2030 2030 "user": user,
2031 2031 "unstable": unstable,
2032 2032 "wdir": wdir,
2033 2033 "_list": _list,
2034 2034 "_intlist": _intlist,
2035 2035 "_hexlist": _hexlist,
2036 2036 }
2037 2037
2038 2038 # symbols which can't be used for a DoS attack for any given input
2039 2039 # (e.g. those which accept regexes as plain strings shouldn't be included)
2040 2040 # functions that just return a lot of changesets (like all) don't count here
2041 2041 safesymbols = set([
2042 2042 "adds",
2043 2043 "all",
2044 2044 "ancestor",
2045 2045 "ancestors",
2046 2046 "_firstancestors",
2047 2047 "author",
2048 2048 "bisect",
2049 2049 "bisected",
2050 2050 "bookmark",
2051 2051 "branch",
2052 2052 "branchpoint",
2053 2053 "bumped",
2054 2054 "bundle",
2055 2055 "children",
2056 2056 "closed",
2057 2057 "converted",
2058 2058 "date",
2059 2059 "desc",
2060 2060 "descendants",
2061 2061 "_firstdescendants",
2062 2062 "destination",
2063 2063 "divergent",
2064 2064 "draft",
2065 2065 "extinct",
2066 2066 "extra",
2067 2067 "file",
2068 2068 "filelog",
2069 2069 "first",
2070 2070 "follow",
2071 2071 "_followfirst",
2072 2072 "head",
2073 2073 "heads",
2074 2074 "hidden",
2075 2075 "id",
2076 2076 "keyword",
2077 2077 "last",
2078 2078 "limit",
2079 2079 "_matchfiles",
2080 2080 "max",
2081 2081 "merge",
2082 2082 "min",
2083 2083 "modifies",
2084 2084 "obsolete",
2085 2085 "only",
2086 2086 "origin",
2087 2087 "outgoing",
2088 2088 "p1",
2089 2089 "p2",
2090 2090 "parents",
2091 2091 "present",
2092 2092 "public",
2093 2093 "_notpublic",
2094 2094 "remote",
2095 2095 "removes",
2096 2096 "rev",
2097 2097 "reverse",
2098 2098 "roots",
2099 2099 "sort",
2100 2100 "secret",
2101 2101 "matching",
2102 2102 "tag",
2103 2103 "tagged",
2104 2104 "user",
2105 2105 "unstable",
2106 2106 "wdir",
2107 2107 "_list",
2108 2108 "_intlist",
2109 2109 "_hexlist",
2110 2110 ])
2111 2111
2112 2112 methods = {
2113 2113 "range": rangeset,
2114 2114 "dagrange": dagrange,
2115 2115 "string": stringset,
2116 2116 "symbol": stringset,
2117 2117 "and": andset,
2118 2118 "or": orset,
2119 2119 "not": notset,
2120 2120 "list": listset,
2121 2121 "func": func,
2122 2122 "ancestor": ancestorspec,
2123 2123 "parent": parentspec,
2124 2124 "parentpost": p1,
2125 2125 }
2126 2126
2127 2127 def optimize(x, small):
2128 2128 if x is None:
2129 2129 return 0, x
2130 2130
2131 2131 smallbonus = 1
2132 2132 if small:
2133 2133 smallbonus = .5
2134 2134
2135 2135 op = x[0]
2136 2136 if op == 'minus':
2137 2137 return optimize(('and', x[1], ('not', x[2])), small)
2138 2138 elif op == 'only':
2139 2139 return optimize(('func', ('symbol', 'only'),
2140 2140 ('list', x[1], x[2])), small)
2141 2141 elif op == 'onlypost':
2142 2142 return optimize(('func', ('symbol', 'only'), x[1]), small)
2143 2143 elif op == 'dagrangepre':
2144 2144 return optimize(('func', ('symbol', 'ancestors'), x[1]), small)
2145 2145 elif op == 'dagrangepost':
2146 2146 return optimize(('func', ('symbol', 'descendants'), x[1]), small)
2147 2147 elif op == 'rangepre':
2148 2148 return optimize(('range', ('string', '0'), x[1]), small)
2149 2149 elif op == 'rangepost':
2150 2150 return optimize(('range', x[1], ('string', 'tip')), small)
2151 2151 elif op == 'negate':
2152 2152 return optimize(('string',
2153 2153 '-' + getstring(x[1], _("can't negate that"))), small)
2154 2154 elif op in 'string symbol negate':
2155 2155 return smallbonus, x # single revisions are small
2156 2156 elif op == 'and':
2157 2157 wa, ta = optimize(x[1], True)
2158 2158 wb, tb = optimize(x[2], True)
2159 2159
2160 2160 # (::x and not ::y)/(not ::y and ::x) have a fast path
2161 2161 def isonly(revs, bases):
2162 2162 return (
2163 2163 revs[0] == 'func'
2164 2164 and getstring(revs[1], _('not a symbol')) == 'ancestors'
2165 2165 and bases[0] == 'not'
2166 2166 and bases[1][0] == 'func'
2167 2167 and getstring(bases[1][1], _('not a symbol')) == 'ancestors')
2168 2168
2169 2169 w = min(wa, wb)
2170 2170 if isonly(ta, tb):
2171 2171 return w, ('func', ('symbol', 'only'), ('list', ta[2], tb[1][2]))
2172 2172 if isonly(tb, ta):
2173 2173 return w, ('func', ('symbol', 'only'), ('list', tb[2], ta[1][2]))
2174 2174
2175 2175 if wa > wb:
2176 2176 return w, (op, tb, ta)
2177 2177 return w, (op, ta, tb)
2178 2178 elif op == 'or':
2179 2179 # fast path for machine-generated expression, that is likely to have
2180 2180 # lots of trivial revisions: 'a + b + c()' to '_list(a b) + c()'
2181 2181 ws, ts, ss = [], [], []
2182 2182 def flushss():
2183 2183 if not ss:
2184 2184 return
2185 2185 if len(ss) == 1:
2186 2186 w, t = ss[0]
2187 2187 else:
2188 2188 s = '\0'.join(t[1] for w, t in ss)
2189 2189 y = ('func', ('symbol', '_list'), ('string', s))
2190 2190 w, t = optimize(y, False)
2191 2191 ws.append(w)
2192 2192 ts.append(t)
2193 2193 del ss[:]
2194 2194 for y in x[1:]:
2195 2195 w, t = optimize(y, False)
2196 2196 if t[0] == 'string' or t[0] == 'symbol':
2197 2197 ss.append((w, t))
2198 2198 continue
2199 2199 flushss()
2200 2200 ws.append(w)
2201 2201 ts.append(t)
2202 2202 flushss()
2203 2203 if len(ts) == 1:
2204 2204 return ws[0], ts[0] # 'or' operation is fully optimized out
2205 2205 # we can't reorder trees by weight because it would change the order.
2206 2206 # ("sort(a + b)" == "sort(b + a)", but "a + b" != "b + a")
2207 2207 # ts = tuple(t for w, t in sorted(zip(ws, ts), key=lambda wt: wt[0]))
2208 2208 return max(ws), (op,) + tuple(ts)
2209 2209 elif op == 'not':
2210 2210 # Optimize not public() to _notpublic() because we have a fast version
2211 2211 if x[1] == ('func', ('symbol', 'public'), None):
2212 2212 newsym = ('func', ('symbol', '_notpublic'), None)
2213 2213 o = optimize(newsym, not small)
2214 2214 return o[0], o[1]
2215 2215 else:
2216 2216 o = optimize(x[1], not small)
2217 2217 return o[0], (op, o[1])
2218 2218 elif op == 'parentpost':
2219 2219 o = optimize(x[1], small)
2220 2220 return o[0], (op, o[1])
2221 2221 elif op == 'group':
2222 2222 return optimize(x[1], small)
2223 2223 elif op in 'dagrange range list parent ancestorspec':
2224 2224 if op == 'parent':
2225 2225 # x^:y means (x^) : y, not x ^ (:y)
2226 2226 post = ('parentpost', x[1])
2227 2227 if x[2][0] == 'dagrangepre':
2228 2228 return optimize(('dagrange', post, x[2][1]), small)
2229 2229 elif x[2][0] == 'rangepre':
2230 2230 return optimize(('range', post, x[2][1]), small)
2231 2231
2232 2232 wa, ta = optimize(x[1], small)
2233 2233 wb, tb = optimize(x[2], small)
2234 2234 return wa + wb, (op, ta, tb)
2235 2235 elif op == 'func':
2236 2236 f = getstring(x[1], _("not a symbol"))
2237 2237 wa, ta = optimize(x[2], small)
2238 2238 if f in ("author branch closed date desc file grep keyword "
2239 2239 "outgoing user"):
2240 2240 w = 10 # slow
2241 2241 elif f in "modifies adds removes":
2242 2242 w = 30 # slower
2243 2243 elif f == "contains":
2244 2244 w = 100 # very slow
2245 2245 elif f == "ancestor":
2246 2246 w = 1 * smallbonus
2247 2247 elif f in "reverse limit first _intlist":
2248 2248 w = 0
2249 2249 elif f in "sort":
2250 2250 w = 10 # assume most sorts look at changelog
2251 2251 else:
2252 2252 w = 1
2253 2253 return w + wa, (op, x[1], ta)
2254 2254 return 1, x
2255 2255
2256 2256 _aliasarg = ('func', ('symbol', '_aliasarg'))
2257 2257 def _getaliasarg(tree):
2258 2258 """If tree matches ('func', ('symbol', '_aliasarg'), ('string', X))
2259 2259 return X, None otherwise.
2260 2260 """
2261 2261 if (len(tree) == 3 and tree[:2] == _aliasarg
2262 2262 and tree[2][0] == 'string'):
2263 2263 return tree[2][1]
2264 2264 return None
2265 2265
2266 2266 def _checkaliasarg(tree, known=None):
2267 2267 """Check tree contains no _aliasarg construct or only ones which
2268 2268 value is in known. Used to avoid alias placeholders injection.
2269 2269 """
2270 2270 if isinstance(tree, tuple):
2271 2271 arg = _getaliasarg(tree)
2272 2272 if arg is not None and (not known or arg not in known):
2273 2273 raise error.UnknownIdentifier('_aliasarg', [])
2274 2274 for t in tree:
2275 2275 _checkaliasarg(t, known)
2276 2276
2277 2277 # the set of valid characters for the initial letter of symbols in
2278 2278 # alias declarations and definitions
2279 2279 _aliassyminitletters = set(c for c in [chr(i) for i in xrange(256)]
2280 2280 if c.isalnum() or c in '._@$' or ord(c) > 127)
2281 2281
2282 2282 def _tokenizealias(program, lookup=None):
2283 2283 """Parse alias declaration/definition into a stream of tokens
2284 2284
2285 2285 This allows symbol names to use also ``$`` as an initial letter
2286 2286 (for backward compatibility), and callers of this function should
2287 2287 examine whether ``$`` is used also for unexpected symbols or not.
2288 2288 """
2289 2289 return tokenize(program, lookup=lookup,
2290 2290 syminitletters=_aliassyminitletters)
2291 2291
2292 2292 def _parsealiasdecl(decl):
2293 2293 """Parse alias declaration ``decl``
2294 2294
2295 2295 This returns ``(name, tree, args, errorstr)`` tuple:
2296 2296
2297 2297 - ``name``: of declared alias (may be ``decl`` itself at error)
2298 2298 - ``tree``: parse result (or ``None`` at error)
2299 2299 - ``args``: list of alias argument names (or None for symbol declaration)
2300 2300 - ``errorstr``: detail about detected error (or None)
2301 2301
2302 2302 >>> _parsealiasdecl('foo')
2303 2303 ('foo', ('symbol', 'foo'), None, None)
2304 2304 >>> _parsealiasdecl('$foo')
2305 2305 ('$foo', None, None, "'$' not for alias arguments")
2306 2306 >>> _parsealiasdecl('foo::bar')
2307 2307 ('foo::bar', None, None, 'invalid format')
2308 2308 >>> _parsealiasdecl('foo bar')
2309 2309 ('foo bar', None, None, 'at 4: invalid token')
2310 2310 >>> _parsealiasdecl('foo()')
2311 2311 ('foo', ('func', ('symbol', 'foo')), [], None)
2312 2312 >>> _parsealiasdecl('$foo()')
2313 2313 ('$foo()', None, None, "'$' not for alias arguments")
2314 2314 >>> _parsealiasdecl('foo($1, $2)')
2315 2315 ('foo', ('func', ('symbol', 'foo')), ['$1', '$2'], None)
2316 2316 >>> _parsealiasdecl('foo(bar_bar, baz.baz)')
2317 2317 ('foo', ('func', ('symbol', 'foo')), ['bar_bar', 'baz.baz'], None)
2318 2318 >>> _parsealiasdecl('foo($1, $2, nested($1, $2))')
2319 2319 ('foo($1, $2, nested($1, $2))', None, None, 'invalid argument list')
2320 2320 >>> _parsealiasdecl('foo(bar($1, $2))')
2321 2321 ('foo(bar($1, $2))', None, None, 'invalid argument list')
2322 2322 >>> _parsealiasdecl('foo("string")')
2323 2323 ('foo("string")', None, None, 'invalid argument list')
2324 2324 >>> _parsealiasdecl('foo($1, $2')
2325 2325 ('foo($1, $2', None, None, 'at 10: unexpected token: end')
2326 2326 >>> _parsealiasdecl('foo("string')
2327 2327 ('foo("string', None, None, 'at 5: unterminated string')
2328 2328 >>> _parsealiasdecl('foo($1, $2, $1)')
2329 2329 ('foo', None, None, 'argument names collide with each other')
2330 2330 """
2331 2331 p = parser.parser(_tokenizealias, elements)
2332 2332 try:
2333 2333 tree, pos = p.parse(decl)
2334 2334 if (pos != len(decl)):
2335 2335 raise error.ParseError(_('invalid token'), pos)
2336 2336
2337 2337 if isvalidsymbol(tree):
2338 2338 # "name = ...." style
2339 2339 name = getsymbol(tree)
2340 2340 if name.startswith('$'):
2341 2341 return (decl, None, None, _("'$' not for alias arguments"))
2342 2342 return (name, ('symbol', name), None, None)
2343 2343
2344 2344 if isvalidfunc(tree):
2345 2345 # "name(arg, ....) = ...." style
2346 2346 name = getfuncname(tree)
2347 2347 if name.startswith('$'):
2348 2348 return (decl, None, None, _("'$' not for alias arguments"))
2349 2349 args = []
2350 2350 for arg in getfuncargs(tree):
2351 2351 if not isvalidsymbol(arg):
2352 2352 return (decl, None, None, _("invalid argument list"))
2353 2353 args.append(getsymbol(arg))
2354 2354 if len(args) != len(set(args)):
2355 2355 return (name, None, None,
2356 2356 _("argument names collide with each other"))
2357 2357 return (name, ('func', ('symbol', name)), args, None)
2358 2358
2359 2359 return (decl, None, None, _("invalid format"))
2360 2360 except error.ParseError, inst:
2361 2361 return (decl, None, None, parseerrordetail(inst))
2362 2362
2363 2363 def _parsealiasdefn(defn, args):
2364 2364 """Parse alias definition ``defn``
2365 2365
2366 2366 This function also replaces alias argument references in the
2367 2367 specified definition by ``_aliasarg(ARGNAME)``.
2368 2368
2369 2369 ``args`` is a list of alias argument names, or None if the alias
2370 2370 is declared as a symbol.
2371 2371
2372 2372 This returns "tree" as parsing result.
2373 2373
2374 2374 >>> args = ['$1', '$2', 'foo']
2375 2375 >>> print prettyformat(_parsealiasdefn('$1 or foo', args))
2376 2376 (or
2377 2377 (func
2378 2378 ('symbol', '_aliasarg')
2379 2379 ('string', '$1'))
2380 2380 (func
2381 2381 ('symbol', '_aliasarg')
2382 2382 ('string', 'foo')))
2383 2383 >>> try:
2384 2384 ... _parsealiasdefn('$1 or $bar', args)
2385 2385 ... except error.ParseError, inst:
2386 2386 ... print parseerrordetail(inst)
2387 2387 at 6: '$' not for alias arguments
2388 2388 >>> args = ['$1', '$10', 'foo']
2389 2389 >>> print prettyformat(_parsealiasdefn('$10 or foobar', args))
2390 2390 (or
2391 2391 (func
2392 2392 ('symbol', '_aliasarg')
2393 2393 ('string', '$10'))
2394 2394 ('symbol', 'foobar'))
2395 2395 >>> print prettyformat(_parsealiasdefn('"$1" or "foo"', args))
2396 2396 (or
2397 2397 ('string', '$1')
2398 2398 ('string', 'foo'))
2399 2399 """
2400 2400 def tokenizedefn(program, lookup=None):
2401 2401 if args:
2402 2402 argset = set(args)
2403 2403 else:
2404 2404 argset = set()
2405 2405
2406 2406 for t, value, pos in _tokenizealias(program, lookup=lookup):
2407 2407 if t == 'symbol':
2408 2408 if value in argset:
2409 2409 # emulate tokenization of "_aliasarg('ARGNAME')":
2410 2410 # "_aliasarg()" is an unknown symbol only used separate
2411 2411 # alias argument placeholders from regular strings.
2412 2412 yield ('symbol', '_aliasarg', pos)
2413 2413 yield ('(', None, pos)
2414 2414 yield ('string', value, pos)
2415 2415 yield (')', None, pos)
2416 2416 continue
2417 2417 elif value.startswith('$'):
2418 2418 raise error.ParseError(_("'$' not for alias arguments"),
2419 2419 pos)
2420 2420 yield (t, value, pos)
2421 2421
2422 2422 p = parser.parser(tokenizedefn, elements)
2423 2423 tree, pos = p.parse(defn)
2424 2424 if pos != len(defn):
2425 2425 raise error.ParseError(_('invalid token'), pos)
2426 2426 return parser.simplifyinfixops(tree, ('or',))
2427 2427
2428 2428 class revsetalias(object):
2429 2429 # whether own `error` information is already shown or not.
2430 2430 # this avoids showing same warning multiple times at each `findaliases`.
2431 2431 warned = False
2432 2432
2433 2433 def __init__(self, name, value):
2434 2434 '''Aliases like:
2435 2435
2436 2436 h = heads(default)
2437 2437 b($1) = ancestors($1) - ancestors(default)
2438 2438 '''
2439 2439 self.name, self.tree, self.args, self.error = _parsealiasdecl(name)
2440 2440 if self.error:
2441 2441 self.error = _('failed to parse the declaration of revset alias'
2442 2442 ' "%s": %s') % (self.name, self.error)
2443 2443 return
2444 2444
2445 2445 try:
2446 2446 self.replacement = _parsealiasdefn(value, self.args)
2447 2447 # Check for placeholder injection
2448 2448 _checkaliasarg(self.replacement, self.args)
2449 2449 except error.ParseError, inst:
2450 2450 self.error = _('failed to parse the definition of revset alias'
2451 2451 ' "%s": %s') % (self.name, parseerrordetail(inst))
2452 2452
2453 2453 def _getalias(aliases, tree):
2454 2454 """If tree looks like an unexpanded alias, return it. Return None
2455 2455 otherwise.
2456 2456 """
2457 2457 if isinstance(tree, tuple) and tree:
2458 2458 if tree[0] == 'symbol' and len(tree) == 2:
2459 2459 name = tree[1]
2460 2460 alias = aliases.get(name)
2461 2461 if alias and alias.args is None and alias.tree == tree:
2462 2462 return alias
2463 2463 if tree[0] == 'func' and len(tree) > 1:
2464 2464 if tree[1][0] == 'symbol' and len(tree[1]) == 2:
2465 2465 name = tree[1][1]
2466 2466 alias = aliases.get(name)
2467 2467 if alias and alias.args is not None and alias.tree == tree[:2]:
2468 2468 return alias
2469 2469 return None
2470 2470
2471 2471 def _expandargs(tree, args):
2472 2472 """Replace _aliasarg instances with the substitution value of the
2473 2473 same name in args, recursively.
2474 2474 """
2475 2475 if not tree or not isinstance(tree, tuple):
2476 2476 return tree
2477 2477 arg = _getaliasarg(tree)
2478 2478 if arg is not None:
2479 2479 return args[arg]
2480 2480 return tuple(_expandargs(t, args) for t in tree)
2481 2481
2482 2482 def _expandaliases(aliases, tree, expanding, cache):
2483 2483 """Expand aliases in tree, recursively.
2484 2484
2485 2485 'aliases' is a dictionary mapping user defined aliases to
2486 2486 revsetalias objects.
2487 2487 """
2488 2488 if not isinstance(tree, tuple):
2489 2489 # Do not expand raw strings
2490 2490 return tree
2491 2491 alias = _getalias(aliases, tree)
2492 2492 if alias is not None:
2493 2493 if alias.error:
2494 2494 raise util.Abort(alias.error)
2495 2495 if alias in expanding:
2496 2496 raise error.ParseError(_('infinite expansion of revset alias "%s" '
2497 2497 'detected') % alias.name)
2498 2498 expanding.append(alias)
2499 2499 if alias.name not in cache:
2500 2500 cache[alias.name] = _expandaliases(aliases, alias.replacement,
2501 2501 expanding, cache)
2502 2502 result = cache[alias.name]
2503 2503 expanding.pop()
2504 2504 if alias.args is not None:
2505 2505 l = getlist(tree[2])
2506 2506 if len(l) != len(alias.args):
2507 2507 raise error.ParseError(
2508 2508 _('invalid number of arguments: %s') % len(l))
2509 2509 l = [_expandaliases(aliases, a, [], cache) for a in l]
2510 2510 result = _expandargs(result, dict(zip(alias.args, l)))
2511 2511 else:
2512 2512 result = tuple(_expandaliases(aliases, t, expanding, cache)
2513 2513 for t in tree)
2514 2514 return result
2515 2515
2516 2516 def findaliases(ui, tree, showwarning=None):
2517 2517 _checkaliasarg(tree)
2518 2518 aliases = {}
2519 2519 for k, v in ui.configitems('revsetalias'):
2520 2520 alias = revsetalias(k, v)
2521 2521 aliases[alias.name] = alias
2522 2522 tree = _expandaliases(aliases, tree, [], {})
2523 2523 if showwarning:
2524 2524 # warn about problematic (but not referred) aliases
2525 2525 for name, alias in sorted(aliases.iteritems()):
2526 2526 if alias.error and not alias.warned:
2527 2527 showwarning(_('warning: %s\n') % (alias.error))
2528 2528 alias.warned = True
2529 2529 return tree
2530 2530
2531 2531 def foldconcat(tree):
2532 2532 """Fold elements to be concatenated by `##`
2533 2533 """
2534 2534 if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
2535 2535 return tree
2536 2536 if tree[0] == '_concat':
2537 2537 pending = [tree]
2538 2538 l = []
2539 2539 while pending:
2540 2540 e = pending.pop()
2541 2541 if e[0] == '_concat':
2542 2542 pending.extend(reversed(e[1:]))
2543 2543 elif e[0] in ('string', 'symbol'):
2544 2544 l.append(e[1])
2545 2545 else:
2546 2546 msg = _("\"##\" can't concatenate \"%s\" element") % (e[0])
2547 2547 raise error.ParseError(msg)
2548 2548 return ('string', ''.join(l))
2549 2549 else:
2550 2550 return tuple(foldconcat(t) for t in tree)
2551 2551
2552 2552 def parse(spec, lookup=None):
2553 2553 p = parser.parser(tokenize, elements)
2554 2554 tree, pos = p.parse(spec, lookup=lookup)
2555 2555 if pos != len(spec):
2556 2556 raise error.ParseError(_("invalid token"), pos)
2557 2557 return parser.simplifyinfixops(tree, ('or',))
2558 2558
2559 2559 def posttreebuilthook(tree, repo):
2560 2560 # hook for extensions to execute code on the optimized tree
2561 2561 pass
2562 2562
2563 2563 def match(ui, spec, repo=None):
2564 2564 if not spec:
2565 2565 raise error.ParseError(_("empty query"))
2566 2566 lookup = None
2567 2567 if repo:
2568 2568 lookup = repo.__contains__
2569 2569 tree = parse(spec, lookup)
2570 2570 if ui:
2571 2571 tree = findaliases(ui, tree, showwarning=ui.warn)
2572 2572 tree = foldconcat(tree)
2573 2573 weight, tree = optimize(tree, True)
2574 2574 posttreebuilthook(tree, repo)
2575 2575 def mfunc(repo, subset=None):
2576 2576 if subset is None:
2577 2577 subset = fullreposet(repo)
2578 2578 if util.safehasattr(subset, 'isascending'):
2579 2579 result = getset(repo, subset, tree)
2580 2580 else:
2581 2581 result = getset(repo, baseset(subset), tree)
2582 2582 return result
2583 2583 return mfunc
2584 2584
2585 2585 def formatspec(expr, *args):
2586 2586 '''
2587 2587 This is a convenience function for using revsets internally, and
2588 2588 escapes arguments appropriately. Aliases are intentionally ignored
2589 2589 so that intended expression behavior isn't accidentally subverted.
2590 2590
2591 2591 Supported arguments:
2592 2592
2593 2593 %r = revset expression, parenthesized
2594 2594 %d = int(arg), no quoting
2595 2595 %s = string(arg), escaped and single-quoted
2596 2596 %b = arg.branch(), escaped and single-quoted
2597 2597 %n = hex(arg), single-quoted
2598 2598 %% = a literal '%'
2599 2599
2600 2600 Prefixing the type with 'l' specifies a parenthesized list of that type.
2601 2601
2602 2602 >>> formatspec('%r:: and %lr', '10 or 11', ("this()", "that()"))
2603 2603 '(10 or 11):: and ((this()) or (that()))'
2604 2604 >>> formatspec('%d:: and not %d::', 10, 20)
2605 2605 '10:: and not 20::'
2606 2606 >>> formatspec('%ld or %ld', [], [1])
2607 2607 "_list('') or 1"
2608 2608 >>> formatspec('keyword(%s)', 'foo\\xe9')
2609 2609 "keyword('foo\\\\xe9')"
2610 2610 >>> b = lambda: 'default'
2611 2611 >>> b.branch = b
2612 2612 >>> formatspec('branch(%b)', b)
2613 2613 "branch('default')"
2614 2614 >>> formatspec('root(%ls)', ['a', 'b', 'c', 'd'])
2615 2615 "root(_list('a\\x00b\\x00c\\x00d'))"
2616 2616 '''
2617 2617
2618 2618 def quote(s):
2619 2619 return repr(str(s))
2620 2620
2621 2621 def argtype(c, arg):
2622 2622 if c == 'd':
2623 2623 return str(int(arg))
2624 2624 elif c == 's':
2625 2625 return quote(arg)
2626 2626 elif c == 'r':
2627 2627 parse(arg) # make sure syntax errors are confined
2628 2628 return '(%s)' % arg
2629 2629 elif c == 'n':
2630 2630 return quote(node.hex(arg))
2631 2631 elif c == 'b':
2632 2632 return quote(arg.branch())
2633 2633
2634 2634 def listexp(s, t):
2635 2635 l = len(s)
2636 2636 if l == 0:
2637 2637 return "_list('')"
2638 2638 elif l == 1:
2639 2639 return argtype(t, s[0])
2640 2640 elif t == 'd':
2641 2641 return "_intlist('%s')" % "\0".join(str(int(a)) for a in s)
2642 2642 elif t == 's':
2643 2643 return "_list('%s')" % "\0".join(s)
2644 2644 elif t == 'n':
2645 2645 return "_hexlist('%s')" % "\0".join(node.hex(a) for a in s)
2646 2646 elif t == 'b':
2647 2647 return "_list('%s')" % "\0".join(a.branch() for a in s)
2648 2648
2649 2649 m = l // 2
2650 2650 return '(%s or %s)' % (listexp(s[:m], t), listexp(s[m:], t))
2651 2651
2652 2652 ret = ''
2653 2653 pos = 0
2654 2654 arg = 0
2655 2655 while pos < len(expr):
2656 2656 c = expr[pos]
2657 2657 if c == '%':
2658 2658 pos += 1
2659 2659 d = expr[pos]
2660 2660 if d == '%':
2661 2661 ret += d
2662 2662 elif d in 'dsnbr':
2663 2663 ret += argtype(d, args[arg])
2664 2664 arg += 1
2665 2665 elif d == 'l':
2666 2666 # a list of some type
2667 2667 pos += 1
2668 2668 d = expr[pos]
2669 2669 ret += listexp(list(args[arg]), d)
2670 2670 arg += 1
2671 2671 else:
2672 2672 raise util.Abort('unexpected revspec format character %s' % d)
2673 2673 else:
2674 2674 ret += c
2675 2675 pos += 1
2676 2676
2677 2677 return ret
2678 2678
2679 2679 def prettyformat(tree):
2680 2680 return parser.prettyformat(tree, ('string', 'symbol'))
2681 2681
2682 2682 def depth(tree):
2683 2683 if isinstance(tree, tuple):
2684 2684 return max(map(depth, tree)) + 1
2685 2685 else:
2686 2686 return 0
2687 2687
2688 2688 def funcsused(tree):
2689 2689 if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
2690 2690 return set()
2691 2691 else:
2692 2692 funcs = set()
2693 2693 for s in tree[1:]:
2694 2694 funcs |= funcsused(s)
2695 2695 if tree[0] == 'func':
2696 2696 funcs.add(tree[1][1])
2697 2697 return funcs
2698 2698
2699 2699 class abstractsmartset(object):
2700 2700
2701 2701 def __nonzero__(self):
2702 2702 """True if the smartset is not empty"""
2703 2703 raise NotImplementedError()
2704 2704
2705 2705 def __contains__(self, rev):
2706 2706 """provide fast membership testing"""
2707 2707 raise NotImplementedError()
2708 2708
2709 2709 def __iter__(self):
2710 2710 """iterate the set in the order it is supposed to be iterated"""
2711 2711 raise NotImplementedError()
2712 2712
2713 2713 # Attributes containing a function to perform a fast iteration in a given
2714 2714 # direction. A smartset can have none, one, or both defined.
2715 2715 #
2716 2716 # Default value is None instead of a function returning None to avoid
2717 2717 # initializing an iterator just for testing if a fast method exists.
2718 2718 fastasc = None
2719 2719 fastdesc = None
2720 2720
2721 2721 def isascending(self):
2722 2722 """True if the set will iterate in ascending order"""
2723 2723 raise NotImplementedError()
2724 2724
2725 2725 def isdescending(self):
2726 2726 """True if the set will iterate in descending order"""
2727 2727 raise NotImplementedError()
2728 2728
2729 2729 def min(self):
2730 2730 """return the minimum element in the set"""
2731 2731 if self.fastasc is not None:
2732 2732 for r in self.fastasc():
2733 2733 return r
2734 2734 raise ValueError('arg is an empty sequence')
2735 2735 return min(self)
2736 2736
2737 2737 def max(self):
2738 2738 """return the maximum element in the set"""
2739 2739 if self.fastdesc is not None:
2740 2740 for r in self.fastdesc():
2741 2741 return r
2742 2742 raise ValueError('arg is an empty sequence')
2743 2743 return max(self)
2744 2744
2745 2745 def first(self):
2746 2746 """return the first element in the set (user iteration perspective)
2747 2747
2748 2748 Return None if the set is empty"""
2749 2749 raise NotImplementedError()
2750 2750
2751 2751 def last(self):
2752 2752 """return the last element in the set (user iteration perspective)
2753 2753
2754 2754 Return None if the set is empty"""
2755 2755 raise NotImplementedError()
2756 2756
2757 2757 def __len__(self):
2758 2758 """return the length of the smartsets
2759 2759
2760 2760 This can be expensive on smartset that could be lazy otherwise."""
2761 2761 raise NotImplementedError()
2762 2762
2763 2763 def reverse(self):
2764 2764 """reverse the expected iteration order"""
2765 2765 raise NotImplementedError()
2766 2766
2767 2767 def sort(self, reverse=True):
2768 2768 """get the set to iterate in an ascending or descending order"""
2769 2769 raise NotImplementedError()
2770 2770
2771 2771 def __and__(self, other):
2772 2772 """Returns a new object with the intersection of the two collections.
2773 2773
2774 2774 This is part of the mandatory API for smartset."""
2775 2775 if isinstance(other, fullreposet):
2776 2776 return self
2777 2777 return self.filter(other.__contains__, cache=False)
2778 2778
2779 2779 def __add__(self, other):
2780 2780 """Returns a new object with the union of the two collections.
2781 2781
2782 2782 This is part of the mandatory API for smartset."""
2783 2783 return addset(self, other)
2784 2784
2785 2785 def __sub__(self, other):
2786 2786 """Returns a new object with the substraction of the two collections.
2787 2787
2788 2788 This is part of the mandatory API for smartset."""
2789 2789 c = other.__contains__
2790 2790 return self.filter(lambda r: not c(r), cache=False)
2791 2791
2792 2792 def filter(self, condition, cache=True):
2793 2793 """Returns this smartset filtered by condition as a new smartset.
2794 2794
2795 2795 `condition` is a callable which takes a revision number and returns a
2796 2796 boolean.
2797 2797
2798 2798 This is part of the mandatory API for smartset."""
2799 2799 # builtin cannot be cached. but do not needs to
2800 2800 if cache and util.safehasattr(condition, 'func_code'):
2801 2801 condition = util.cachefunc(condition)
2802 2802 return filteredset(self, condition)
2803 2803
2804 2804 class baseset(abstractsmartset):
2805 2805 """Basic data structure that represents a revset and contains the basic
2806 2806 operation that it should be able to perform.
2807 2807
2808 2808 Every method in this class should be implemented by any smartset class.
2809 2809 """
2810 2810 def __init__(self, data=()):
2811 2811 if not isinstance(data, list):
2812 2812 data = list(data)
2813 2813 self._list = data
2814 2814 self._ascending = None
2815 2815
2816 2816 @util.propertycache
2817 2817 def _set(self):
2818 2818 return set(self._list)
2819 2819
2820 2820 @util.propertycache
2821 2821 def _asclist(self):
2822 2822 asclist = self._list[:]
2823 2823 asclist.sort()
2824 2824 return asclist
2825 2825
2826 2826 def __iter__(self):
2827 2827 if self._ascending is None:
2828 2828 return iter(self._list)
2829 2829 elif self._ascending:
2830 2830 return iter(self._asclist)
2831 2831 else:
2832 2832 return reversed(self._asclist)
2833 2833
2834 2834 def fastasc(self):
2835 2835 return iter(self._asclist)
2836 2836
2837 2837 def fastdesc(self):
2838 2838 return reversed(self._asclist)
2839 2839
2840 2840 @util.propertycache
2841 2841 def __contains__(self):
2842 2842 return self._set.__contains__
2843 2843
2844 2844 def __nonzero__(self):
2845 2845 return bool(self._list)
2846 2846
2847 2847 def sort(self, reverse=False):
2848 2848 self._ascending = not bool(reverse)
2849 2849
2850 2850 def reverse(self):
2851 2851 if self._ascending is None:
2852 2852 self._list.reverse()
2853 2853 else:
2854 2854 self._ascending = not self._ascending
2855 2855
2856 2856 def __len__(self):
2857 2857 return len(self._list)
2858 2858
2859 2859 def isascending(self):
2860 2860 """Returns True if the collection is ascending order, False if not.
2861 2861
2862 2862 This is part of the mandatory API for smartset."""
2863 2863 if len(self) <= 1:
2864 2864 return True
2865 2865 return self._ascending is not None and self._ascending
2866 2866
2867 2867 def isdescending(self):
2868 2868 """Returns True if the collection is descending order, False if not.
2869 2869
2870 2870 This is part of the mandatory API for smartset."""
2871 2871 if len(self) <= 1:
2872 2872 return True
2873 2873 return self._ascending is not None and not self._ascending
2874 2874
2875 2875 def first(self):
2876 2876 if self:
2877 2877 if self._ascending is None:
2878 2878 return self._list[0]
2879 2879 elif self._ascending:
2880 2880 return self._asclist[0]
2881 2881 else:
2882 2882 return self._asclist[-1]
2883 2883 return None
2884 2884
2885 2885 def last(self):
2886 2886 if self:
2887 2887 if self._ascending is None:
2888 2888 return self._list[-1]
2889 2889 elif self._ascending:
2890 2890 return self._asclist[-1]
2891 2891 else:
2892 2892 return self._asclist[0]
2893 2893 return None
2894 2894
2895 2895 def __repr__(self):
2896 2896 d = {None: '', False: '-', True: '+'}[self._ascending]
2897 2897 return '<%s%s %r>' % (type(self).__name__, d, self._list)
2898 2898
2899 2899 class filteredset(abstractsmartset):
2900 2900 """Duck type for baseset class which iterates lazily over the revisions in
2901 2901 the subset and contains a function which tests for membership in the
2902 2902 revset
2903 2903 """
2904 2904 def __init__(self, subset, condition=lambda x: True):
2905 2905 """
2906 2906 condition: a function that decide whether a revision in the subset
2907 2907 belongs to the revset or not.
2908 2908 """
2909 2909 self._subset = subset
2910 2910 self._condition = condition
2911 2911 self._cache = {}
2912 2912
2913 2913 def __contains__(self, x):
2914 2914 c = self._cache
2915 2915 if x not in c:
2916 2916 v = c[x] = x in self._subset and self._condition(x)
2917 2917 return v
2918 2918 return c[x]
2919 2919
2920 2920 def __iter__(self):
2921 2921 return self._iterfilter(self._subset)
2922 2922
2923 2923 def _iterfilter(self, it):
2924 2924 cond = self._condition
2925 2925 for x in it:
2926 2926 if cond(x):
2927 2927 yield x
2928 2928
2929 2929 @property
2930 2930 def fastasc(self):
2931 2931 it = self._subset.fastasc
2932 2932 if it is None:
2933 2933 return None
2934 2934 return lambda: self._iterfilter(it())
2935 2935
2936 2936 @property
2937 2937 def fastdesc(self):
2938 2938 it = self._subset.fastdesc
2939 2939 if it is None:
2940 2940 return None
2941 2941 return lambda: self._iterfilter(it())
2942 2942
2943 2943 def __nonzero__(self):
2944 2944 for r in self:
2945 2945 return True
2946 2946 return False
2947 2947
2948 2948 def __len__(self):
2949 2949 # Basic implementation to be changed in future patches.
2950 2950 l = baseset([r for r in self])
2951 2951 return len(l)
2952 2952
2953 2953 def sort(self, reverse=False):
2954 2954 self._subset.sort(reverse=reverse)
2955 2955
2956 2956 def reverse(self):
2957 2957 self._subset.reverse()
2958 2958
2959 2959 def isascending(self):
2960 2960 return self._subset.isascending()
2961 2961
2962 2962 def isdescending(self):
2963 2963 return self._subset.isdescending()
2964 2964
2965 2965 def first(self):
2966 2966 for x in self:
2967 2967 return x
2968 2968 return None
2969 2969
2970 2970 def last(self):
2971 2971 it = None
2972 2972 if self._subset.isascending:
2973 2973 it = self.fastdesc
2974 2974 elif self._subset.isdescending:
2975 2975 it = self.fastdesc
2976 2976 if it is None:
2977 2977 # slowly consume everything. This needs improvement
2978 2978 it = lambda: reversed(list(self))
2979 2979 for x in it():
2980 2980 return x
2981 2981 return None
2982 2982
2983 2983 def __repr__(self):
2984 2984 return '<%s %r>' % (type(self).__name__, self._subset)
2985 2985
2986 2986 # this function will be removed, or merged to addset or orset, when
2987 2987 # - scmutil.revrange() can be rewritten to not combine calculated smartsets
2988 2988 # - or addset can handle more than two sets without balanced tree
2989 2989 def _combinesets(subsets):
2990 2990 """Create balanced tree of addsets representing union of given sets"""
2991 2991 if not subsets:
2992 2992 return baseset()
2993 2993 if len(subsets) == 1:
2994 2994 return subsets[0]
2995 2995 p = len(subsets) // 2
2996 2996 xs = _combinesets(subsets[:p])
2997 2997 ys = _combinesets(subsets[p:])
2998 2998 return addset(xs, ys)
2999 2999
3000 3000 def _iterordered(ascending, iter1, iter2):
3001 3001 """produce an ordered iteration from two iterators with the same order
3002 3002
3003 3003 The ascending is used to indicated the iteration direction.
3004 3004 """
3005 3005 choice = max
3006 3006 if ascending:
3007 3007 choice = min
3008 3008
3009 3009 val1 = None
3010 3010 val2 = None
3011 3011 try:
3012 3012 # Consume both iterators in an ordered way until one is empty
3013 3013 while True:
3014 3014 if val1 is None:
3015 3015 val1 = iter1.next()
3016 3016 if val2 is None:
3017 3017 val2 = iter2.next()
3018 3018 next = choice(val1, val2)
3019 3019 yield next
3020 3020 if val1 == next:
3021 3021 val1 = None
3022 3022 if val2 == next:
3023 3023 val2 = None
3024 3024 except StopIteration:
3025 3025 # Flush any remaining values and consume the other one
3026 3026 it = iter2
3027 3027 if val1 is not None:
3028 3028 yield val1
3029 3029 it = iter1
3030 3030 elif val2 is not None:
3031 3031 # might have been equality and both are empty
3032 3032 yield val2
3033 3033 for val in it:
3034 3034 yield val
3035 3035
3036 3036 class addset(abstractsmartset):
3037 3037 """Represent the addition of two sets
3038 3038
3039 3039 Wrapper structure for lazily adding two structures without losing much
3040 3040 performance on the __contains__ method
3041 3041
3042 3042 If the ascending attribute is set, that means the two structures are
3043 3043 ordered in either an ascending or descending way. Therefore, we can add
3044 3044 them maintaining the order by iterating over both at the same time
3045 3045
3046 3046 >>> xs = baseset([0, 3, 2])
3047 3047 >>> ys = baseset([5, 2, 4])
3048 3048
3049 3049 >>> rs = addset(xs, ys)
3050 3050 >>> bool(rs), 0 in rs, 1 in rs, 5 in rs, rs.first(), rs.last()
3051 3051 (True, True, False, True, 0, 4)
3052 3052 >>> rs = addset(xs, baseset([]))
3053 3053 >>> bool(rs), 0 in rs, 1 in rs, rs.first(), rs.last()
3054 3054 (True, True, False, 0, 2)
3055 3055 >>> rs = addset(baseset([]), baseset([]))
3056 3056 >>> bool(rs), 0 in rs, rs.first(), rs.last()
3057 3057 (False, False, None, None)
3058 3058
3059 3059 iterate unsorted:
3060 3060 >>> rs = addset(xs, ys)
3061 3061 >>> [x for x in rs] # without _genlist
3062 3062 [0, 3, 2, 5, 4]
3063 3063 >>> assert not rs._genlist
3064 3064 >>> len(rs)
3065 3065 5
3066 3066 >>> [x for x in rs] # with _genlist
3067 3067 [0, 3, 2, 5, 4]
3068 3068 >>> assert rs._genlist
3069 3069
3070 3070 iterate ascending:
3071 3071 >>> rs = addset(xs, ys, ascending=True)
3072 3072 >>> [x for x in rs], [x for x in rs.fastasc()] # without _asclist
3073 3073 ([0, 2, 3, 4, 5], [0, 2, 3, 4, 5])
3074 3074 >>> assert not rs._asclist
3075 3075 >>> len(rs)
3076 3076 5
3077 3077 >>> [x for x in rs], [x for x in rs.fastasc()]
3078 3078 ([0, 2, 3, 4, 5], [0, 2, 3, 4, 5])
3079 3079 >>> assert rs._asclist
3080 3080
3081 3081 iterate descending:
3082 3082 >>> rs = addset(xs, ys, ascending=False)
3083 3083 >>> [x for x in rs], [x for x in rs.fastdesc()] # without _asclist
3084 3084 ([5, 4, 3, 2, 0], [5, 4, 3, 2, 0])
3085 3085 >>> assert not rs._asclist
3086 3086 >>> len(rs)
3087 3087 5
3088 3088 >>> [x for x in rs], [x for x in rs.fastdesc()]
3089 3089 ([5, 4, 3, 2, 0], [5, 4, 3, 2, 0])
3090 3090 >>> assert rs._asclist
3091 3091
3092 3092 iterate ascending without fastasc:
3093 3093 >>> rs = addset(xs, generatorset(ys), ascending=True)
3094 3094 >>> assert rs.fastasc is None
3095 3095 >>> [x for x in rs]
3096 3096 [0, 2, 3, 4, 5]
3097 3097
3098 3098 iterate descending without fastdesc:
3099 3099 >>> rs = addset(generatorset(xs), ys, ascending=False)
3100 3100 >>> assert rs.fastdesc is None
3101 3101 >>> [x for x in rs]
3102 3102 [5, 4, 3, 2, 0]
3103 3103 """
3104 3104 def __init__(self, revs1, revs2, ascending=None):
3105 3105 self._r1 = revs1
3106 3106 self._r2 = revs2
3107 3107 self._iter = None
3108 3108 self._ascending = ascending
3109 3109 self._genlist = None
3110 3110 self._asclist = None
3111 3111
3112 3112 def __len__(self):
3113 3113 return len(self._list)
3114 3114
3115 3115 def __nonzero__(self):
3116 3116 return bool(self._r1) or bool(self._r2)
3117 3117
3118 3118 @util.propertycache
3119 3119 def _list(self):
3120 3120 if not self._genlist:
3121 3121 self._genlist = baseset(iter(self))
3122 3122 return self._genlist
3123 3123
3124 3124 def __iter__(self):
3125 3125 """Iterate over both collections without repeating elements
3126 3126
3127 3127 If the ascending attribute is not set, iterate over the first one and
3128 3128 then over the second one checking for membership on the first one so we
3129 3129 dont yield any duplicates.
3130 3130
3131 3131 If the ascending attribute is set, iterate over both collections at the
3132 3132 same time, yielding only one value at a time in the given order.
3133 3133 """
3134 3134 if self._ascending is None:
3135 3135 if self._genlist:
3136 3136 return iter(self._genlist)
3137 3137 def arbitraryordergen():
3138 3138 for r in self._r1:
3139 3139 yield r
3140 3140 inr1 = self._r1.__contains__
3141 3141 for r in self._r2:
3142 3142 if not inr1(r):
3143 3143 yield r
3144 3144 return arbitraryordergen()
3145 3145 # try to use our own fast iterator if it exists
3146 3146 self._trysetasclist()
3147 3147 if self._ascending:
3148 3148 attr = 'fastasc'
3149 3149 else:
3150 3150 attr = 'fastdesc'
3151 3151 it = getattr(self, attr)
3152 3152 if it is not None:
3153 3153 return it()
3154 3154 # maybe half of the component supports fast
3155 3155 # get iterator for _r1
3156 3156 iter1 = getattr(self._r1, attr)
3157 3157 if iter1 is None:
3158 3158 # let's avoid side effect (not sure it matters)
3159 3159 iter1 = iter(sorted(self._r1, reverse=not self._ascending))
3160 3160 else:
3161 3161 iter1 = iter1()
3162 3162 # get iterator for _r2
3163 3163 iter2 = getattr(self._r2, attr)
3164 3164 if iter2 is None:
3165 3165 # let's avoid side effect (not sure it matters)
3166 3166 iter2 = iter(sorted(self._r2, reverse=not self._ascending))
3167 3167 else:
3168 3168 iter2 = iter2()
3169 3169 return _iterordered(self._ascending, iter1, iter2)
3170 3170
3171 3171 def _trysetasclist(self):
3172 3172 """populate the _asclist attribute if possible and necessary"""
3173 3173 if self._genlist is not None and self._asclist is None:
3174 3174 self._asclist = sorted(self._genlist)
3175 3175
3176 3176 @property
3177 3177 def fastasc(self):
3178 3178 self._trysetasclist()
3179 3179 if self._asclist is not None:
3180 3180 return self._asclist.__iter__
3181 3181 iter1 = self._r1.fastasc
3182 3182 iter2 = self._r2.fastasc
3183 3183 if None in (iter1, iter2):
3184 3184 return None
3185 3185 return lambda: _iterordered(True, iter1(), iter2())
3186 3186
3187 3187 @property
3188 3188 def fastdesc(self):
3189 3189 self._trysetasclist()
3190 3190 if self._asclist is not None:
3191 3191 return self._asclist.__reversed__
3192 3192 iter1 = self._r1.fastdesc
3193 3193 iter2 = self._r2.fastdesc
3194 3194 if None in (iter1, iter2):
3195 3195 return None
3196 3196 return lambda: _iterordered(False, iter1(), iter2())
3197 3197
3198 3198 def __contains__(self, x):
3199 3199 return x in self._r1 or x in self._r2
3200 3200
3201 3201 def sort(self, reverse=False):
3202 3202 """Sort the added set
3203 3203
3204 3204 For this we use the cached list with all the generated values and if we
3205 3205 know they are ascending or descending we can sort them in a smart way.
3206 3206 """
3207 3207 self._ascending = not reverse
3208 3208
3209 3209 def isascending(self):
3210 3210 return self._ascending is not None and self._ascending
3211 3211
3212 3212 def isdescending(self):
3213 3213 return self._ascending is not None and not self._ascending
3214 3214
3215 3215 def reverse(self):
3216 3216 if self._ascending is None:
3217 3217 self._list.reverse()
3218 3218 else:
3219 3219 self._ascending = not self._ascending
3220 3220
3221 3221 def first(self):
3222 3222 for x in self:
3223 3223 return x
3224 3224 return None
3225 3225
3226 3226 def last(self):
3227 3227 self.reverse()
3228 3228 val = self.first()
3229 3229 self.reverse()
3230 3230 return val
3231 3231
3232 3232 def __repr__(self):
3233 3233 d = {None: '', False: '-', True: '+'}[self._ascending]
3234 3234 return '<%s%s %r, %r>' % (type(self).__name__, d, self._r1, self._r2)
3235 3235
3236 3236 class generatorset(abstractsmartset):
3237 3237 """Wrap a generator for lazy iteration
3238 3238
3239 3239 Wrapper structure for generators that provides lazy membership and can
3240 3240 be iterated more than once.
3241 3241 When asked for membership it generates values until either it finds the
3242 3242 requested one or has gone through all the elements in the generator
3243 3243 """
3244 3244 def __init__(self, gen, iterasc=None):
3245 3245 """
3246 3246 gen: a generator producing the values for the generatorset.
3247 3247 """
3248 3248 self._gen = gen
3249 3249 self._asclist = None
3250 3250 self._cache = {}
3251 3251 self._genlist = []
3252 3252 self._finished = False
3253 3253 self._ascending = True
3254 3254 if iterasc is not None:
3255 3255 if iterasc:
3256 3256 self.fastasc = self._iterator
3257 3257 self.__contains__ = self._asccontains
3258 3258 else:
3259 3259 self.fastdesc = self._iterator
3260 3260 self.__contains__ = self._desccontains
3261 3261
3262 3262 def __nonzero__(self):
3263 3263 # Do not use 'for r in self' because it will enforce the iteration
3264 3264 # order (default ascending), possibly unrolling a whole descending
3265 3265 # iterator.
3266 3266 if self._genlist:
3267 3267 return True
3268 3268 for r in self._consumegen():
3269 3269 return True
3270 3270 return False
3271 3271
3272 3272 def __contains__(self, x):
3273 3273 if x in self._cache:
3274 3274 return self._cache[x]
3275 3275
3276 3276 # Use new values only, as existing values would be cached.
3277 3277 for l in self._consumegen():
3278 3278 if l == x:
3279 3279 return True
3280 3280
3281 3281 self._cache[x] = False
3282 3282 return False
3283 3283
3284 3284 def _asccontains(self, x):
3285 3285 """version of contains optimised for ascending generator"""
3286 3286 if x in self._cache:
3287 3287 return self._cache[x]
3288 3288
3289 3289 # Use new values only, as existing values would be cached.
3290 3290 for l in self._consumegen():
3291 3291 if l == x:
3292 3292 return True
3293 3293 if l > x:
3294 3294 break
3295 3295
3296 3296 self._cache[x] = False
3297 3297 return False
3298 3298
3299 3299 def _desccontains(self, x):
3300 3300 """version of contains optimised for descending generator"""
3301 3301 if x in self._cache:
3302 3302 return self._cache[x]
3303 3303
3304 3304 # Use new values only, as existing values would be cached.
3305 3305 for l in self._consumegen():
3306 3306 if l == x:
3307 3307 return True
3308 3308 if l < x:
3309 3309 break
3310 3310
3311 3311 self._cache[x] = False
3312 3312 return False
3313 3313
3314 3314 def __iter__(self):
3315 3315 if self._ascending:
3316 3316 it = self.fastasc
3317 3317 else:
3318 3318 it = self.fastdesc
3319 3319 if it is not None:
3320 3320 return it()
3321 3321 # we need to consume the iterator
3322 3322 for x in self._consumegen():
3323 3323 pass
3324 3324 # recall the same code
3325 3325 return iter(self)
3326 3326
3327 3327 def _iterator(self):
3328 3328 if self._finished:
3329 3329 return iter(self._genlist)
3330 3330
3331 3331 # We have to use this complex iteration strategy to allow multiple
3332 3332 # iterations at the same time. We need to be able to catch revision
3333 3333 # removed from _consumegen and added to genlist in another instance.
3334 3334 #
3335 3335 # Getting rid of it would provide an about 15% speed up on this
3336 3336 # iteration.
3337 3337 genlist = self._genlist
3338 3338 nextrev = self._consumegen().next
3339 3339 _len = len # cache global lookup
3340 3340 def gen():
3341 3341 i = 0
3342 3342 while True:
3343 3343 if i < _len(genlist):
3344 3344 yield genlist[i]
3345 3345 else:
3346 3346 yield nextrev()
3347 3347 i += 1
3348 3348 return gen()
3349 3349
3350 3350 def _consumegen(self):
3351 3351 cache = self._cache
3352 3352 genlist = self._genlist.append
3353 3353 for item in self._gen:
3354 3354 cache[item] = True
3355 3355 genlist(item)
3356 3356 yield item
3357 3357 if not self._finished:
3358 3358 self._finished = True
3359 3359 asc = self._genlist[:]
3360 3360 asc.sort()
3361 3361 self._asclist = asc
3362 3362 self.fastasc = asc.__iter__
3363 3363 self.fastdesc = asc.__reversed__
3364 3364
3365 3365 def __len__(self):
3366 3366 for x in self._consumegen():
3367 3367 pass
3368 3368 return len(self._genlist)
3369 3369
3370 3370 def sort(self, reverse=False):
3371 3371 self._ascending = not reverse
3372 3372
3373 3373 def reverse(self):
3374 3374 self._ascending = not self._ascending
3375 3375
3376 3376 def isascending(self):
3377 3377 return self._ascending
3378 3378
3379 3379 def isdescending(self):
3380 3380 return not self._ascending
3381 3381
3382 3382 def first(self):
3383 3383 if self._ascending:
3384 3384 it = self.fastasc
3385 3385 else:
3386 3386 it = self.fastdesc
3387 3387 if it is None:
3388 3388 # we need to consume all and try again
3389 3389 for x in self._consumegen():
3390 3390 pass
3391 3391 return self.first()
3392 3392 return next(it(), None)
3393 3393
3394 3394 def last(self):
3395 3395 if self._ascending:
3396 3396 it = self.fastdesc
3397 3397 else:
3398 3398 it = self.fastasc
3399 3399 if it is None:
3400 3400 # we need to consume all and try again
3401 3401 for x in self._consumegen():
3402 3402 pass
3403 3403 return self.first()
3404 3404 return next(it(), None)
3405 3405
3406 3406 def __repr__(self):
3407 3407 d = {False: '-', True: '+'}[self._ascending]
3408 3408 return '<%s%s>' % (type(self).__name__, d)
3409 3409
3410 3410 class spanset(abstractsmartset):
3411 3411 """Duck type for baseset class which represents a range of revisions and
3412 3412 can work lazily and without having all the range in memory
3413 3413
3414 3414 Note that spanset(x, y) behave almost like xrange(x, y) except for two
3415 3415 notable points:
3416 3416 - when x < y it will be automatically descending,
3417 3417 - revision filtered with this repoview will be skipped.
3418 3418
3419 3419 """
3420 3420 def __init__(self, repo, start=0, end=None):
3421 3421 """
3422 3422 start: first revision included the set
3423 3423 (default to 0)
3424 3424 end: first revision excluded (last+1)
3425 3425 (default to len(repo)
3426 3426
3427 3427 Spanset will be descending if `end` < `start`.
3428 3428 """
3429 3429 if end is None:
3430 3430 end = len(repo)
3431 3431 self._ascending = start <= end
3432 3432 if not self._ascending:
3433 3433 start, end = end + 1, start +1
3434 3434 self._start = start
3435 3435 self._end = end
3436 3436 self._hiddenrevs = repo.changelog.filteredrevs
3437 3437
3438 3438 def sort(self, reverse=False):
3439 3439 self._ascending = not reverse
3440 3440
3441 3441 def reverse(self):
3442 3442 self._ascending = not self._ascending
3443 3443
3444 3444 def _iterfilter(self, iterrange):
3445 3445 s = self._hiddenrevs
3446 3446 for r in iterrange:
3447 3447 if r not in s:
3448 3448 yield r
3449 3449
3450 3450 def __iter__(self):
3451 3451 if self._ascending:
3452 3452 return self.fastasc()
3453 3453 else:
3454 3454 return self.fastdesc()
3455 3455
3456 3456 def fastasc(self):
3457 3457 iterrange = xrange(self._start, self._end)
3458 3458 if self._hiddenrevs:
3459 3459 return self._iterfilter(iterrange)
3460 3460 return iter(iterrange)
3461 3461
3462 3462 def fastdesc(self):
3463 3463 iterrange = xrange(self._end - 1, self._start - 1, -1)
3464 3464 if self._hiddenrevs:
3465 3465 return self._iterfilter(iterrange)
3466 3466 return iter(iterrange)
3467 3467
3468 3468 def __contains__(self, rev):
3469 3469 hidden = self._hiddenrevs
3470 3470 return ((self._start <= rev < self._end)
3471 3471 and not (hidden and rev in hidden))
3472 3472
3473 3473 def __nonzero__(self):
3474 3474 for r in self:
3475 3475 return True
3476 3476 return False
3477 3477
3478 3478 def __len__(self):
3479 3479 if not self._hiddenrevs:
3480 3480 return abs(self._end - self._start)
3481 3481 else:
3482 3482 count = 0
3483 3483 start = self._start
3484 3484 end = self._end
3485 3485 for rev in self._hiddenrevs:
3486 3486 if (end < rev <= start) or (start <= rev < end):
3487 3487 count += 1
3488 3488 return abs(self._end - self._start) - count
3489 3489
3490 3490 def isascending(self):
3491 3491 return self._ascending
3492 3492
3493 3493 def isdescending(self):
3494 3494 return not self._ascending
3495 3495
3496 3496 def first(self):
3497 3497 if self._ascending:
3498 3498 it = self.fastasc
3499 3499 else:
3500 3500 it = self.fastdesc
3501 3501 for x in it():
3502 3502 return x
3503 3503 return None
3504 3504
3505 3505 def last(self):
3506 3506 if self._ascending:
3507 3507 it = self.fastdesc
3508 3508 else:
3509 3509 it = self.fastasc
3510 3510 for x in it():
3511 3511 return x
3512 3512 return None
3513 3513
3514 3514 def __repr__(self):
3515 3515 d = {False: '-', True: '+'}[self._ascending]
3516 3516 return '<%s%s %d:%d>' % (type(self).__name__, d,
3517 3517 self._start, self._end - 1)
3518 3518
3519 3519 class fullreposet(spanset):
3520 3520 """a set containing all revisions in the repo
3521 3521
3522 3522 This class exists to host special optimization and magic to handle virtual
3523 3523 revisions such as "null".
3524 3524 """
3525 3525
3526 3526 def __init__(self, repo):
3527 3527 super(fullreposet, self).__init__(repo)
3528 3528
3529 3529 def __and__(self, other):
3530 3530 """As self contains the whole repo, all of the other set should also be
3531 3531 in self. Therefore `self & other = other`.
3532 3532
3533 3533 This boldly assumes the other contains valid revs only.
3534 3534 """
3535 3535 # other not a smartset, make is so
3536 3536 if not util.safehasattr(other, 'isascending'):
3537 3537 # filter out hidden revision
3538 3538 # (this boldly assumes all smartset are pure)
3539 3539 #
3540 3540 # `other` was used with "&", let's assume this is a set like
3541 3541 # object.
3542 3542 other = baseset(other - self._hiddenrevs)
3543 3543
3544 # XXX As fullreposet is also used as bootstrap, this is wrong.
3545 #
3546 # With a giveme312() revset returning [3,1,2], this makes
3547 # 'hg log -r "giveme312()"' -> 1, 2, 3 (wrong)
3548 # We cannot just drop it because other usage still need to sort it:
3549 # 'hg log -r "all() and giveme312()"' -> 1, 2, 3 (right)
3550 #
3551 # There is also some faulty revset implementations that rely on it
3552 # (eg: children as of its state in e8075329c5fb)
3553 #
3554 # When we fix the two points above we can move this into the if clause
3544 3555 other.sort(reverse=self.isdescending())
3545 3556 return other
3546 3557
3547 3558 def prettyformatset(revs):
3548 3559 lines = []
3549 3560 rs = repr(revs)
3550 3561 p = 0
3551 3562 while p < len(rs):
3552 3563 q = rs.find('<', p + 1)
3553 3564 if q < 0:
3554 3565 q = len(rs)
3555 3566 l = rs.count('<', 0, p) - rs.count('>', 0, p)
3556 3567 assert l >= 0
3557 3568 lines.append((l, rs[p:q].rstrip()))
3558 3569 p = q
3559 3570 return '\n'.join(' ' * l + s for l, s in lines)
3560 3571
3561 3572 # tell hggettext to extract docstrings from these functions:
3562 3573 i18nfunctions = symbols.values()
General Comments 0
You need to be logged in to leave comments. Login now