##// END OF EJS Templates
revset: fix iteration over ordered addset composed of non-ordered operands...
Pierre-Yves David -
r25115:5548f558 default
parent child Browse files
Show More
@@ -1,3504 +1,3512
1 1 # revset.py - revision set queries for mercurial
2 2 #
3 3 # Copyright 2010 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 import re
9 9 import parser, util, error, hbisect, phases
10 10 import node
11 11 import heapq
12 12 import match as matchmod
13 13 from i18n import _
14 14 import encoding
15 15 import obsolete as obsmod
16 16 import pathutil
17 17 import repoview
18 18
19 19 def _revancestors(repo, revs, followfirst):
20 20 """Like revlog.ancestors(), but supports followfirst."""
21 21 if followfirst:
22 22 cut = 1
23 23 else:
24 24 cut = None
25 25 cl = repo.changelog
26 26
27 27 def iterate():
28 28 revs.sort(reverse=True)
29 29 irevs = iter(revs)
30 30 h = []
31 31 try:
32 32 inputrev = irevs.next()
33 33 heapq.heappush(h, -inputrev)
34 34 except StopIteration:
35 35 return
36 36
37 37 seen = set()
38 38 while h:
39 39 current = -heapq.heappop(h)
40 40 if current == inputrev:
41 41 try:
42 42 inputrev = irevs.next()
43 43 heapq.heappush(h, -inputrev)
44 44 except StopIteration:
45 45 pass
46 46 if current not in seen:
47 47 seen.add(current)
48 48 yield current
49 49 for parent in cl.parentrevs(current)[:cut]:
50 50 if parent != node.nullrev:
51 51 heapq.heappush(h, -parent)
52 52
53 53 return generatorset(iterate(), iterasc=False)
54 54
55 55 def _revdescendants(repo, revs, followfirst):
56 56 """Like revlog.descendants() but supports followfirst."""
57 57 if followfirst:
58 58 cut = 1
59 59 else:
60 60 cut = None
61 61
62 62 def iterate():
63 63 cl = repo.changelog
64 64 first = min(revs)
65 65 nullrev = node.nullrev
66 66 if first == nullrev:
67 67 # Are there nodes with a null first parent and a non-null
68 68 # second one? Maybe. Do we care? Probably not.
69 69 for i in cl:
70 70 yield i
71 71 else:
72 72 seen = set(revs)
73 73 for i in cl.revs(first + 1):
74 74 for x in cl.parentrevs(i)[:cut]:
75 75 if x != nullrev and x in seen:
76 76 seen.add(i)
77 77 yield i
78 78 break
79 79
80 80 return generatorset(iterate(), iterasc=True)
81 81
82 82 def _revsbetween(repo, roots, heads):
83 83 """Return all paths between roots and heads, inclusive of both endpoint
84 84 sets."""
85 85 if not roots:
86 86 return baseset()
87 87 parentrevs = repo.changelog.parentrevs
88 88 visit = list(heads)
89 89 reachable = set()
90 90 seen = {}
91 91 minroot = min(roots)
92 92 roots = set(roots)
93 93 # open-code the post-order traversal due to the tiny size of
94 94 # sys.getrecursionlimit()
95 95 while visit:
96 96 rev = visit.pop()
97 97 if rev in roots:
98 98 reachable.add(rev)
99 99 parents = parentrevs(rev)
100 100 seen[rev] = parents
101 101 for parent in parents:
102 102 if parent >= minroot and parent not in seen:
103 103 visit.append(parent)
104 104 if not reachable:
105 105 return baseset()
106 106 for rev in sorted(seen):
107 107 for parent in seen[rev]:
108 108 if parent in reachable:
109 109 reachable.add(rev)
110 110 return baseset(sorted(reachable))
111 111
112 112 elements = {
113 113 "(": (21, ("group", 1, ")"), ("func", 1, ")")),
114 114 "##": (20, None, ("_concat", 20)),
115 115 "~": (18, None, ("ancestor", 18)),
116 116 "^": (18, None, ("parent", 18), ("parentpost", 18)),
117 117 "-": (5, ("negate", 19), ("minus", 5)),
118 118 "::": (17, ("dagrangepre", 17), ("dagrange", 17),
119 119 ("dagrangepost", 17)),
120 120 "..": (17, ("dagrangepre", 17), ("dagrange", 17),
121 121 ("dagrangepost", 17)),
122 122 ":": (15, ("rangepre", 15), ("range", 15), ("rangepost", 15)),
123 123 "not": (10, ("not", 10)),
124 124 "!": (10, ("not", 10)),
125 125 "and": (5, None, ("and", 5)),
126 126 "&": (5, None, ("and", 5)),
127 127 "%": (5, None, ("only", 5), ("onlypost", 5)),
128 128 "or": (4, None, ("or", 4)),
129 129 "|": (4, None, ("or", 4)),
130 130 "+": (4, None, ("or", 4)),
131 131 ",": (2, None, ("list", 2)),
132 132 ")": (0, None, None),
133 133 "symbol": (0, ("symbol",), None),
134 134 "string": (0, ("string",), None),
135 135 "end": (0, None, None),
136 136 }
137 137
138 138 keywords = set(['and', 'or', 'not'])
139 139
140 140 # default set of valid characters for the initial letter of symbols
141 141 _syminitletters = set(c for c in [chr(i) for i in xrange(256)]
142 142 if c.isalnum() or c in '._@' or ord(c) > 127)
143 143
144 144 # default set of valid characters for non-initial letters of symbols
145 145 _symletters = set(c for c in [chr(i) for i in xrange(256)]
146 146 if c.isalnum() or c in '-._/@' or ord(c) > 127)
147 147
148 148 def tokenize(program, lookup=None, syminitletters=None, symletters=None):
149 149 '''
150 150 Parse a revset statement into a stream of tokens
151 151
152 152 ``syminitletters`` is the set of valid characters for the initial
153 153 letter of symbols.
154 154
155 155 By default, character ``c`` is recognized as valid for initial
156 156 letter of symbols, if ``c.isalnum() or c in '._@' or ord(c) > 127``.
157 157
158 158 ``symletters`` is the set of valid characters for non-initial
159 159 letters of symbols.
160 160
161 161 By default, character ``c`` is recognized as valid for non-initial
162 162 letters of symbols, if ``c.isalnum() or c in '-._/@' or ord(c) > 127``.
163 163
164 164 Check that @ is a valid unquoted token character (issue3686):
165 165 >>> list(tokenize("@::"))
166 166 [('symbol', '@', 0), ('::', None, 1), ('end', None, 3)]
167 167
168 168 '''
169 169 if syminitletters is None:
170 170 syminitletters = _syminitletters
171 171 if symletters is None:
172 172 symletters = _symletters
173 173
174 174 pos, l = 0, len(program)
175 175 while pos < l:
176 176 c = program[pos]
177 177 if c.isspace(): # skip inter-token whitespace
178 178 pass
179 179 elif c == ':' and program[pos:pos + 2] == '::': # look ahead carefully
180 180 yield ('::', None, pos)
181 181 pos += 1 # skip ahead
182 182 elif c == '.' and program[pos:pos + 2] == '..': # look ahead carefully
183 183 yield ('..', None, pos)
184 184 pos += 1 # skip ahead
185 185 elif c == '#' and program[pos:pos + 2] == '##': # look ahead carefully
186 186 yield ('##', None, pos)
187 187 pos += 1 # skip ahead
188 188 elif c in "():,-|&+!~^%": # handle simple operators
189 189 yield (c, None, pos)
190 190 elif (c in '"\'' or c == 'r' and
191 191 program[pos:pos + 2] in ("r'", 'r"')): # handle quoted strings
192 192 if c == 'r':
193 193 pos += 1
194 194 c = program[pos]
195 195 decode = lambda x: x
196 196 else:
197 197 decode = lambda x: x.decode('string-escape')
198 198 pos += 1
199 199 s = pos
200 200 while pos < l: # find closing quote
201 201 d = program[pos]
202 202 if d == '\\': # skip over escaped characters
203 203 pos += 2
204 204 continue
205 205 if d == c:
206 206 yield ('string', decode(program[s:pos]), s)
207 207 break
208 208 pos += 1
209 209 else:
210 210 raise error.ParseError(_("unterminated string"), s)
211 211 # gather up a symbol/keyword
212 212 elif c in syminitletters:
213 213 s = pos
214 214 pos += 1
215 215 while pos < l: # find end of symbol
216 216 d = program[pos]
217 217 if d not in symletters:
218 218 break
219 219 if d == '.' and program[pos - 1] == '.': # special case for ..
220 220 pos -= 1
221 221 break
222 222 pos += 1
223 223 sym = program[s:pos]
224 224 if sym in keywords: # operator keywords
225 225 yield (sym, None, s)
226 226 elif '-' in sym:
227 227 # some jerk gave us foo-bar-baz, try to check if it's a symbol
228 228 if lookup and lookup(sym):
229 229 # looks like a real symbol
230 230 yield ('symbol', sym, s)
231 231 else:
232 232 # looks like an expression
233 233 parts = sym.split('-')
234 234 for p in parts[:-1]:
235 235 if p: # possible consecutive -
236 236 yield ('symbol', p, s)
237 237 s += len(p)
238 238 yield ('-', None, pos)
239 239 s += 1
240 240 if parts[-1]: # possible trailing -
241 241 yield ('symbol', parts[-1], s)
242 242 else:
243 243 yield ('symbol', sym, s)
244 244 pos -= 1
245 245 else:
246 246 raise error.ParseError(_("syntax error in revset '%s'") %
247 247 program, pos)
248 248 pos += 1
249 249 yield ('end', None, pos)
250 250
251 251 def parseerrordetail(inst):
252 252 """Compose error message from specified ParseError object
253 253 """
254 254 if len(inst.args) > 1:
255 255 return _('at %s: %s') % (inst.args[1], inst.args[0])
256 256 else:
257 257 return inst.args[0]
258 258
259 259 # helpers
260 260
261 261 def getstring(x, err):
262 262 if x and (x[0] == 'string' or x[0] == 'symbol'):
263 263 return x[1]
264 264 raise error.ParseError(err)
265 265
266 266 def getlist(x):
267 267 if not x:
268 268 return []
269 269 if x[0] == 'list':
270 270 return getlist(x[1]) + [x[2]]
271 271 return [x]
272 272
273 273 def getargs(x, min, max, err):
274 274 l = getlist(x)
275 275 if len(l) < min or (max >= 0 and len(l) > max):
276 276 raise error.ParseError(err)
277 277 return l
278 278
279 279 def isvalidsymbol(tree):
280 280 """Examine whether specified ``tree`` is valid ``symbol`` or not
281 281 """
282 282 return tree[0] == 'symbol' and len(tree) > 1
283 283
284 284 def getsymbol(tree):
285 285 """Get symbol name from valid ``symbol`` in ``tree``
286 286
287 287 This assumes that ``tree`` is already examined by ``isvalidsymbol``.
288 288 """
289 289 return tree[1]
290 290
291 291 def isvalidfunc(tree):
292 292 """Examine whether specified ``tree`` is valid ``func`` or not
293 293 """
294 294 return tree[0] == 'func' and len(tree) > 1 and isvalidsymbol(tree[1])
295 295
296 296 def getfuncname(tree):
297 297 """Get function name from valid ``func`` in ``tree``
298 298
299 299 This assumes that ``tree`` is already examined by ``isvalidfunc``.
300 300 """
301 301 return getsymbol(tree[1])
302 302
303 303 def getfuncargs(tree):
304 304 """Get list of function arguments from valid ``func`` in ``tree``
305 305
306 306 This assumes that ``tree`` is already examined by ``isvalidfunc``.
307 307 """
308 308 if len(tree) > 2:
309 309 return getlist(tree[2])
310 310 else:
311 311 return []
312 312
313 313 def getset(repo, subset, x):
314 314 if not x:
315 315 raise error.ParseError(_("missing argument"))
316 316 s = methods[x[0]](repo, subset, *x[1:])
317 317 if util.safehasattr(s, 'isascending'):
318 318 return s
319 319 return baseset(s)
320 320
321 321 def _getrevsource(repo, r):
322 322 extra = repo[r].extra()
323 323 for label in ('source', 'transplant_source', 'rebase_source'):
324 324 if label in extra:
325 325 try:
326 326 return repo[extra[label]].rev()
327 327 except error.RepoLookupError:
328 328 pass
329 329 return None
330 330
331 331 # operator methods
332 332
333 333 def stringset(repo, subset, x):
334 334 x = repo[x].rev()
335 335 if x in subset:
336 336 return baseset([x])
337 337 return baseset()
338 338
339 339 def rangeset(repo, subset, x, y):
340 340 m = getset(repo, fullreposet(repo), x)
341 341 n = getset(repo, fullreposet(repo), y)
342 342
343 343 if not m or not n:
344 344 return baseset()
345 345 m, n = m.first(), n.last()
346 346
347 347 if m < n:
348 348 r = spanset(repo, m, n + 1)
349 349 else:
350 350 r = spanset(repo, m, n - 1)
351 351 return r & subset
352 352
353 353 def dagrange(repo, subset, x, y):
354 354 r = fullreposet(repo)
355 355 xs = _revsbetween(repo, getset(repo, r, x), getset(repo, r, y))
356 356 return xs & subset
357 357
358 358 def andset(repo, subset, x, y):
359 359 return getset(repo, getset(repo, subset, x), y)
360 360
361 361 def orset(repo, subset, x, y):
362 362 xl = getset(repo, subset, x)
363 363 yl = getset(repo, subset - xl, y)
364 364 return xl + yl
365 365
366 366 def notset(repo, subset, x):
367 367 return subset - getset(repo, subset, x)
368 368
369 369 def listset(repo, subset, a, b):
370 370 raise error.ParseError(_("can't use a list in this context"))
371 371
372 372 def func(repo, subset, a, b):
373 373 if a[0] == 'symbol' and a[1] in symbols:
374 374 return symbols[a[1]](repo, subset, b)
375 375 raise error.UnknownIdentifier(a[1], symbols.keys())
376 376
377 377 # functions
378 378
379 379 def adds(repo, subset, x):
380 380 """``adds(pattern)``
381 381 Changesets that add a file matching pattern.
382 382
383 383 The pattern without explicit kind like ``glob:`` is expected to be
384 384 relative to the current directory and match against a file or a
385 385 directory.
386 386 """
387 387 # i18n: "adds" is a keyword
388 388 pat = getstring(x, _("adds requires a pattern"))
389 389 return checkstatus(repo, subset, pat, 1)
390 390
391 391 def ancestor(repo, subset, x):
392 392 """``ancestor(*changeset)``
393 393 A greatest common ancestor of the changesets.
394 394
395 395 Accepts 0 or more changesets.
396 396 Will return empty list when passed no args.
397 397 Greatest common ancestor of a single changeset is that changeset.
398 398 """
399 399 # i18n: "ancestor" is a keyword
400 400 l = getlist(x)
401 401 rl = fullreposet(repo)
402 402 anc = None
403 403
404 404 # (getset(repo, rl, i) for i in l) generates a list of lists
405 405 for revs in (getset(repo, rl, i) for i in l):
406 406 for r in revs:
407 407 if anc is None:
408 408 anc = repo[r]
409 409 else:
410 410 anc = anc.ancestor(repo[r])
411 411
412 412 if anc is not None and anc.rev() in subset:
413 413 return baseset([anc.rev()])
414 414 return baseset()
415 415
416 416 def _ancestors(repo, subset, x, followfirst=False):
417 417 heads = getset(repo, fullreposet(repo), x)
418 418 if not heads:
419 419 return baseset()
420 420 s = _revancestors(repo, heads, followfirst)
421 421 return subset & s
422 422
423 423 def ancestors(repo, subset, x):
424 424 """``ancestors(set)``
425 425 Changesets that are ancestors of a changeset in set.
426 426 """
427 427 return _ancestors(repo, subset, x)
428 428
429 429 def _firstancestors(repo, subset, x):
430 430 # ``_firstancestors(set)``
431 431 # Like ``ancestors(set)`` but follows only the first parents.
432 432 return _ancestors(repo, subset, x, followfirst=True)
433 433
434 434 def ancestorspec(repo, subset, x, n):
435 435 """``set~n``
436 436 Changesets that are the Nth ancestor (first parents only) of a changeset
437 437 in set.
438 438 """
439 439 try:
440 440 n = int(n[1])
441 441 except (TypeError, ValueError):
442 442 raise error.ParseError(_("~ expects a number"))
443 443 ps = set()
444 444 cl = repo.changelog
445 445 for r in getset(repo, fullreposet(repo), x):
446 446 for i in range(n):
447 447 r = cl.parentrevs(r)[0]
448 448 ps.add(r)
449 449 return subset & ps
450 450
451 451 def author(repo, subset, x):
452 452 """``author(string)``
453 453 Alias for ``user(string)``.
454 454 """
455 455 # i18n: "author" is a keyword
456 456 n = encoding.lower(getstring(x, _("author requires a string")))
457 457 kind, pattern, matcher = _substringmatcher(n)
458 458 return subset.filter(lambda x: matcher(encoding.lower(repo[x].user())))
459 459
460 460 def bisect(repo, subset, x):
461 461 """``bisect(string)``
462 462 Changesets marked in the specified bisect status:
463 463
464 464 - ``good``, ``bad``, ``skip``: csets explicitly marked as good/bad/skip
465 465 - ``goods``, ``bads`` : csets topologically good/bad
466 466 - ``range`` : csets taking part in the bisection
467 467 - ``pruned`` : csets that are goods, bads or skipped
468 468 - ``untested`` : csets whose fate is yet unknown
469 469 - ``ignored`` : csets ignored due to DAG topology
470 470 - ``current`` : the cset currently being bisected
471 471 """
472 472 # i18n: "bisect" is a keyword
473 473 status = getstring(x, _("bisect requires a string")).lower()
474 474 state = set(hbisect.get(repo, status))
475 475 return subset & state
476 476
477 477 # Backward-compatibility
478 478 # - no help entry so that we do not advertise it any more
479 479 def bisected(repo, subset, x):
480 480 return bisect(repo, subset, x)
481 481
482 482 def bookmark(repo, subset, x):
483 483 """``bookmark([name])``
484 484 The named bookmark or all bookmarks.
485 485
486 486 If `name` starts with `re:`, the remainder of the name is treated as
487 487 a regular expression. To match a bookmark that actually starts with `re:`,
488 488 use the prefix `literal:`.
489 489 """
490 490 # i18n: "bookmark" is a keyword
491 491 args = getargs(x, 0, 1, _('bookmark takes one or no arguments'))
492 492 if args:
493 493 bm = getstring(args[0],
494 494 # i18n: "bookmark" is a keyword
495 495 _('the argument to bookmark must be a string'))
496 496 kind, pattern, matcher = _stringmatcher(bm)
497 497 bms = set()
498 498 if kind == 'literal':
499 499 bmrev = repo._bookmarks.get(pattern, None)
500 500 if not bmrev:
501 501 raise error.RepoLookupError(_("bookmark '%s' does not exist")
502 502 % bm)
503 503 bms.add(repo[bmrev].rev())
504 504 else:
505 505 matchrevs = set()
506 506 for name, bmrev in repo._bookmarks.iteritems():
507 507 if matcher(name):
508 508 matchrevs.add(bmrev)
509 509 if not matchrevs:
510 510 raise error.RepoLookupError(_("no bookmarks exist"
511 511 " that match '%s'") % pattern)
512 512 for bmrev in matchrevs:
513 513 bms.add(repo[bmrev].rev())
514 514 else:
515 515 bms = set([repo[r].rev()
516 516 for r in repo._bookmarks.values()])
517 517 bms -= set([node.nullrev])
518 518 return subset & bms
519 519
520 520 def branch(repo, subset, x):
521 521 """``branch(string or set)``
522 522 All changesets belonging to the given branch or the branches of the given
523 523 changesets.
524 524
525 525 If `string` starts with `re:`, the remainder of the name is treated as
526 526 a regular expression. To match a branch that actually starts with `re:`,
527 527 use the prefix `literal:`.
528 528 """
529 529 getbi = repo.revbranchcache().branchinfo
530 530
531 531 try:
532 532 b = getstring(x, '')
533 533 except error.ParseError:
534 534 # not a string, but another revspec, e.g. tip()
535 535 pass
536 536 else:
537 537 kind, pattern, matcher = _stringmatcher(b)
538 538 if kind == 'literal':
539 539 # note: falls through to the revspec case if no branch with
540 540 # this name exists
541 541 if pattern in repo.branchmap():
542 542 return subset.filter(lambda r: matcher(getbi(r)[0]))
543 543 else:
544 544 return subset.filter(lambda r: matcher(getbi(r)[0]))
545 545
546 546 s = getset(repo, fullreposet(repo), x)
547 547 b = set()
548 548 for r in s:
549 549 b.add(getbi(r)[0])
550 550 c = s.__contains__
551 551 return subset.filter(lambda r: c(r) or getbi(r)[0] in b)
552 552
553 553 def bumped(repo, subset, x):
554 554 """``bumped()``
555 555 Mutable changesets marked as successors of public changesets.
556 556
557 557 Only non-public and non-obsolete changesets can be `bumped`.
558 558 """
559 559 # i18n: "bumped" is a keyword
560 560 getargs(x, 0, 0, _("bumped takes no arguments"))
561 561 bumped = obsmod.getrevs(repo, 'bumped')
562 562 return subset & bumped
563 563
564 564 def bundle(repo, subset, x):
565 565 """``bundle()``
566 566 Changesets in the bundle.
567 567
568 568 Bundle must be specified by the -R option."""
569 569
570 570 try:
571 571 bundlerevs = repo.changelog.bundlerevs
572 572 except AttributeError:
573 573 raise util.Abort(_("no bundle provided - specify with -R"))
574 574 return subset & bundlerevs
575 575
576 576 def checkstatus(repo, subset, pat, field):
577 577 hasset = matchmod.patkind(pat) == 'set'
578 578
579 579 mcache = [None]
580 580 def matches(x):
581 581 c = repo[x]
582 582 if not mcache[0] or hasset:
583 583 mcache[0] = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
584 584 m = mcache[0]
585 585 fname = None
586 586 if not m.anypats() and len(m.files()) == 1:
587 587 fname = m.files()[0]
588 588 if fname is not None:
589 589 if fname not in c.files():
590 590 return False
591 591 else:
592 592 for f in c.files():
593 593 if m(f):
594 594 break
595 595 else:
596 596 return False
597 597 files = repo.status(c.p1().node(), c.node())[field]
598 598 if fname is not None:
599 599 if fname in files:
600 600 return True
601 601 else:
602 602 for f in files:
603 603 if m(f):
604 604 return True
605 605
606 606 return subset.filter(matches)
607 607
608 608 def _children(repo, narrow, parentset):
609 609 cs = set()
610 610 if not parentset:
611 611 return baseset(cs)
612 612 pr = repo.changelog.parentrevs
613 613 minrev = min(parentset)
614 614 for r in narrow:
615 615 if r <= minrev:
616 616 continue
617 617 for p in pr(r):
618 618 if p in parentset:
619 619 cs.add(r)
620 620 return baseset(cs)
621 621
622 622 def children(repo, subset, x):
623 623 """``children(set)``
624 624 Child changesets of changesets in set.
625 625 """
626 626 s = getset(repo, fullreposet(repo), x)
627 627 cs = _children(repo, subset, s)
628 628 return subset & cs
629 629
630 630 def closed(repo, subset, x):
631 631 """``closed()``
632 632 Changeset is closed.
633 633 """
634 634 # i18n: "closed" is a keyword
635 635 getargs(x, 0, 0, _("closed takes no arguments"))
636 636 return subset.filter(lambda r: repo[r].closesbranch())
637 637
638 638 def contains(repo, subset, x):
639 639 """``contains(pattern)``
640 640 The revision's manifest contains a file matching pattern (but might not
641 641 modify it). See :hg:`help patterns` for information about file patterns.
642 642
643 643 The pattern without explicit kind like ``glob:`` is expected to be
644 644 relative to the current directory and match against a file exactly
645 645 for efficiency.
646 646 """
647 647 # i18n: "contains" is a keyword
648 648 pat = getstring(x, _("contains requires a pattern"))
649 649
650 650 def matches(x):
651 651 if not matchmod.patkind(pat):
652 652 pats = pathutil.canonpath(repo.root, repo.getcwd(), pat)
653 653 if pats in repo[x]:
654 654 return True
655 655 else:
656 656 c = repo[x]
657 657 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
658 658 for f in c.manifest():
659 659 if m(f):
660 660 return True
661 661 return False
662 662
663 663 return subset.filter(matches)
664 664
665 665 def converted(repo, subset, x):
666 666 """``converted([id])``
667 667 Changesets converted from the given identifier in the old repository if
668 668 present, or all converted changesets if no identifier is specified.
669 669 """
670 670
671 671 # There is exactly no chance of resolving the revision, so do a simple
672 672 # string compare and hope for the best
673 673
674 674 rev = None
675 675 # i18n: "converted" is a keyword
676 676 l = getargs(x, 0, 1, _('converted takes one or no arguments'))
677 677 if l:
678 678 # i18n: "converted" is a keyword
679 679 rev = getstring(l[0], _('converted requires a revision'))
680 680
681 681 def _matchvalue(r):
682 682 source = repo[r].extra().get('convert_revision', None)
683 683 return source is not None and (rev is None or source.startswith(rev))
684 684
685 685 return subset.filter(lambda r: _matchvalue(r))
686 686
687 687 def date(repo, subset, x):
688 688 """``date(interval)``
689 689 Changesets within the interval, see :hg:`help dates`.
690 690 """
691 691 # i18n: "date" is a keyword
692 692 ds = getstring(x, _("date requires a string"))
693 693 dm = util.matchdate(ds)
694 694 return subset.filter(lambda x: dm(repo[x].date()[0]))
695 695
696 696 def desc(repo, subset, x):
697 697 """``desc(string)``
698 698 Search commit message for string. The match is case-insensitive.
699 699 """
700 700 # i18n: "desc" is a keyword
701 701 ds = encoding.lower(getstring(x, _("desc requires a string")))
702 702
703 703 def matches(x):
704 704 c = repo[x]
705 705 return ds in encoding.lower(c.description())
706 706
707 707 return subset.filter(matches)
708 708
709 709 def _descendants(repo, subset, x, followfirst=False):
710 710 roots = getset(repo, fullreposet(repo), x)
711 711 if not roots:
712 712 return baseset()
713 713 s = _revdescendants(repo, roots, followfirst)
714 714
715 715 # Both sets need to be ascending in order to lazily return the union
716 716 # in the correct order.
717 717 base = subset & roots
718 718 desc = subset & s
719 719 result = base + desc
720 720 if subset.isascending():
721 721 result.sort()
722 722 elif subset.isdescending():
723 723 result.sort(reverse=True)
724 724 else:
725 725 result = subset & result
726 726 return result
727 727
728 728 def descendants(repo, subset, x):
729 729 """``descendants(set)``
730 730 Changesets which are descendants of changesets in set.
731 731 """
732 732 return _descendants(repo, subset, x)
733 733
734 734 def _firstdescendants(repo, subset, x):
735 735 # ``_firstdescendants(set)``
736 736 # Like ``descendants(set)`` but follows only the first parents.
737 737 return _descendants(repo, subset, x, followfirst=True)
738 738
739 739 def destination(repo, subset, x):
740 740 """``destination([set])``
741 741 Changesets that were created by a graft, transplant or rebase operation,
742 742 with the given revisions specified as the source. Omitting the optional set
743 743 is the same as passing all().
744 744 """
745 745 if x is not None:
746 746 sources = getset(repo, fullreposet(repo), x)
747 747 else:
748 748 sources = fullreposet(repo)
749 749
750 750 dests = set()
751 751
752 752 # subset contains all of the possible destinations that can be returned, so
753 753 # iterate over them and see if their source(s) were provided in the arg set.
754 754 # Even if the immediate src of r is not in the arg set, src's source (or
755 755 # further back) may be. Scanning back further than the immediate src allows
756 756 # transitive transplants and rebases to yield the same results as transitive
757 757 # grafts.
758 758 for r in subset:
759 759 src = _getrevsource(repo, r)
760 760 lineage = None
761 761
762 762 while src is not None:
763 763 if lineage is None:
764 764 lineage = list()
765 765
766 766 lineage.append(r)
767 767
768 768 # The visited lineage is a match if the current source is in the arg
769 769 # set. Since every candidate dest is visited by way of iterating
770 770 # subset, any dests further back in the lineage will be tested by a
771 771 # different iteration over subset. Likewise, if the src was already
772 772 # selected, the current lineage can be selected without going back
773 773 # further.
774 774 if src in sources or src in dests:
775 775 dests.update(lineage)
776 776 break
777 777
778 778 r = src
779 779 src = _getrevsource(repo, r)
780 780
781 781 return subset.filter(dests.__contains__)
782 782
783 783 def divergent(repo, subset, x):
784 784 """``divergent()``
785 785 Final successors of changesets with an alternative set of final successors.
786 786 """
787 787 # i18n: "divergent" is a keyword
788 788 getargs(x, 0, 0, _("divergent takes no arguments"))
789 789 divergent = obsmod.getrevs(repo, 'divergent')
790 790 return subset & divergent
791 791
792 792 def draft(repo, subset, x):
793 793 """``draft()``
794 794 Changeset in draft phase."""
795 795 # i18n: "draft" is a keyword
796 796 getargs(x, 0, 0, _("draft takes no arguments"))
797 797 phase = repo._phasecache.phase
798 798 target = phases.draft
799 799 condition = lambda r: phase(repo, r) == target
800 800 return subset.filter(condition, cache=False)
801 801
802 802 def extinct(repo, subset, x):
803 803 """``extinct()``
804 804 Obsolete changesets with obsolete descendants only.
805 805 """
806 806 # i18n: "extinct" is a keyword
807 807 getargs(x, 0, 0, _("extinct takes no arguments"))
808 808 extincts = obsmod.getrevs(repo, 'extinct')
809 809 return subset & extincts
810 810
811 811 def extra(repo, subset, x):
812 812 """``extra(label, [value])``
813 813 Changesets with the given label in the extra metadata, with the given
814 814 optional value.
815 815
816 816 If `value` starts with `re:`, the remainder of the value is treated as
817 817 a regular expression. To match a value that actually starts with `re:`,
818 818 use the prefix `literal:`.
819 819 """
820 820
821 821 # i18n: "extra" is a keyword
822 822 l = getargs(x, 1, 2, _('extra takes at least 1 and at most 2 arguments'))
823 823 # i18n: "extra" is a keyword
824 824 label = getstring(l[0], _('first argument to extra must be a string'))
825 825 value = None
826 826
827 827 if len(l) > 1:
828 828 # i18n: "extra" is a keyword
829 829 value = getstring(l[1], _('second argument to extra must be a string'))
830 830 kind, value, matcher = _stringmatcher(value)
831 831
832 832 def _matchvalue(r):
833 833 extra = repo[r].extra()
834 834 return label in extra and (value is None or matcher(extra[label]))
835 835
836 836 return subset.filter(lambda r: _matchvalue(r))
837 837
838 838 def filelog(repo, subset, x):
839 839 """``filelog(pattern)``
840 840 Changesets connected to the specified filelog.
841 841
842 842 For performance reasons, visits only revisions mentioned in the file-level
843 843 filelog, rather than filtering through all changesets (much faster, but
844 844 doesn't include deletes or duplicate changes). For a slower, more accurate
845 845 result, use ``file()``.
846 846
847 847 The pattern without explicit kind like ``glob:`` is expected to be
848 848 relative to the current directory and match against a file exactly
849 849 for efficiency.
850 850
851 851 If some linkrev points to revisions filtered by the current repoview, we'll
852 852 work around it to return a non-filtered value.
853 853 """
854 854
855 855 # i18n: "filelog" is a keyword
856 856 pat = getstring(x, _("filelog requires a pattern"))
857 857 s = set()
858 858 cl = repo.changelog
859 859
860 860 if not matchmod.patkind(pat):
861 861 f = pathutil.canonpath(repo.root, repo.getcwd(), pat)
862 862 files = [f]
863 863 else:
864 864 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=repo[None])
865 865 files = (f for f in repo[None] if m(f))
866 866
867 867 for f in files:
868 868 backrevref = {} # final value for: filerev -> changerev
869 869 lowestchild = {} # lowest known filerev child of a filerev
870 870 delayed = [] # filerev with filtered linkrev, for post-processing
871 871 lowesthead = None # cache for manifest content of all head revisions
872 872 fl = repo.file(f)
873 873 for fr in list(fl):
874 874 rev = fl.linkrev(fr)
875 875 if rev not in cl:
876 876 # changerev pointed in linkrev is filtered
877 877 # record it for post processing.
878 878 delayed.append((fr, rev))
879 879 continue
880 880 for p in fl.parentrevs(fr):
881 881 if 0 <= p and p not in lowestchild:
882 882 lowestchild[p] = fr
883 883 backrevref[fr] = rev
884 884 s.add(rev)
885 885
886 886 # Post-processing of all filerevs we skipped because they were
887 887 # filtered. If such filerevs have known and unfiltered children, this
888 888 # means they have an unfiltered appearance out there. We'll use linkrev
889 889 # adjustment to find one of these appearances. The lowest known child
890 890 # will be used as a starting point because it is the best upper-bound we
891 891 # have.
892 892 #
893 893 # This approach will fail when an unfiltered but linkrev-shadowed
894 894 # appearance exists in a head changeset without unfiltered filerev
895 895 # children anywhere.
896 896 while delayed:
897 897 # must be a descending iteration. To slowly fill lowest child
898 898 # information that is of potential use by the next item.
899 899 fr, rev = delayed.pop()
900 900 lkr = rev
901 901
902 902 child = lowestchild.get(fr)
903 903
904 904 if child is None:
905 905 # search for existence of this file revision in a head revision.
906 906 # There are three possibilities:
907 907 # - the revision exists in a head and we can find an
908 908 # introduction from there,
909 909 # - the revision does not exist in a head because it has been
910 910 # changed since its introduction: we would have found a child
911 911 # and be in the other 'else' clause,
912 912 # - all versions of the revision are hidden.
913 913 if lowesthead is None:
914 914 lowesthead = {}
915 915 for h in repo.heads():
916 916 fnode = repo[h].manifest().get(f)
917 917 if fnode is not None:
918 918 lowesthead[fl.rev(fnode)] = h
919 919 headrev = lowesthead.get(fr)
920 920 if headrev is None:
921 921 # content is nowhere unfiltered
922 922 continue
923 923 rev = repo[headrev][f].introrev()
924 924 else:
925 925 # the lowest known child is a good upper bound
926 926 childcrev = backrevref[child]
927 927 # XXX this does not guarantee returning the lowest
928 928 # introduction of this revision, but this gives a
929 929 # result which is a good start and will fit in most
930 930 # cases. We probably need to fix the multiple
931 931 # introductions case properly (report each
932 932 # introduction, even for identical file revisions)
933 933 # once and for all at some point anyway.
934 934 for p in repo[childcrev][f].parents():
935 935 if p.filerev() == fr:
936 936 rev = p.rev()
937 937 break
938 938 if rev == lkr: # no shadowed entry found
939 939 # XXX This should never happen unless some manifest points
940 940 # to biggish file revisions (like a revision that uses a
941 941 # parent that never appears in the manifest ancestors)
942 942 continue
943 943
944 944 # Fill the data for the next iteration.
945 945 for p in fl.parentrevs(fr):
946 946 if 0 <= p and p not in lowestchild:
947 947 lowestchild[p] = fr
948 948 backrevref[fr] = rev
949 949 s.add(rev)
950 950
951 951 return subset & s
952 952
953 953 def first(repo, subset, x):
954 954 """``first(set, [n])``
955 955 An alias for limit().
956 956 """
957 957 return limit(repo, subset, x)
958 958
959 959 def _follow(repo, subset, x, name, followfirst=False):
960 960 l = getargs(x, 0, 1, _("%s takes no arguments or a filename") % name)
961 961 c = repo['.']
962 962 if l:
963 963 x = getstring(l[0], _("%s expected a filename") % name)
964 964 if x in c:
965 965 cx = c[x]
966 966 s = set(ctx.rev() for ctx in cx.ancestors(followfirst=followfirst))
967 967 # include the revision responsible for the most recent version
968 968 s.add(cx.introrev())
969 969 else:
970 970 return baseset()
971 971 else:
972 972 s = _revancestors(repo, baseset([c.rev()]), followfirst)
973 973
974 974 return subset & s
975 975
976 976 def follow(repo, subset, x):
977 977 """``follow([file])``
978 978 An alias for ``::.`` (ancestors of the working directory's first parent).
979 979 If a filename is specified, the history of the given file is followed,
980 980 including copies.
981 981 """
982 982 return _follow(repo, subset, x, 'follow')
983 983
984 984 def _followfirst(repo, subset, x):
985 985 # ``followfirst([file])``
986 986 # Like ``follow([file])`` but follows only the first parent of
987 987 # every revision or file revision.
988 988 return _follow(repo, subset, x, '_followfirst', followfirst=True)
989 989
990 990 def getall(repo, subset, x):
991 991 """``all()``
992 992 All changesets, the same as ``0:tip``.
993 993 """
994 994 # i18n: "all" is a keyword
995 995 getargs(x, 0, 0, _("all takes no arguments"))
996 996 return subset & spanset(repo) # drop "null" if any
997 997
998 998 def grep(repo, subset, x):
999 999 """``grep(regex)``
1000 1000 Like ``keyword(string)`` but accepts a regex. Use ``grep(r'...')``
1001 1001 to ensure special escape characters are handled correctly. Unlike
1002 1002 ``keyword(string)``, the match is case-sensitive.
1003 1003 """
1004 1004 try:
1005 1005 # i18n: "grep" is a keyword
1006 1006 gr = re.compile(getstring(x, _("grep requires a string")))
1007 1007 except re.error, e:
1008 1008 raise error.ParseError(_('invalid match pattern: %s') % e)
1009 1009
1010 1010 def matches(x):
1011 1011 c = repo[x]
1012 1012 for e in c.files() + [c.user(), c.description()]:
1013 1013 if gr.search(e):
1014 1014 return True
1015 1015 return False
1016 1016
1017 1017 return subset.filter(matches)
1018 1018
1019 1019 def _matchfiles(repo, subset, x):
1020 1020 # _matchfiles takes a revset list of prefixed arguments:
1021 1021 #
1022 1022 # [p:foo, i:bar, x:baz]
1023 1023 #
1024 1024 # builds a match object from them and filters subset. Allowed
1025 1025 # prefixes are 'p:' for regular patterns, 'i:' for include
1026 1026 # patterns and 'x:' for exclude patterns. Use 'r:' prefix to pass
1027 1027 # a revision identifier, or the empty string to reference the
1028 1028 # working directory, from which the match object is
1029 1029 # initialized. Use 'd:' to set the default matching mode, default
1030 1030 # to 'glob'. At most one 'r:' and 'd:' argument can be passed.
1031 1031
1032 1032 # i18n: "_matchfiles" is a keyword
1033 1033 l = getargs(x, 1, -1, _("_matchfiles requires at least one argument"))
1034 1034 pats, inc, exc = [], [], []
1035 1035 rev, default = None, None
1036 1036 for arg in l:
1037 1037 # i18n: "_matchfiles" is a keyword
1038 1038 s = getstring(arg, _("_matchfiles requires string arguments"))
1039 1039 prefix, value = s[:2], s[2:]
1040 1040 if prefix == 'p:':
1041 1041 pats.append(value)
1042 1042 elif prefix == 'i:':
1043 1043 inc.append(value)
1044 1044 elif prefix == 'x:':
1045 1045 exc.append(value)
1046 1046 elif prefix == 'r:':
1047 1047 if rev is not None:
1048 1048 # i18n: "_matchfiles" is a keyword
1049 1049 raise error.ParseError(_('_matchfiles expected at most one '
1050 1050 'revision'))
1051 1051 if value != '': # empty means working directory; leave rev as None
1052 1052 rev = value
1053 1053 elif prefix == 'd:':
1054 1054 if default is not None:
1055 1055 # i18n: "_matchfiles" is a keyword
1056 1056 raise error.ParseError(_('_matchfiles expected at most one '
1057 1057 'default mode'))
1058 1058 default = value
1059 1059 else:
1060 1060 # i18n: "_matchfiles" is a keyword
1061 1061 raise error.ParseError(_('invalid _matchfiles prefix: %s') % prefix)
1062 1062 if not default:
1063 1063 default = 'glob'
1064 1064
1065 1065 m = matchmod.match(repo.root, repo.getcwd(), pats, include=inc,
1066 1066 exclude=exc, ctx=repo[rev], default=default)
1067 1067
1068 1068 def matches(x):
1069 1069 for f in repo[x].files():
1070 1070 if m(f):
1071 1071 return True
1072 1072 return False
1073 1073
1074 1074 return subset.filter(matches)
1075 1075
1076 1076 def hasfile(repo, subset, x):
1077 1077 """``file(pattern)``
1078 1078 Changesets affecting files matched by pattern.
1079 1079
1080 1080 For a faster but less accurate result, consider using ``filelog()``
1081 1081 instead.
1082 1082
1083 1083 This predicate uses ``glob:`` as the default kind of pattern.
1084 1084 """
1085 1085 # i18n: "file" is a keyword
1086 1086 pat = getstring(x, _("file requires a pattern"))
1087 1087 return _matchfiles(repo, subset, ('string', 'p:' + pat))
1088 1088
1089 1089 def head(repo, subset, x):
1090 1090 """``head()``
1091 1091 Changeset is a named branch head.
1092 1092 """
1093 1093 # i18n: "head" is a keyword
1094 1094 getargs(x, 0, 0, _("head takes no arguments"))
1095 1095 hs = set()
1096 1096 for b, ls in repo.branchmap().iteritems():
1097 1097 hs.update(repo[h].rev() for h in ls)
1098 1098 return baseset(hs).filter(subset.__contains__)
1099 1099
1100 1100 def heads(repo, subset, x):
1101 1101 """``heads(set)``
1102 1102 Members of set with no children in set.
1103 1103 """
1104 1104 s = getset(repo, subset, x)
1105 1105 ps = parents(repo, subset, x)
1106 1106 return s - ps
1107 1107
1108 1108 def hidden(repo, subset, x):
1109 1109 """``hidden()``
1110 1110 Hidden changesets.
1111 1111 """
1112 1112 # i18n: "hidden" is a keyword
1113 1113 getargs(x, 0, 0, _("hidden takes no arguments"))
1114 1114 hiddenrevs = repoview.filterrevs(repo, 'visible')
1115 1115 return subset & hiddenrevs
1116 1116
1117 1117 def keyword(repo, subset, x):
1118 1118 """``keyword(string)``
1119 1119 Search commit message, user name, and names of changed files for
1120 1120 string. The match is case-insensitive.
1121 1121 """
1122 1122 # i18n: "keyword" is a keyword
1123 1123 kw = encoding.lower(getstring(x, _("keyword requires a string")))
1124 1124
1125 1125 def matches(r):
1126 1126 c = repo[r]
1127 1127 return util.any(kw in encoding.lower(t) for t in c.files() + [c.user(),
1128 1128 c.description()])
1129 1129
1130 1130 return subset.filter(matches)
1131 1131
1132 1132 def limit(repo, subset, x):
1133 1133 """``limit(set, [n])``
1134 1134 First n members of set, defaulting to 1.
1135 1135 """
1136 1136 # i18n: "limit" is a keyword
1137 1137 l = getargs(x, 1, 2, _("limit requires one or two arguments"))
1138 1138 try:
1139 1139 lim = 1
1140 1140 if len(l) == 2:
1141 1141 # i18n: "limit" is a keyword
1142 1142 lim = int(getstring(l[1], _("limit requires a number")))
1143 1143 except (TypeError, ValueError):
1144 1144 # i18n: "limit" is a keyword
1145 1145 raise error.ParseError(_("limit expects a number"))
1146 1146 ss = subset
1147 1147 os = getset(repo, fullreposet(repo), l[0])
1148 1148 result = []
1149 1149 it = iter(os)
1150 1150 for x in xrange(lim):
1151 1151 try:
1152 1152 y = it.next()
1153 1153 if y in ss:
1154 1154 result.append(y)
1155 1155 except (StopIteration):
1156 1156 break
1157 1157 return baseset(result)
1158 1158
1159 1159 def last(repo, subset, x):
1160 1160 """``last(set, [n])``
1161 1161 Last n members of set, defaulting to 1.
1162 1162 """
1163 1163 # i18n: "last" is a keyword
1164 1164 l = getargs(x, 1, 2, _("last requires one or two arguments"))
1165 1165 try:
1166 1166 lim = 1
1167 1167 if len(l) == 2:
1168 1168 # i18n: "last" is a keyword
1169 1169 lim = int(getstring(l[1], _("last requires a number")))
1170 1170 except (TypeError, ValueError):
1171 1171 # i18n: "last" is a keyword
1172 1172 raise error.ParseError(_("last expects a number"))
1173 1173 ss = subset
1174 1174 os = getset(repo, fullreposet(repo), l[0])
1175 1175 os.reverse()
1176 1176 result = []
1177 1177 it = iter(os)
1178 1178 for x in xrange(lim):
1179 1179 try:
1180 1180 y = it.next()
1181 1181 if y in ss:
1182 1182 result.append(y)
1183 1183 except (StopIteration):
1184 1184 break
1185 1185 return baseset(result)
1186 1186
1187 1187 def maxrev(repo, subset, x):
1188 1188 """``max(set)``
1189 1189 Changeset with highest revision number in set.
1190 1190 """
1191 1191 os = getset(repo, fullreposet(repo), x)
1192 1192 if os:
1193 1193 m = os.max()
1194 1194 if m in subset:
1195 1195 return baseset([m])
1196 1196 return baseset()
1197 1197
1198 1198 def merge(repo, subset, x):
1199 1199 """``merge()``
1200 1200 Changeset is a merge changeset.
1201 1201 """
1202 1202 # i18n: "merge" is a keyword
1203 1203 getargs(x, 0, 0, _("merge takes no arguments"))
1204 1204 cl = repo.changelog
1205 1205 return subset.filter(lambda r: cl.parentrevs(r)[1] != -1)
1206 1206
1207 1207 def branchpoint(repo, subset, x):
1208 1208 """``branchpoint()``
1209 1209 Changesets with more than one child.
1210 1210 """
1211 1211 # i18n: "branchpoint" is a keyword
1212 1212 getargs(x, 0, 0, _("branchpoint takes no arguments"))
1213 1213 cl = repo.changelog
1214 1214 if not subset:
1215 1215 return baseset()
1216 1216 baserev = min(subset)
1217 1217 parentscount = [0]*(len(repo) - baserev)
1218 1218 for r in cl.revs(start=baserev + 1):
1219 1219 for p in cl.parentrevs(r):
1220 1220 if p >= baserev:
1221 1221 parentscount[p - baserev] += 1
1222 1222 return subset.filter(lambda r: parentscount[r - baserev] > 1)
1223 1223
1224 1224 def minrev(repo, subset, x):
1225 1225 """``min(set)``
1226 1226 Changeset with lowest revision number in set.
1227 1227 """
1228 1228 os = getset(repo, fullreposet(repo), x)
1229 1229 if os:
1230 1230 m = os.min()
1231 1231 if m in subset:
1232 1232 return baseset([m])
1233 1233 return baseset()
1234 1234
1235 1235 def modifies(repo, subset, x):
1236 1236 """``modifies(pattern)``
1237 1237 Changesets modifying files matched by pattern.
1238 1238
1239 1239 The pattern without explicit kind like ``glob:`` is expected to be
1240 1240 relative to the current directory and match against a file or a
1241 1241 directory.
1242 1242 """
1243 1243 # i18n: "modifies" is a keyword
1244 1244 pat = getstring(x, _("modifies requires a pattern"))
1245 1245 return checkstatus(repo, subset, pat, 0)
1246 1246
1247 1247 def named(repo, subset, x):
1248 1248 """``named(namespace)``
1249 1249 The changesets in a given namespace.
1250 1250
1251 1251 If `namespace` starts with `re:`, the remainder of the string is treated as
1252 1252 a regular expression. To match a namespace that actually starts with `re:`,
1253 1253 use the prefix `literal:`.
1254 1254 """
1255 1255 # i18n: "named" is a keyword
1256 1256 args = getargs(x, 1, 1, _('named requires a namespace argument'))
1257 1257
1258 1258 ns = getstring(args[0],
1259 1259 # i18n: "named" is a keyword
1260 1260 _('the argument to named must be a string'))
1261 1261 kind, pattern, matcher = _stringmatcher(ns)
1262 1262 namespaces = set()
1263 1263 if kind == 'literal':
1264 1264 if pattern not in repo.names:
1265 1265 raise error.RepoLookupError(_("namespace '%s' does not exist")
1266 1266 % ns)
1267 1267 namespaces.add(repo.names[pattern])
1268 1268 else:
1269 1269 for name, ns in repo.names.iteritems():
1270 1270 if matcher(name):
1271 1271 namespaces.add(ns)
1272 1272 if not namespaces:
1273 1273 raise error.RepoLookupError(_("no namespace exists"
1274 1274 " that match '%s'") % pattern)
1275 1275
1276 1276 names = set()
1277 1277 for ns in namespaces:
1278 1278 for name in ns.listnames(repo):
1279 1279 if name not in ns.deprecated:
1280 1280 names.update(repo[n].rev() for n in ns.nodes(repo, name))
1281 1281
1282 1282 names -= set([node.nullrev])
1283 1283 return subset & names
1284 1284
1285 1285 def node_(repo, subset, x):
1286 1286 """``id(string)``
1287 1287 Revision non-ambiguously specified by the given hex string prefix.
1288 1288 """
1289 1289 # i18n: "id" is a keyword
1290 1290 l = getargs(x, 1, 1, _("id requires one argument"))
1291 1291 # i18n: "id" is a keyword
1292 1292 n = getstring(l[0], _("id requires a string"))
1293 1293 if len(n) == 40:
1294 1294 try:
1295 1295 rn = repo.changelog.rev(node.bin(n))
1296 1296 except (LookupError, TypeError):
1297 1297 rn = None
1298 1298 else:
1299 1299 rn = None
1300 1300 pm = repo.changelog._partialmatch(n)
1301 1301 if pm is not None:
1302 1302 rn = repo.changelog.rev(pm)
1303 1303
1304 1304 if rn is None:
1305 1305 return baseset()
1306 1306 result = baseset([rn])
1307 1307 return result & subset
1308 1308
1309 1309 def obsolete(repo, subset, x):
1310 1310 """``obsolete()``
1311 1311 Mutable changeset with a newer version."""
1312 1312 # i18n: "obsolete" is a keyword
1313 1313 getargs(x, 0, 0, _("obsolete takes no arguments"))
1314 1314 obsoletes = obsmod.getrevs(repo, 'obsolete')
1315 1315 return subset & obsoletes
1316 1316
1317 1317 def only(repo, subset, x):
1318 1318 """``only(set, [set])``
1319 1319 Changesets that are ancestors of the first set that are not ancestors
1320 1320 of any other head in the repo. If a second set is specified, the result
1321 1321 is ancestors of the first set that are not ancestors of the second set
1322 1322 (i.e. ::<set1> - ::<set2>).
1323 1323 """
1324 1324 cl = repo.changelog
1325 1325 # i18n: "only" is a keyword
1326 1326 args = getargs(x, 1, 2, _('only takes one or two arguments'))
1327 1327 include = getset(repo, fullreposet(repo), args[0])
1328 1328 if len(args) == 1:
1329 1329 if not include:
1330 1330 return baseset()
1331 1331
1332 1332 descendants = set(_revdescendants(repo, include, False))
1333 1333 exclude = [rev for rev in cl.headrevs()
1334 1334 if not rev in descendants and not rev in include]
1335 1335 else:
1336 1336 exclude = getset(repo, fullreposet(repo), args[1])
1337 1337
1338 1338 results = set(cl.findmissingrevs(common=exclude, heads=include))
1339 1339 return subset & results
1340 1340
1341 1341 def origin(repo, subset, x):
1342 1342 """``origin([set])``
1343 1343 Changesets that were specified as a source for the grafts, transplants or
1344 1344 rebases that created the given revisions. Omitting the optional set is the
1345 1345 same as passing all(). If a changeset created by these operations is itself
1346 1346 specified as a source for one of these operations, only the source changeset
1347 1347 for the first operation is selected.
1348 1348 """
1349 1349 if x is not None:
1350 1350 dests = getset(repo, fullreposet(repo), x)
1351 1351 else:
1352 1352 dests = fullreposet(repo)
1353 1353
1354 1354 def _firstsrc(rev):
1355 1355 src = _getrevsource(repo, rev)
1356 1356 if src is None:
1357 1357 return None
1358 1358
1359 1359 while True:
1360 1360 prev = _getrevsource(repo, src)
1361 1361
1362 1362 if prev is None:
1363 1363 return src
1364 1364 src = prev
1365 1365
1366 1366 o = set([_firstsrc(r) for r in dests])
1367 1367 o -= set([None])
1368 1368 return subset & o
1369 1369
1370 1370 def outgoing(repo, subset, x):
1371 1371 """``outgoing([path])``
1372 1372 Changesets not found in the specified destination repository, or the
1373 1373 default push location.
1374 1374 """
1375 1375 # Avoid cycles.
1376 1376 import discovery
1377 1377 import hg
1378 1378 # i18n: "outgoing" is a keyword
1379 1379 l = getargs(x, 0, 1, _("outgoing takes one or no arguments"))
1380 1380 # i18n: "outgoing" is a keyword
1381 1381 dest = l and getstring(l[0], _("outgoing requires a repository path")) or ''
1382 1382 dest = repo.ui.expandpath(dest or 'default-push', dest or 'default')
1383 1383 dest, branches = hg.parseurl(dest)
1384 1384 revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
1385 1385 if revs:
1386 1386 revs = [repo.lookup(rev) for rev in revs]
1387 1387 other = hg.peer(repo, {}, dest)
1388 1388 repo.ui.pushbuffer()
1389 1389 outgoing = discovery.findcommonoutgoing(repo, other, onlyheads=revs)
1390 1390 repo.ui.popbuffer()
1391 1391 cl = repo.changelog
1392 1392 o = set([cl.rev(r) for r in outgoing.missing])
1393 1393 return subset & o
1394 1394
1395 1395 def p1(repo, subset, x):
1396 1396 """``p1([set])``
1397 1397 First parent of changesets in set, or the working directory.
1398 1398 """
1399 1399 if x is None:
1400 1400 p = repo[x].p1().rev()
1401 1401 if p >= 0:
1402 1402 return subset & baseset([p])
1403 1403 return baseset()
1404 1404
1405 1405 ps = set()
1406 1406 cl = repo.changelog
1407 1407 for r in getset(repo, fullreposet(repo), x):
1408 1408 ps.add(cl.parentrevs(r)[0])
1409 1409 ps -= set([node.nullrev])
1410 1410 return subset & ps
1411 1411
1412 1412 def p2(repo, subset, x):
1413 1413 """``p2([set])``
1414 1414 Second parent of changesets in set, or the working directory.
1415 1415 """
1416 1416 if x is None:
1417 1417 ps = repo[x].parents()
1418 1418 try:
1419 1419 p = ps[1].rev()
1420 1420 if p >= 0:
1421 1421 return subset & baseset([p])
1422 1422 return baseset()
1423 1423 except IndexError:
1424 1424 return baseset()
1425 1425
1426 1426 ps = set()
1427 1427 cl = repo.changelog
1428 1428 for r in getset(repo, fullreposet(repo), x):
1429 1429 ps.add(cl.parentrevs(r)[1])
1430 1430 ps -= set([node.nullrev])
1431 1431 return subset & ps
1432 1432
1433 1433 def parents(repo, subset, x):
1434 1434 """``parents([set])``
1435 1435 The set of all parents for all changesets in set, or the working directory.
1436 1436 """
1437 1437 if x is None:
1438 1438 ps = set(p.rev() for p in repo[x].parents())
1439 1439 else:
1440 1440 ps = set()
1441 1441 cl = repo.changelog
1442 1442 for r in getset(repo, fullreposet(repo), x):
1443 1443 ps.update(cl.parentrevs(r))
1444 1444 ps -= set([node.nullrev])
1445 1445 return subset & ps
1446 1446
1447 1447 def parentspec(repo, subset, x, n):
1448 1448 """``set^0``
1449 1449 The set.
1450 1450 ``set^1`` (or ``set^``), ``set^2``
1451 1451 First or second parent, respectively, of all changesets in set.
1452 1452 """
1453 1453 try:
1454 1454 n = int(n[1])
1455 1455 if n not in (0, 1, 2):
1456 1456 raise ValueError
1457 1457 except (TypeError, ValueError):
1458 1458 raise error.ParseError(_("^ expects a number 0, 1, or 2"))
1459 1459 ps = set()
1460 1460 cl = repo.changelog
1461 1461 for r in getset(repo, fullreposet(repo), x):
1462 1462 if n == 0:
1463 1463 ps.add(r)
1464 1464 elif n == 1:
1465 1465 ps.add(cl.parentrevs(r)[0])
1466 1466 elif n == 2:
1467 1467 parents = cl.parentrevs(r)
1468 1468 if len(parents) > 1:
1469 1469 ps.add(parents[1])
1470 1470 return subset & ps
1471 1471
1472 1472 def present(repo, subset, x):
1473 1473 """``present(set)``
1474 1474 An empty set, if any revision in set isn't found; otherwise,
1475 1475 all revisions in set.
1476 1476
1477 1477 If any of specified revisions is not present in the local repository,
1478 1478 the query is normally aborted. But this predicate allows the query
1479 1479 to continue even in such cases.
1480 1480 """
1481 1481 try:
1482 1482 return getset(repo, subset, x)
1483 1483 except error.RepoLookupError:
1484 1484 return baseset()
1485 1485
1486 1486 def public(repo, subset, x):
1487 1487 """``public()``
1488 1488 Changeset in public phase."""
1489 1489 # i18n: "public" is a keyword
1490 1490 getargs(x, 0, 0, _("public takes no arguments"))
1491 1491 phase = repo._phasecache.phase
1492 1492 target = phases.public
1493 1493 condition = lambda r: phase(repo, r) == target
1494 1494 return subset.filter(condition, cache=False)
1495 1495
1496 1496 def remote(repo, subset, x):
1497 1497 """``remote([id [,path]])``
1498 1498 Local revision that corresponds to the given identifier in a
1499 1499 remote repository, if present. Here, the '.' identifier is a
1500 1500 synonym for the current local branch.
1501 1501 """
1502 1502
1503 1503 import hg # avoid start-up nasties
1504 1504 # i18n: "remote" is a keyword
1505 1505 l = getargs(x, 0, 2, _("remote takes one, two or no arguments"))
1506 1506
1507 1507 q = '.'
1508 1508 if len(l) > 0:
1509 1509 # i18n: "remote" is a keyword
1510 1510 q = getstring(l[0], _("remote requires a string id"))
1511 1511 if q == '.':
1512 1512 q = repo['.'].branch()
1513 1513
1514 1514 dest = ''
1515 1515 if len(l) > 1:
1516 1516 # i18n: "remote" is a keyword
1517 1517 dest = getstring(l[1], _("remote requires a repository path"))
1518 1518 dest = repo.ui.expandpath(dest or 'default')
1519 1519 dest, branches = hg.parseurl(dest)
1520 1520 revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
1521 1521 if revs:
1522 1522 revs = [repo.lookup(rev) for rev in revs]
1523 1523 other = hg.peer(repo, {}, dest)
1524 1524 n = other.lookup(q)
1525 1525 if n in repo:
1526 1526 r = repo[n].rev()
1527 1527 if r in subset:
1528 1528 return baseset([r])
1529 1529 return baseset()
1530 1530
1531 1531 def removes(repo, subset, x):
1532 1532 """``removes(pattern)``
1533 1533 Changesets which remove files matching pattern.
1534 1534
1535 1535 The pattern without explicit kind like ``glob:`` is expected to be
1536 1536 relative to the current directory and match against a file or a
1537 1537 directory.
1538 1538 """
1539 1539 # i18n: "removes" is a keyword
1540 1540 pat = getstring(x, _("removes requires a pattern"))
1541 1541 return checkstatus(repo, subset, pat, 2)
1542 1542
1543 1543 def rev(repo, subset, x):
1544 1544 """``rev(number)``
1545 1545 Revision with the given numeric identifier.
1546 1546 """
1547 1547 # i18n: "rev" is a keyword
1548 1548 l = getargs(x, 1, 1, _("rev requires one argument"))
1549 1549 try:
1550 1550 # i18n: "rev" is a keyword
1551 1551 l = int(getstring(l[0], _("rev requires a number")))
1552 1552 except (TypeError, ValueError):
1553 1553 # i18n: "rev" is a keyword
1554 1554 raise error.ParseError(_("rev expects a number"))
1555 1555 if l not in repo.changelog and l != node.nullrev:
1556 1556 return baseset()
1557 1557 return subset & baseset([l])
1558 1558
1559 1559 def matching(repo, subset, x):
1560 1560 """``matching(revision [, field])``
1561 1561 Changesets in which a given set of fields match the set of fields in the
1562 1562 selected revision or set.
1563 1563
1564 1564 To match more than one field pass the list of fields to match separated
1565 1565 by spaces (e.g. ``author description``).
1566 1566
1567 1567 Valid fields are most regular revision fields and some special fields.
1568 1568
1569 1569 Regular revision fields are ``description``, ``author``, ``branch``,
1570 1570 ``date``, ``files``, ``phase``, ``parents``, ``substate``, ``user``
1571 1571 and ``diff``.
1572 1572 Note that ``author`` and ``user`` are synonyms. ``diff`` refers to the
1573 1573 contents of the revision. Two revisions matching their ``diff`` will
1574 1574 also match their ``files``.
1575 1575
1576 1576 Special fields are ``summary`` and ``metadata``:
1577 1577 ``summary`` matches the first line of the description.
1578 1578 ``metadata`` is equivalent to matching ``description user date``
1579 1579 (i.e. it matches the main metadata fields).
1580 1580
1581 1581 ``metadata`` is the default field which is used when no fields are
1582 1582 specified. You can match more than one field at a time.
1583 1583 """
1584 1584 # i18n: "matching" is a keyword
1585 1585 l = getargs(x, 1, 2, _("matching takes 1 or 2 arguments"))
1586 1586
1587 1587 revs = getset(repo, fullreposet(repo), l[0])
1588 1588
1589 1589 fieldlist = ['metadata']
1590 1590 if len(l) > 1:
1591 1591 fieldlist = getstring(l[1],
1592 1592 # i18n: "matching" is a keyword
1593 1593 _("matching requires a string "
1594 1594 "as its second argument")).split()
1595 1595
1596 1596 # Make sure that there are no repeated fields,
1597 1597 # expand the 'special' 'metadata' field type
1598 1598 # and check the 'files' whenever we check the 'diff'
1599 1599 fields = []
1600 1600 for field in fieldlist:
1601 1601 if field == 'metadata':
1602 1602 fields += ['user', 'description', 'date']
1603 1603 elif field == 'diff':
1604 1604 # a revision matching the diff must also match the files
1605 1605 # since matching the diff is very costly, make sure to
1606 1606 # also match the files first
1607 1607 fields += ['files', 'diff']
1608 1608 else:
1609 1609 if field == 'author':
1610 1610 field = 'user'
1611 1611 fields.append(field)
1612 1612 fields = set(fields)
1613 1613 if 'summary' in fields and 'description' in fields:
1614 1614 # If a revision matches its description it also matches its summary
1615 1615 fields.discard('summary')
1616 1616
1617 1617 # We may want to match more than one field
1618 1618 # Not all fields take the same amount of time to be matched
1619 1619 # Sort the selected fields in order of increasing matching cost
1620 1620 fieldorder = ['phase', 'parents', 'user', 'date', 'branch', 'summary',
1621 1621 'files', 'description', 'substate', 'diff']
1622 1622 def fieldkeyfunc(f):
1623 1623 try:
1624 1624 return fieldorder.index(f)
1625 1625 except ValueError:
1626 1626 # assume an unknown field is very costly
1627 1627 return len(fieldorder)
1628 1628 fields = list(fields)
1629 1629 fields.sort(key=fieldkeyfunc)
1630 1630
1631 1631 # Each field will be matched with its own "getfield" function
1632 1632 # which will be added to the getfieldfuncs array of functions
1633 1633 getfieldfuncs = []
1634 1634 _funcs = {
1635 1635 'user': lambda r: repo[r].user(),
1636 1636 'branch': lambda r: repo[r].branch(),
1637 1637 'date': lambda r: repo[r].date(),
1638 1638 'description': lambda r: repo[r].description(),
1639 1639 'files': lambda r: repo[r].files(),
1640 1640 'parents': lambda r: repo[r].parents(),
1641 1641 'phase': lambda r: repo[r].phase(),
1642 1642 'substate': lambda r: repo[r].substate,
1643 1643 'summary': lambda r: repo[r].description().splitlines()[0],
1644 1644 'diff': lambda r: list(repo[r].diff(git=True),)
1645 1645 }
1646 1646 for info in fields:
1647 1647 getfield = _funcs.get(info, None)
1648 1648 if getfield is None:
1649 1649 raise error.ParseError(
1650 1650 # i18n: "matching" is a keyword
1651 1651 _("unexpected field name passed to matching: %s") % info)
1652 1652 getfieldfuncs.append(getfield)
1653 1653 # convert the getfield array of functions into a "getinfo" function
1654 1654 # which returns an array of field values (or a single value if there
1655 1655 # is only one field to match)
1656 1656 getinfo = lambda r: [f(r) for f in getfieldfuncs]
1657 1657
1658 1658 def matches(x):
1659 1659 for rev in revs:
1660 1660 target = getinfo(rev)
1661 1661 match = True
1662 1662 for n, f in enumerate(getfieldfuncs):
1663 1663 if target[n] != f(x):
1664 1664 match = False
1665 1665 if match:
1666 1666 return True
1667 1667 return False
1668 1668
1669 1669 return subset.filter(matches)
1670 1670
1671 1671 def reverse(repo, subset, x):
1672 1672 """``reverse(set)``
1673 1673 Reverse order of set.
1674 1674 """
1675 1675 l = getset(repo, subset, x)
1676 1676 l.reverse()
1677 1677 return l
1678 1678
1679 1679 def roots(repo, subset, x):
1680 1680 """``roots(set)``
1681 1681 Changesets in set with no parent changeset in set.
1682 1682 """
1683 1683 s = getset(repo, fullreposet(repo), x)
1684 1684 subset = subset & s# baseset([r for r in s if r in subset])
1685 1685 cs = _children(repo, subset, s)
1686 1686 return subset - cs
1687 1687
1688 1688 def secret(repo, subset, x):
1689 1689 """``secret()``
1690 1690 Changeset in secret phase."""
1691 1691 # i18n: "secret" is a keyword
1692 1692 getargs(x, 0, 0, _("secret takes no arguments"))
1693 1693 phase = repo._phasecache.phase
1694 1694 target = phases.secret
1695 1695 condition = lambda r: phase(repo, r) == target
1696 1696 return subset.filter(condition, cache=False)
1697 1697
1698 1698 def sort(repo, subset, x):
1699 1699 """``sort(set[, [-]key...])``
1700 1700 Sort set by keys. The default sort order is ascending, specify a key
1701 1701 as ``-key`` to sort in descending order.
1702 1702
1703 1703 The keys can be:
1704 1704
1705 1705 - ``rev`` for the revision number,
1706 1706 - ``branch`` for the branch name,
1707 1707 - ``desc`` for the commit message (description),
1708 1708 - ``user`` for user name (``author`` can be used as an alias),
1709 1709 - ``date`` for the commit date
1710 1710 """
1711 1711 # i18n: "sort" is a keyword
1712 1712 l = getargs(x, 1, 2, _("sort requires one or two arguments"))
1713 1713 keys = "rev"
1714 1714 if len(l) == 2:
1715 1715 # i18n: "sort" is a keyword
1716 1716 keys = getstring(l[1], _("sort spec must be a string"))
1717 1717
1718 1718 s = l[0]
1719 1719 keys = keys.split()
1720 1720 l = []
1721 1721 def invert(s):
1722 1722 return "".join(chr(255 - ord(c)) for c in s)
1723 1723 revs = getset(repo, subset, s)
1724 1724 if keys == ["rev"]:
1725 1725 revs.sort()
1726 1726 return revs
1727 1727 elif keys == ["-rev"]:
1728 1728 revs.sort(reverse=True)
1729 1729 return revs
1730 1730 for r in revs:
1731 1731 c = repo[r]
1732 1732 e = []
1733 1733 for k in keys:
1734 1734 if k == 'rev':
1735 1735 e.append(r)
1736 1736 elif k == '-rev':
1737 1737 e.append(-r)
1738 1738 elif k == 'branch':
1739 1739 e.append(c.branch())
1740 1740 elif k == '-branch':
1741 1741 e.append(invert(c.branch()))
1742 1742 elif k == 'desc':
1743 1743 e.append(c.description())
1744 1744 elif k == '-desc':
1745 1745 e.append(invert(c.description()))
1746 1746 elif k in 'user author':
1747 1747 e.append(c.user())
1748 1748 elif k in '-user -author':
1749 1749 e.append(invert(c.user()))
1750 1750 elif k == 'date':
1751 1751 e.append(c.date()[0])
1752 1752 elif k == '-date':
1753 1753 e.append(-c.date()[0])
1754 1754 else:
1755 1755 raise error.ParseError(_("unknown sort key %r") % k)
1756 1756 e.append(r)
1757 1757 l.append(e)
1758 1758 l.sort()
1759 1759 return baseset([e[-1] for e in l])
1760 1760
1761 1761 def subrepo(repo, subset, x):
1762 1762 """``subrepo([pattern])``
1763 1763 Changesets that add, modify or remove the given subrepo. If no subrepo
1764 1764 pattern is named, any subrepo changes are returned.
1765 1765 """
1766 1766 # i18n: "subrepo" is a keyword
1767 1767 args = getargs(x, 0, 1, _('subrepo takes at most one argument'))
1768 1768 if len(args) != 0:
1769 1769 pat = getstring(args[0], _("subrepo requires a pattern"))
1770 1770
1771 1771 m = matchmod.exact(repo.root, repo.root, ['.hgsubstate'])
1772 1772
1773 1773 def submatches(names):
1774 1774 k, p, m = _stringmatcher(pat)
1775 1775 for name in names:
1776 1776 if m(name):
1777 1777 yield name
1778 1778
1779 1779 def matches(x):
1780 1780 c = repo[x]
1781 1781 s = repo.status(c.p1().node(), c.node(), match=m)
1782 1782
1783 1783 if len(args) == 0:
1784 1784 return s.added or s.modified or s.removed
1785 1785
1786 1786 if s.added:
1787 1787 return util.any(submatches(c.substate.keys()))
1788 1788
1789 1789 if s.modified:
1790 1790 subs = set(c.p1().substate.keys())
1791 1791 subs.update(c.substate.keys())
1792 1792
1793 1793 for path in submatches(subs):
1794 1794 if c.p1().substate.get(path) != c.substate.get(path):
1795 1795 return True
1796 1796
1797 1797 if s.removed:
1798 1798 return util.any(submatches(c.p1().substate.keys()))
1799 1799
1800 1800 return False
1801 1801
1802 1802 return subset.filter(matches)
1803 1803
1804 1804 def _stringmatcher(pattern):
1805 1805 """
1806 1806 accepts a string, possibly starting with 're:' or 'literal:' prefix.
1807 1807 returns the matcher name, pattern, and matcher function.
1808 1808 missing or unknown prefixes are treated as literal matches.
1809 1809
1810 1810 helper for tests:
1811 1811 >>> def test(pattern, *tests):
1812 1812 ... kind, pattern, matcher = _stringmatcher(pattern)
1813 1813 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
1814 1814
1815 1815 exact matching (no prefix):
1816 1816 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
1817 1817 ('literal', 'abcdefg', [False, False, True])
1818 1818
1819 1819 regex matching ('re:' prefix)
1820 1820 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
1821 1821 ('re', 'a.+b', [False, False, True])
1822 1822
1823 1823 force exact matches ('literal:' prefix)
1824 1824 >>> test('literal:re:foobar', 'foobar', 're:foobar')
1825 1825 ('literal', 're:foobar', [False, True])
1826 1826
1827 1827 unknown prefixes are ignored and treated as literals
1828 1828 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
1829 1829 ('literal', 'foo:bar', [False, False, True])
1830 1830 """
1831 1831 if pattern.startswith('re:'):
1832 1832 pattern = pattern[3:]
1833 1833 try:
1834 1834 regex = re.compile(pattern)
1835 1835 except re.error, e:
1836 1836 raise error.ParseError(_('invalid regular expression: %s')
1837 1837 % e)
1838 1838 return 're', pattern, regex.search
1839 1839 elif pattern.startswith('literal:'):
1840 1840 pattern = pattern[8:]
1841 1841 return 'literal', pattern, pattern.__eq__
1842 1842
1843 1843 def _substringmatcher(pattern):
1844 1844 kind, pattern, matcher = _stringmatcher(pattern)
1845 1845 if kind == 'literal':
1846 1846 matcher = lambda s: pattern in s
1847 1847 return kind, pattern, matcher
1848 1848
1849 1849 def tag(repo, subset, x):
1850 1850 """``tag([name])``
1851 1851 The specified tag by name, or all tagged revisions if no name is given.
1852 1852
1853 1853 If `name` starts with `re:`, the remainder of the name is treated as
1854 1854 a regular expression. To match a tag that actually starts with `re:`,
1855 1855 use the prefix `literal:`.
1856 1856 """
1857 1857 # i18n: "tag" is a keyword
1858 1858 args = getargs(x, 0, 1, _("tag takes one or no arguments"))
1859 1859 cl = repo.changelog
1860 1860 if args:
1861 1861 pattern = getstring(args[0],
1862 1862 # i18n: "tag" is a keyword
1863 1863 _('the argument to tag must be a string'))
1864 1864 kind, pattern, matcher = _stringmatcher(pattern)
1865 1865 if kind == 'literal':
1866 1866 # avoid resolving all tags
1867 1867 tn = repo._tagscache.tags.get(pattern, None)
1868 1868 if tn is None:
1869 1869 raise error.RepoLookupError(_("tag '%s' does not exist")
1870 1870 % pattern)
1871 1871 s = set([repo[tn].rev()])
1872 1872 else:
1873 1873 s = set([cl.rev(n) for t, n in repo.tagslist() if matcher(t)])
1874 1874 else:
1875 1875 s = set([cl.rev(n) for t, n in repo.tagslist() if t != 'tip'])
1876 1876 return subset & s
1877 1877
1878 1878 def tagged(repo, subset, x):
1879 1879 return tag(repo, subset, x)
1880 1880
1881 1881 def unstable(repo, subset, x):
1882 1882 """``unstable()``
1883 1883 Non-obsolete changesets with obsolete ancestors.
1884 1884 """
1885 1885 # i18n: "unstable" is a keyword
1886 1886 getargs(x, 0, 0, _("unstable takes no arguments"))
1887 1887 unstables = obsmod.getrevs(repo, 'unstable')
1888 1888 return subset & unstables
1889 1889
1890 1890
1891 1891 def user(repo, subset, x):
1892 1892 """``user(string)``
1893 1893 User name contains string. The match is case-insensitive.
1894 1894
1895 1895 If `string` starts with `re:`, the remainder of the string is treated as
1896 1896 a regular expression. To match a user that actually contains `re:`, use
1897 1897 the prefix `literal:`.
1898 1898 """
1899 1899 return author(repo, subset, x)
1900 1900
1901 1901 # experimental
1902 1902 def wdir(repo, subset, x):
1903 1903 # i18n: "wdir" is a keyword
1904 1904 getargs(x, 0, 0, _("wdir takes no arguments"))
1905 1905 if None in subset:
1906 1906 return baseset([None])
1907 1907 return baseset()
1908 1908
1909 1909 # for internal use
1910 1910 def _list(repo, subset, x):
1911 1911 s = getstring(x, "internal error")
1912 1912 if not s:
1913 1913 return baseset()
1914 1914 ls = [repo[r].rev() for r in s.split('\0')]
1915 1915 s = subset
1916 1916 return baseset([r for r in ls if r in s])
1917 1917
1918 1918 # for internal use
1919 1919 def _intlist(repo, subset, x):
1920 1920 s = getstring(x, "internal error")
1921 1921 if not s:
1922 1922 return baseset()
1923 1923 ls = [int(r) for r in s.split('\0')]
1924 1924 s = subset
1925 1925 return baseset([r for r in ls if r in s])
1926 1926
1927 1927 # for internal use
1928 1928 def _hexlist(repo, subset, x):
1929 1929 s = getstring(x, "internal error")
1930 1930 if not s:
1931 1931 return baseset()
1932 1932 cl = repo.changelog
1933 1933 ls = [cl.rev(node.bin(r)) for r in s.split('\0')]
1934 1934 s = subset
1935 1935 return baseset([r for r in ls if r in s])
1936 1936
1937 1937 symbols = {
1938 1938 "adds": adds,
1939 1939 "all": getall,
1940 1940 "ancestor": ancestor,
1941 1941 "ancestors": ancestors,
1942 1942 "_firstancestors": _firstancestors,
1943 1943 "author": author,
1944 1944 "bisect": bisect,
1945 1945 "bisected": bisected,
1946 1946 "bookmark": bookmark,
1947 1947 "branch": branch,
1948 1948 "branchpoint": branchpoint,
1949 1949 "bumped": bumped,
1950 1950 "bundle": bundle,
1951 1951 "children": children,
1952 1952 "closed": closed,
1953 1953 "contains": contains,
1954 1954 "converted": converted,
1955 1955 "date": date,
1956 1956 "desc": desc,
1957 1957 "descendants": descendants,
1958 1958 "_firstdescendants": _firstdescendants,
1959 1959 "destination": destination,
1960 1960 "divergent": divergent,
1961 1961 "draft": draft,
1962 1962 "extinct": extinct,
1963 1963 "extra": extra,
1964 1964 "file": hasfile,
1965 1965 "filelog": filelog,
1966 1966 "first": first,
1967 1967 "follow": follow,
1968 1968 "_followfirst": _followfirst,
1969 1969 "grep": grep,
1970 1970 "head": head,
1971 1971 "heads": heads,
1972 1972 "hidden": hidden,
1973 1973 "id": node_,
1974 1974 "keyword": keyword,
1975 1975 "last": last,
1976 1976 "limit": limit,
1977 1977 "_matchfiles": _matchfiles,
1978 1978 "max": maxrev,
1979 1979 "merge": merge,
1980 1980 "min": minrev,
1981 1981 "modifies": modifies,
1982 1982 "named": named,
1983 1983 "obsolete": obsolete,
1984 1984 "only": only,
1985 1985 "origin": origin,
1986 1986 "outgoing": outgoing,
1987 1987 "p1": p1,
1988 1988 "p2": p2,
1989 1989 "parents": parents,
1990 1990 "present": present,
1991 1991 "public": public,
1992 1992 "remote": remote,
1993 1993 "removes": removes,
1994 1994 "rev": rev,
1995 1995 "reverse": reverse,
1996 1996 "roots": roots,
1997 1997 "sort": sort,
1998 1998 "secret": secret,
1999 1999 "subrepo": subrepo,
2000 2000 "matching": matching,
2001 2001 "tag": tag,
2002 2002 "tagged": tagged,
2003 2003 "user": user,
2004 2004 "unstable": unstable,
2005 2005 "wdir": wdir,
2006 2006 "_list": _list,
2007 2007 "_intlist": _intlist,
2008 2008 "_hexlist": _hexlist,
2009 2009 }
2010 2010
2011 2011 # symbols which can't be used for a DoS attack for any given input
2012 2012 # (e.g. those which accept regexes as plain strings shouldn't be included)
2013 2013 # functions that just return a lot of changesets (like all) don't count here
2014 2014 safesymbols = set([
2015 2015 "adds",
2016 2016 "all",
2017 2017 "ancestor",
2018 2018 "ancestors",
2019 2019 "_firstancestors",
2020 2020 "author",
2021 2021 "bisect",
2022 2022 "bisected",
2023 2023 "bookmark",
2024 2024 "branch",
2025 2025 "branchpoint",
2026 2026 "bumped",
2027 2027 "bundle",
2028 2028 "children",
2029 2029 "closed",
2030 2030 "converted",
2031 2031 "date",
2032 2032 "desc",
2033 2033 "descendants",
2034 2034 "_firstdescendants",
2035 2035 "destination",
2036 2036 "divergent",
2037 2037 "draft",
2038 2038 "extinct",
2039 2039 "extra",
2040 2040 "file",
2041 2041 "filelog",
2042 2042 "first",
2043 2043 "follow",
2044 2044 "_followfirst",
2045 2045 "head",
2046 2046 "heads",
2047 2047 "hidden",
2048 2048 "id",
2049 2049 "keyword",
2050 2050 "last",
2051 2051 "limit",
2052 2052 "_matchfiles",
2053 2053 "max",
2054 2054 "merge",
2055 2055 "min",
2056 2056 "modifies",
2057 2057 "obsolete",
2058 2058 "only",
2059 2059 "origin",
2060 2060 "outgoing",
2061 2061 "p1",
2062 2062 "p2",
2063 2063 "parents",
2064 2064 "present",
2065 2065 "public",
2066 2066 "remote",
2067 2067 "removes",
2068 2068 "rev",
2069 2069 "reverse",
2070 2070 "roots",
2071 2071 "sort",
2072 2072 "secret",
2073 2073 "matching",
2074 2074 "tag",
2075 2075 "tagged",
2076 2076 "user",
2077 2077 "unstable",
2078 2078 "wdir",
2079 2079 "_list",
2080 2080 "_intlist",
2081 2081 "_hexlist",
2082 2082 ])
2083 2083
2084 2084 methods = {
2085 2085 "range": rangeset,
2086 2086 "dagrange": dagrange,
2087 2087 "string": stringset,
2088 2088 "symbol": stringset,
2089 2089 "and": andset,
2090 2090 "or": orset,
2091 2091 "not": notset,
2092 2092 "list": listset,
2093 2093 "func": func,
2094 2094 "ancestor": ancestorspec,
2095 2095 "parent": parentspec,
2096 2096 "parentpost": p1,
2097 2097 }
2098 2098
2099 2099 def optimize(x, small):
2100 2100 if x is None:
2101 2101 return 0, x
2102 2102
2103 2103 smallbonus = 1
2104 2104 if small:
2105 2105 smallbonus = .5
2106 2106
2107 2107 op = x[0]
2108 2108 if op == 'minus':
2109 2109 return optimize(('and', x[1], ('not', x[2])), small)
2110 2110 elif op == 'only':
2111 2111 return optimize(('func', ('symbol', 'only'),
2112 2112 ('list', x[1], x[2])), small)
2113 2113 elif op == 'onlypost':
2114 2114 return optimize(('func', ('symbol', 'only'), x[1]), small)
2115 2115 elif op == 'dagrangepre':
2116 2116 return optimize(('func', ('symbol', 'ancestors'), x[1]), small)
2117 2117 elif op == 'dagrangepost':
2118 2118 return optimize(('func', ('symbol', 'descendants'), x[1]), small)
2119 2119 elif op == 'rangepre':
2120 2120 return optimize(('range', ('string', '0'), x[1]), small)
2121 2121 elif op == 'rangepost':
2122 2122 return optimize(('range', x[1], ('string', 'tip')), small)
2123 2123 elif op == 'negate':
2124 2124 return optimize(('string',
2125 2125 '-' + getstring(x[1], _("can't negate that"))), small)
2126 2126 elif op in 'string symbol negate':
2127 2127 return smallbonus, x # single revisions are small
2128 2128 elif op == 'and':
2129 2129 wa, ta = optimize(x[1], True)
2130 2130 wb, tb = optimize(x[2], True)
2131 2131
2132 2132 # (::x and not ::y)/(not ::y and ::x) have a fast path
2133 2133 def isonly(revs, bases):
2134 2134 return (
2135 2135 revs[0] == 'func'
2136 2136 and getstring(revs[1], _('not a symbol')) == 'ancestors'
2137 2137 and bases[0] == 'not'
2138 2138 and bases[1][0] == 'func'
2139 2139 and getstring(bases[1][1], _('not a symbol')) == 'ancestors')
2140 2140
2141 2141 w = min(wa, wb)
2142 2142 if isonly(ta, tb):
2143 2143 return w, ('func', ('symbol', 'only'), ('list', ta[2], tb[1][2]))
2144 2144 if isonly(tb, ta):
2145 2145 return w, ('func', ('symbol', 'only'), ('list', tb[2], ta[1][2]))
2146 2146
2147 2147 if wa > wb:
2148 2148 return w, (op, tb, ta)
2149 2149 return w, (op, ta, tb)
2150 2150 elif op == 'or':
2151 2151 wa, ta = optimize(x[1], False)
2152 2152 wb, tb = optimize(x[2], False)
2153 2153 if wb < wa:
2154 2154 wb, wa = wa, wb
2155 2155 return max(wa, wb), (op, ta, tb)
2156 2156 elif op == 'not':
2157 2157 o = optimize(x[1], not small)
2158 2158 return o[0], (op, o[1])
2159 2159 elif op == 'parentpost':
2160 2160 o = optimize(x[1], small)
2161 2161 return o[0], (op, o[1])
2162 2162 elif op == 'group':
2163 2163 return optimize(x[1], small)
2164 2164 elif op in 'dagrange range list parent ancestorspec':
2165 2165 if op == 'parent':
2166 2166 # x^:y means (x^) : y, not x ^ (:y)
2167 2167 post = ('parentpost', x[1])
2168 2168 if x[2][0] == 'dagrangepre':
2169 2169 return optimize(('dagrange', post, x[2][1]), small)
2170 2170 elif x[2][0] == 'rangepre':
2171 2171 return optimize(('range', post, x[2][1]), small)
2172 2172
2173 2173 wa, ta = optimize(x[1], small)
2174 2174 wb, tb = optimize(x[2], small)
2175 2175 return wa + wb, (op, ta, tb)
2176 2176 elif op == 'func':
2177 2177 f = getstring(x[1], _("not a symbol"))
2178 2178 wa, ta = optimize(x[2], small)
2179 2179 if f in ("author branch closed date desc file grep keyword "
2180 2180 "outgoing user"):
2181 2181 w = 10 # slow
2182 2182 elif f in "modifies adds removes":
2183 2183 w = 30 # slower
2184 2184 elif f == "contains":
2185 2185 w = 100 # very slow
2186 2186 elif f == "ancestor":
2187 2187 w = 1 * smallbonus
2188 2188 elif f in "reverse limit first _intlist":
2189 2189 w = 0
2190 2190 elif f in "sort":
2191 2191 w = 10 # assume most sorts look at changelog
2192 2192 else:
2193 2193 w = 1
2194 2194 return w + wa, (op, x[1], ta)
2195 2195 return 1, x
2196 2196
2197 2197 _aliasarg = ('func', ('symbol', '_aliasarg'))
2198 2198 def _getaliasarg(tree):
2199 2199 """If tree matches ('func', ('symbol', '_aliasarg'), ('string', X))
2200 2200 return X, None otherwise.
2201 2201 """
2202 2202 if (len(tree) == 3 and tree[:2] == _aliasarg
2203 2203 and tree[2][0] == 'string'):
2204 2204 return tree[2][1]
2205 2205 return None
2206 2206
2207 2207 def _checkaliasarg(tree, known=None):
2208 2208 """Check tree contains no _aliasarg construct or only ones which
2209 2209 value is in known. Used to avoid alias placeholders injection.
2210 2210 """
2211 2211 if isinstance(tree, tuple):
2212 2212 arg = _getaliasarg(tree)
2213 2213 if arg is not None and (not known or arg not in known):
2214 2214 raise error.UnknownIdentifier('_aliasarg', [])
2215 2215 for t in tree:
2216 2216 _checkaliasarg(t, known)
2217 2217
2218 2218 # the set of valid characters for the initial letter of symbols in
2219 2219 # alias declarations and definitions
2220 2220 _aliassyminitletters = set(c for c in [chr(i) for i in xrange(256)]
2221 2221 if c.isalnum() or c in '._@$' or ord(c) > 127)
2222 2222
2223 2223 def _tokenizealias(program, lookup=None):
2224 2224 """Parse alias declaration/definition into a stream of tokens
2225 2225
2226 2226 This allows symbol names to use also ``$`` as an initial letter
2227 2227 (for backward compatibility), and callers of this function should
2228 2228 examine whether ``$`` is used also for unexpected symbols or not.
2229 2229 """
2230 2230 return tokenize(program, lookup=lookup,
2231 2231 syminitletters=_aliassyminitletters)
2232 2232
2233 2233 def _parsealiasdecl(decl):
2234 2234 """Parse alias declaration ``decl``
2235 2235
2236 2236 This returns ``(name, tree, args, errorstr)`` tuple:
2237 2237
2238 2238 - ``name``: of declared alias (may be ``decl`` itself at error)
2239 2239 - ``tree``: parse result (or ``None`` at error)
2240 2240 - ``args``: list of alias argument names (or None for symbol declaration)
2241 2241 - ``errorstr``: detail about detected error (or None)
2242 2242
2243 2243 >>> _parsealiasdecl('foo')
2244 2244 ('foo', ('symbol', 'foo'), None, None)
2245 2245 >>> _parsealiasdecl('$foo')
2246 2246 ('$foo', None, None, "'$' not for alias arguments")
2247 2247 >>> _parsealiasdecl('foo::bar')
2248 2248 ('foo::bar', None, None, 'invalid format')
2249 2249 >>> _parsealiasdecl('foo bar')
2250 2250 ('foo bar', None, None, 'at 4: invalid token')
2251 2251 >>> _parsealiasdecl('foo()')
2252 2252 ('foo', ('func', ('symbol', 'foo')), [], None)
2253 2253 >>> _parsealiasdecl('$foo()')
2254 2254 ('$foo()', None, None, "'$' not for alias arguments")
2255 2255 >>> _parsealiasdecl('foo($1, $2)')
2256 2256 ('foo', ('func', ('symbol', 'foo')), ['$1', '$2'], None)
2257 2257 >>> _parsealiasdecl('foo(bar_bar, baz.baz)')
2258 2258 ('foo', ('func', ('symbol', 'foo')), ['bar_bar', 'baz.baz'], None)
2259 2259 >>> _parsealiasdecl('foo($1, $2, nested($1, $2))')
2260 2260 ('foo($1, $2, nested($1, $2))', None, None, 'invalid argument list')
2261 2261 >>> _parsealiasdecl('foo(bar($1, $2))')
2262 2262 ('foo(bar($1, $2))', None, None, 'invalid argument list')
2263 2263 >>> _parsealiasdecl('foo("string")')
2264 2264 ('foo("string")', None, None, 'invalid argument list')
2265 2265 >>> _parsealiasdecl('foo($1, $2')
2266 2266 ('foo($1, $2', None, None, 'at 10: unexpected token: end')
2267 2267 >>> _parsealiasdecl('foo("string')
2268 2268 ('foo("string', None, None, 'at 5: unterminated string')
2269 2269 >>> _parsealiasdecl('foo($1, $2, $1)')
2270 2270 ('foo', None, None, 'argument names collide with each other')
2271 2271 """
2272 2272 p = parser.parser(_tokenizealias, elements)
2273 2273 try:
2274 2274 tree, pos = p.parse(decl)
2275 2275 if (pos != len(decl)):
2276 2276 raise error.ParseError(_('invalid token'), pos)
2277 2277
2278 2278 if isvalidsymbol(tree):
2279 2279 # "name = ...." style
2280 2280 name = getsymbol(tree)
2281 2281 if name.startswith('$'):
2282 2282 return (decl, None, None, _("'$' not for alias arguments"))
2283 2283 return (name, ('symbol', name), None, None)
2284 2284
2285 2285 if isvalidfunc(tree):
2286 2286 # "name(arg, ....) = ...." style
2287 2287 name = getfuncname(tree)
2288 2288 if name.startswith('$'):
2289 2289 return (decl, None, None, _("'$' not for alias arguments"))
2290 2290 args = []
2291 2291 for arg in getfuncargs(tree):
2292 2292 if not isvalidsymbol(arg):
2293 2293 return (decl, None, None, _("invalid argument list"))
2294 2294 args.append(getsymbol(arg))
2295 2295 if len(args) != len(set(args)):
2296 2296 return (name, None, None,
2297 2297 _("argument names collide with each other"))
2298 2298 return (name, ('func', ('symbol', name)), args, None)
2299 2299
2300 2300 return (decl, None, None, _("invalid format"))
2301 2301 except error.ParseError, inst:
2302 2302 return (decl, None, None, parseerrordetail(inst))
2303 2303
2304 2304 def _parsealiasdefn(defn, args):
2305 2305 """Parse alias definition ``defn``
2306 2306
2307 2307 This function also replaces alias argument references in the
2308 2308 specified definition by ``_aliasarg(ARGNAME)``.
2309 2309
2310 2310 ``args`` is a list of alias argument names, or None if the alias
2311 2311 is declared as a symbol.
2312 2312
2313 2313 This returns "tree" as parsing result.
2314 2314
2315 2315 >>> args = ['$1', '$2', 'foo']
2316 2316 >>> print prettyformat(_parsealiasdefn('$1 or foo', args))
2317 2317 (or
2318 2318 (func
2319 2319 ('symbol', '_aliasarg')
2320 2320 ('string', '$1'))
2321 2321 (func
2322 2322 ('symbol', '_aliasarg')
2323 2323 ('string', 'foo')))
2324 2324 >>> try:
2325 2325 ... _parsealiasdefn('$1 or $bar', args)
2326 2326 ... except error.ParseError, inst:
2327 2327 ... print parseerrordetail(inst)
2328 2328 at 6: '$' not for alias arguments
2329 2329 >>> args = ['$1', '$10', 'foo']
2330 2330 >>> print prettyformat(_parsealiasdefn('$10 or foobar', args))
2331 2331 (or
2332 2332 (func
2333 2333 ('symbol', '_aliasarg')
2334 2334 ('string', '$10'))
2335 2335 ('symbol', 'foobar'))
2336 2336 >>> print prettyformat(_parsealiasdefn('"$1" or "foo"', args))
2337 2337 (or
2338 2338 ('string', '$1')
2339 2339 ('string', 'foo'))
2340 2340 """
2341 2341 def tokenizedefn(program, lookup=None):
2342 2342 if args:
2343 2343 argset = set(args)
2344 2344 else:
2345 2345 argset = set()
2346 2346
2347 2347 for t, value, pos in _tokenizealias(program, lookup=lookup):
2348 2348 if t == 'symbol':
2349 2349 if value in argset:
2350 2350 # emulate tokenization of "_aliasarg('ARGNAME')":
2351 2351 # "_aliasarg()" is an unknown symbol only used separate
2352 2352 # alias argument placeholders from regular strings.
2353 2353 yield ('symbol', '_aliasarg', pos)
2354 2354 yield ('(', None, pos)
2355 2355 yield ('string', value, pos)
2356 2356 yield (')', None, pos)
2357 2357 continue
2358 2358 elif value.startswith('$'):
2359 2359 raise error.ParseError(_("'$' not for alias arguments"),
2360 2360 pos)
2361 2361 yield (t, value, pos)
2362 2362
2363 2363 p = parser.parser(tokenizedefn, elements)
2364 2364 tree, pos = p.parse(defn)
2365 2365 if pos != len(defn):
2366 2366 raise error.ParseError(_('invalid token'), pos)
2367 2367 return tree
2368 2368
2369 2369 class revsetalias(object):
2370 2370 # whether own `error` information is already shown or not.
2371 2371 # this avoids showing same warning multiple times at each `findaliases`.
2372 2372 warned = False
2373 2373
2374 2374 def __init__(self, name, value):
2375 2375 '''Aliases like:
2376 2376
2377 2377 h = heads(default)
2378 2378 b($1) = ancestors($1) - ancestors(default)
2379 2379 '''
2380 2380 self.name, self.tree, self.args, self.error = _parsealiasdecl(name)
2381 2381 if self.error:
2382 2382 self.error = _('failed to parse the declaration of revset alias'
2383 2383 ' "%s": %s') % (self.name, self.error)
2384 2384 return
2385 2385
2386 2386 try:
2387 2387 self.replacement = _parsealiasdefn(value, self.args)
2388 2388 # Check for placeholder injection
2389 2389 _checkaliasarg(self.replacement, self.args)
2390 2390 except error.ParseError, inst:
2391 2391 self.error = _('failed to parse the definition of revset alias'
2392 2392 ' "%s": %s') % (self.name, parseerrordetail(inst))
2393 2393
2394 2394 def _getalias(aliases, tree):
2395 2395 """If tree looks like an unexpanded alias, return it. Return None
2396 2396 otherwise.
2397 2397 """
2398 2398 if isinstance(tree, tuple) and tree:
2399 2399 if tree[0] == 'symbol' and len(tree) == 2:
2400 2400 name = tree[1]
2401 2401 alias = aliases.get(name)
2402 2402 if alias and alias.args is None and alias.tree == tree:
2403 2403 return alias
2404 2404 if tree[0] == 'func' and len(tree) > 1:
2405 2405 if tree[1][0] == 'symbol' and len(tree[1]) == 2:
2406 2406 name = tree[1][1]
2407 2407 alias = aliases.get(name)
2408 2408 if alias and alias.args is not None and alias.tree == tree[:2]:
2409 2409 return alias
2410 2410 return None
2411 2411
2412 2412 def _expandargs(tree, args):
2413 2413 """Replace _aliasarg instances with the substitution value of the
2414 2414 same name in args, recursively.
2415 2415 """
2416 2416 if not tree or not isinstance(tree, tuple):
2417 2417 return tree
2418 2418 arg = _getaliasarg(tree)
2419 2419 if arg is not None:
2420 2420 return args[arg]
2421 2421 return tuple(_expandargs(t, args) for t in tree)
2422 2422
2423 2423 def _expandaliases(aliases, tree, expanding, cache):
2424 2424 """Expand aliases in tree, recursively.
2425 2425
2426 2426 'aliases' is a dictionary mapping user defined aliases to
2427 2427 revsetalias objects.
2428 2428 """
2429 2429 if not isinstance(tree, tuple):
2430 2430 # Do not expand raw strings
2431 2431 return tree
2432 2432 alias = _getalias(aliases, tree)
2433 2433 if alias is not None:
2434 2434 if alias.error:
2435 2435 raise util.Abort(alias.error)
2436 2436 if alias in expanding:
2437 2437 raise error.ParseError(_('infinite expansion of revset alias "%s" '
2438 2438 'detected') % alias.name)
2439 2439 expanding.append(alias)
2440 2440 if alias.name not in cache:
2441 2441 cache[alias.name] = _expandaliases(aliases, alias.replacement,
2442 2442 expanding, cache)
2443 2443 result = cache[alias.name]
2444 2444 expanding.pop()
2445 2445 if alias.args is not None:
2446 2446 l = getlist(tree[2])
2447 2447 if len(l) != len(alias.args):
2448 2448 raise error.ParseError(
2449 2449 _('invalid number of arguments: %s') % len(l))
2450 2450 l = [_expandaliases(aliases, a, [], cache) for a in l]
2451 2451 result = _expandargs(result, dict(zip(alias.args, l)))
2452 2452 else:
2453 2453 result = tuple(_expandaliases(aliases, t, expanding, cache)
2454 2454 for t in tree)
2455 2455 return result
2456 2456
2457 2457 def findaliases(ui, tree, showwarning=None):
2458 2458 _checkaliasarg(tree)
2459 2459 aliases = {}
2460 2460 for k, v in ui.configitems('revsetalias'):
2461 2461 alias = revsetalias(k, v)
2462 2462 aliases[alias.name] = alias
2463 2463 tree = _expandaliases(aliases, tree, [], {})
2464 2464 if showwarning:
2465 2465 # warn about problematic (but not referred) aliases
2466 2466 for name, alias in sorted(aliases.iteritems()):
2467 2467 if alias.error and not alias.warned:
2468 2468 showwarning(_('warning: %s\n') % (alias.error))
2469 2469 alias.warned = True
2470 2470 return tree
2471 2471
2472 2472 def foldconcat(tree):
2473 2473 """Fold elements to be concatenated by `##`
2474 2474 """
2475 2475 if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
2476 2476 return tree
2477 2477 if tree[0] == '_concat':
2478 2478 pending = [tree]
2479 2479 l = []
2480 2480 while pending:
2481 2481 e = pending.pop()
2482 2482 if e[0] == '_concat':
2483 2483 pending.extend(reversed(e[1:]))
2484 2484 elif e[0] in ('string', 'symbol'):
2485 2485 l.append(e[1])
2486 2486 else:
2487 2487 msg = _("\"##\" can't concatenate \"%s\" element") % (e[0])
2488 2488 raise error.ParseError(msg)
2489 2489 return ('string', ''.join(l))
2490 2490 else:
2491 2491 return tuple(foldconcat(t) for t in tree)
2492 2492
2493 2493 def parse(spec, lookup=None):
2494 2494 p = parser.parser(tokenize, elements)
2495 2495 return p.parse(spec, lookup=lookup)
2496 2496
2497 2497 def posttreebuilthook(tree, repo):
2498 2498 # hook for extensions to execute code on the optimized tree
2499 2499 pass
2500 2500
2501 2501 def match(ui, spec, repo=None):
2502 2502 if not spec:
2503 2503 raise error.ParseError(_("empty query"))
2504 2504 lookup = None
2505 2505 if repo:
2506 2506 lookup = repo.__contains__
2507 2507 tree, pos = parse(spec, lookup)
2508 2508 if (pos != len(spec)):
2509 2509 raise error.ParseError(_("invalid token"), pos)
2510 2510 if ui:
2511 2511 tree = findaliases(ui, tree, showwarning=ui.warn)
2512 2512 tree = foldconcat(tree)
2513 2513 weight, tree = optimize(tree, True)
2514 2514 posttreebuilthook(tree, repo)
2515 2515 def mfunc(repo, subset=None):
2516 2516 if subset is None:
2517 2517 subset = fullreposet(repo)
2518 2518 if util.safehasattr(subset, 'isascending'):
2519 2519 result = getset(repo, subset, tree)
2520 2520 else:
2521 2521 result = getset(repo, baseset(subset), tree)
2522 2522 return result
2523 2523 return mfunc
2524 2524
2525 2525 def formatspec(expr, *args):
2526 2526 '''
2527 2527 This is a convenience function for using revsets internally, and
2528 2528 escapes arguments appropriately. Aliases are intentionally ignored
2529 2529 so that intended expression behavior isn't accidentally subverted.
2530 2530
2531 2531 Supported arguments:
2532 2532
2533 2533 %r = revset expression, parenthesized
2534 2534 %d = int(arg), no quoting
2535 2535 %s = string(arg), escaped and single-quoted
2536 2536 %b = arg.branch(), escaped and single-quoted
2537 2537 %n = hex(arg), single-quoted
2538 2538 %% = a literal '%'
2539 2539
2540 2540 Prefixing the type with 'l' specifies a parenthesized list of that type.
2541 2541
2542 2542 >>> formatspec('%r:: and %lr', '10 or 11', ("this()", "that()"))
2543 2543 '(10 or 11):: and ((this()) or (that()))'
2544 2544 >>> formatspec('%d:: and not %d::', 10, 20)
2545 2545 '10:: and not 20::'
2546 2546 >>> formatspec('%ld or %ld', [], [1])
2547 2547 "_list('') or 1"
2548 2548 >>> formatspec('keyword(%s)', 'foo\\xe9')
2549 2549 "keyword('foo\\\\xe9')"
2550 2550 >>> b = lambda: 'default'
2551 2551 >>> b.branch = b
2552 2552 >>> formatspec('branch(%b)', b)
2553 2553 "branch('default')"
2554 2554 >>> formatspec('root(%ls)', ['a', 'b', 'c', 'd'])
2555 2555 "root(_list('a\\x00b\\x00c\\x00d'))"
2556 2556 '''
2557 2557
2558 2558 def quote(s):
2559 2559 return repr(str(s))
2560 2560
2561 2561 def argtype(c, arg):
2562 2562 if c == 'd':
2563 2563 return str(int(arg))
2564 2564 elif c == 's':
2565 2565 return quote(arg)
2566 2566 elif c == 'r':
2567 2567 parse(arg) # make sure syntax errors are confined
2568 2568 return '(%s)' % arg
2569 2569 elif c == 'n':
2570 2570 return quote(node.hex(arg))
2571 2571 elif c == 'b':
2572 2572 return quote(arg.branch())
2573 2573
2574 2574 def listexp(s, t):
2575 2575 l = len(s)
2576 2576 if l == 0:
2577 2577 return "_list('')"
2578 2578 elif l == 1:
2579 2579 return argtype(t, s[0])
2580 2580 elif t == 'd':
2581 2581 return "_intlist('%s')" % "\0".join(str(int(a)) for a in s)
2582 2582 elif t == 's':
2583 2583 return "_list('%s')" % "\0".join(s)
2584 2584 elif t == 'n':
2585 2585 return "_hexlist('%s')" % "\0".join(node.hex(a) for a in s)
2586 2586 elif t == 'b':
2587 2587 return "_list('%s')" % "\0".join(a.branch() for a in s)
2588 2588
2589 2589 m = l // 2
2590 2590 return '(%s or %s)' % (listexp(s[:m], t), listexp(s[m:], t))
2591 2591
2592 2592 ret = ''
2593 2593 pos = 0
2594 2594 arg = 0
2595 2595 while pos < len(expr):
2596 2596 c = expr[pos]
2597 2597 if c == '%':
2598 2598 pos += 1
2599 2599 d = expr[pos]
2600 2600 if d == '%':
2601 2601 ret += d
2602 2602 elif d in 'dsnbr':
2603 2603 ret += argtype(d, args[arg])
2604 2604 arg += 1
2605 2605 elif d == 'l':
2606 2606 # a list of some type
2607 2607 pos += 1
2608 2608 d = expr[pos]
2609 2609 ret += listexp(list(args[arg]), d)
2610 2610 arg += 1
2611 2611 else:
2612 2612 raise util.Abort('unexpected revspec format character %s' % d)
2613 2613 else:
2614 2614 ret += c
2615 2615 pos += 1
2616 2616
2617 2617 return ret
2618 2618
2619 2619 def prettyformat(tree):
2620 2620 def _prettyformat(tree, level, lines):
2621 2621 if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
2622 2622 lines.append((level, str(tree)))
2623 2623 else:
2624 2624 lines.append((level, '(%s' % tree[0]))
2625 2625 for s in tree[1:]:
2626 2626 _prettyformat(s, level + 1, lines)
2627 2627 lines[-1:] = [(lines[-1][0], lines[-1][1] + ')')]
2628 2628
2629 2629 lines = []
2630 2630 _prettyformat(tree, 0, lines)
2631 2631 output = '\n'.join((' '*l + s) for l, s in lines)
2632 2632 return output
2633 2633
2634 2634 def depth(tree):
2635 2635 if isinstance(tree, tuple):
2636 2636 return max(map(depth, tree)) + 1
2637 2637 else:
2638 2638 return 0
2639 2639
2640 2640 def funcsused(tree):
2641 2641 if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
2642 2642 return set()
2643 2643 else:
2644 2644 funcs = set()
2645 2645 for s in tree[1:]:
2646 2646 funcs |= funcsused(s)
2647 2647 if tree[0] == 'func':
2648 2648 funcs.add(tree[1][1])
2649 2649 return funcs
2650 2650
2651 2651 class abstractsmartset(object):
2652 2652
2653 2653 def __nonzero__(self):
2654 2654 """True if the smartset is not empty"""
2655 2655 raise NotImplementedError()
2656 2656
2657 2657 def __contains__(self, rev):
2658 2658 """provide fast membership testing"""
2659 2659 raise NotImplementedError()
2660 2660
2661 2661 def __iter__(self):
2662 2662 """iterate the set in the order it is supposed to be iterated"""
2663 2663 raise NotImplementedError()
2664 2664
2665 2665 # Attributes containing a function to perform a fast iteration in a given
2666 2666 # direction. A smartset can have none, one, or both defined.
2667 2667 #
2668 2668 # Default value is None instead of a function returning None to avoid
2669 2669 # initializing an iterator just for testing if a fast method exists.
2670 2670 fastasc = None
2671 2671 fastdesc = None
2672 2672
2673 2673 def isascending(self):
2674 2674 """True if the set will iterate in ascending order"""
2675 2675 raise NotImplementedError()
2676 2676
2677 2677 def isdescending(self):
2678 2678 """True if the set will iterate in descending order"""
2679 2679 raise NotImplementedError()
2680 2680
2681 2681 def min(self):
2682 2682 """return the minimum element in the set"""
2683 2683 if self.fastasc is not None:
2684 2684 for r in self.fastasc():
2685 2685 return r
2686 2686 raise ValueError('arg is an empty sequence')
2687 2687 return min(self)
2688 2688
2689 2689 def max(self):
2690 2690 """return the maximum element in the set"""
2691 2691 if self.fastdesc is not None:
2692 2692 for r in self.fastdesc():
2693 2693 return r
2694 2694 raise ValueError('arg is an empty sequence')
2695 2695 return max(self)
2696 2696
2697 2697 def first(self):
2698 2698 """return the first element in the set (user iteration perspective)
2699 2699
2700 2700 Return None if the set is empty"""
2701 2701 raise NotImplementedError()
2702 2702
2703 2703 def last(self):
2704 2704 """return the last element in the set (user iteration perspective)
2705 2705
2706 2706 Return None if the set is empty"""
2707 2707 raise NotImplementedError()
2708 2708
2709 2709 def __len__(self):
2710 2710 """return the length of the smartsets
2711 2711
2712 2712 This can be expensive on smartset that could be lazy otherwise."""
2713 2713 raise NotImplementedError()
2714 2714
2715 2715 def reverse(self):
2716 2716 """reverse the expected iteration order"""
2717 2717 raise NotImplementedError()
2718 2718
2719 2719 def sort(self, reverse=True):
2720 2720 """get the set to iterate in an ascending or descending order"""
2721 2721 raise NotImplementedError()
2722 2722
2723 2723 def __and__(self, other):
2724 2724 """Returns a new object with the intersection of the two collections.
2725 2725
2726 2726 This is part of the mandatory API for smartset."""
2727 2727 if isinstance(other, fullreposet):
2728 2728 return self
2729 2729 return self.filter(other.__contains__, cache=False)
2730 2730
2731 2731 def __add__(self, other):
2732 2732 """Returns a new object with the union of the two collections.
2733 2733
2734 2734 This is part of the mandatory API for smartset."""
2735 2735 return addset(self, other)
2736 2736
2737 2737 def __sub__(self, other):
2738 2738 """Returns a new object with the substraction of the two collections.
2739 2739
2740 2740 This is part of the mandatory API for smartset."""
2741 2741 c = other.__contains__
2742 2742 return self.filter(lambda r: not c(r), cache=False)
2743 2743
2744 2744 def filter(self, condition, cache=True):
2745 2745 """Returns this smartset filtered by condition as a new smartset.
2746 2746
2747 2747 `condition` is a callable which takes a revision number and returns a
2748 2748 boolean.
2749 2749
2750 2750 This is part of the mandatory API for smartset."""
2751 2751 # builtin cannot be cached. but do not needs to
2752 2752 if cache and util.safehasattr(condition, 'func_code'):
2753 2753 condition = util.cachefunc(condition)
2754 2754 return filteredset(self, condition)
2755 2755
2756 2756 class baseset(abstractsmartset):
2757 2757 """Basic data structure that represents a revset and contains the basic
2758 2758 operation that it should be able to perform.
2759 2759
2760 2760 Every method in this class should be implemented by any smartset class.
2761 2761 """
2762 2762 def __init__(self, data=()):
2763 2763 if not isinstance(data, list):
2764 2764 data = list(data)
2765 2765 self._list = data
2766 2766 self._ascending = None
2767 2767
2768 2768 @util.propertycache
2769 2769 def _set(self):
2770 2770 return set(self._list)
2771 2771
2772 2772 @util.propertycache
2773 2773 def _asclist(self):
2774 2774 asclist = self._list[:]
2775 2775 asclist.sort()
2776 2776 return asclist
2777 2777
2778 2778 def __iter__(self):
2779 2779 if self._ascending is None:
2780 2780 return iter(self._list)
2781 2781 elif self._ascending:
2782 2782 return iter(self._asclist)
2783 2783 else:
2784 2784 return reversed(self._asclist)
2785 2785
2786 2786 def fastasc(self):
2787 2787 return iter(self._asclist)
2788 2788
2789 2789 def fastdesc(self):
2790 2790 return reversed(self._asclist)
2791 2791
2792 2792 @util.propertycache
2793 2793 def __contains__(self):
2794 2794 return self._set.__contains__
2795 2795
2796 2796 def __nonzero__(self):
2797 2797 return bool(self._list)
2798 2798
2799 2799 def sort(self, reverse=False):
2800 2800 self._ascending = not bool(reverse)
2801 2801
2802 2802 def reverse(self):
2803 2803 if self._ascending is None:
2804 2804 self._list.reverse()
2805 2805 else:
2806 2806 self._ascending = not self._ascending
2807 2807
2808 2808 def __len__(self):
2809 2809 return len(self._list)
2810 2810
2811 2811 def isascending(self):
2812 2812 """Returns True if the collection is ascending order, False if not.
2813 2813
2814 2814 This is part of the mandatory API for smartset."""
2815 2815 if len(self) <= 1:
2816 2816 return True
2817 2817 return self._ascending is not None and self._ascending
2818 2818
2819 2819 def isdescending(self):
2820 2820 """Returns True if the collection is descending order, False if not.
2821 2821
2822 2822 This is part of the mandatory API for smartset."""
2823 2823 if len(self) <= 1:
2824 2824 return True
2825 2825 return self._ascending is not None and not self._ascending
2826 2826
2827 2827 def first(self):
2828 2828 if self:
2829 2829 if self._ascending is None:
2830 2830 return self._list[0]
2831 2831 elif self._ascending:
2832 2832 return self._asclist[0]
2833 2833 else:
2834 2834 return self._asclist[-1]
2835 2835 return None
2836 2836
2837 2837 def last(self):
2838 2838 if self:
2839 2839 if self._ascending is None:
2840 2840 return self._list[-1]
2841 2841 elif self._ascending:
2842 2842 return self._asclist[-1]
2843 2843 else:
2844 2844 return self._asclist[0]
2845 2845 return None
2846 2846
2847 2847 def __repr__(self):
2848 2848 d = {None: '', False: '-', True: '+'}[self._ascending]
2849 2849 return '<%s%s %r>' % (type(self).__name__, d, self._list)
2850 2850
2851 2851 class filteredset(abstractsmartset):
2852 2852 """Duck type for baseset class which iterates lazily over the revisions in
2853 2853 the subset and contains a function which tests for membership in the
2854 2854 revset
2855 2855 """
2856 2856 def __init__(self, subset, condition=lambda x: True):
2857 2857 """
2858 2858 condition: a function that decide whether a revision in the subset
2859 2859 belongs to the revset or not.
2860 2860 """
2861 2861 self._subset = subset
2862 2862 self._condition = condition
2863 2863 self._cache = {}
2864 2864
2865 2865 def __contains__(self, x):
2866 2866 c = self._cache
2867 2867 if x not in c:
2868 2868 v = c[x] = x in self._subset and self._condition(x)
2869 2869 return v
2870 2870 return c[x]
2871 2871
2872 2872 def __iter__(self):
2873 2873 return self._iterfilter(self._subset)
2874 2874
2875 2875 def _iterfilter(self, it):
2876 2876 cond = self._condition
2877 2877 for x in it:
2878 2878 if cond(x):
2879 2879 yield x
2880 2880
2881 2881 @property
2882 2882 def fastasc(self):
2883 2883 it = self._subset.fastasc
2884 2884 if it is None:
2885 2885 return None
2886 2886 return lambda: self._iterfilter(it())
2887 2887
2888 2888 @property
2889 2889 def fastdesc(self):
2890 2890 it = self._subset.fastdesc
2891 2891 if it is None:
2892 2892 return None
2893 2893 return lambda: self._iterfilter(it())
2894 2894
2895 2895 def __nonzero__(self):
2896 2896 for r in self:
2897 2897 return True
2898 2898 return False
2899 2899
2900 2900 def __len__(self):
2901 2901 # Basic implementation to be changed in future patches.
2902 2902 l = baseset([r for r in self])
2903 2903 return len(l)
2904 2904
2905 2905 def sort(self, reverse=False):
2906 2906 self._subset.sort(reverse=reverse)
2907 2907
2908 2908 def reverse(self):
2909 2909 self._subset.reverse()
2910 2910
2911 2911 def isascending(self):
2912 2912 return self._subset.isascending()
2913 2913
2914 2914 def isdescending(self):
2915 2915 return self._subset.isdescending()
2916 2916
2917 2917 def first(self):
2918 2918 for x in self:
2919 2919 return x
2920 2920 return None
2921 2921
2922 2922 def last(self):
2923 2923 it = None
2924 2924 if self._subset.isascending:
2925 2925 it = self.fastdesc
2926 2926 elif self._subset.isdescending:
2927 2927 it = self.fastdesc
2928 2928 if it is None:
2929 2929 # slowly consume everything. This needs improvement
2930 2930 it = lambda: reversed(list(self))
2931 2931 for x in it():
2932 2932 return x
2933 2933 return None
2934 2934
2935 2935 def __repr__(self):
2936 2936 return '<%s %r>' % (type(self).__name__, self._subset)
2937 2937
2938 2938 class addset(abstractsmartset):
2939 2939 """Represent the addition of two sets
2940 2940
2941 2941 Wrapper structure for lazily adding two structures without losing much
2942 2942 performance on the __contains__ method
2943 2943
2944 2944 If the ascending attribute is set, that means the two structures are
2945 2945 ordered in either an ascending or descending way. Therefore, we can add
2946 2946 them maintaining the order by iterating over both at the same time
2947 2947
2948 2948 >>> xs = baseset([0, 3, 2])
2949 2949 >>> ys = baseset([5, 2, 4])
2950 2950
2951 2951 >>> rs = addset(xs, ys)
2952 2952 >>> bool(rs), 0 in rs, 1 in rs, 5 in rs, rs.first(), rs.last()
2953 2953 (True, True, False, True, 0, 4)
2954 2954 >>> rs = addset(xs, baseset([]))
2955 2955 >>> bool(rs), 0 in rs, 1 in rs, rs.first(), rs.last()
2956 2956 (True, True, False, 0, 2)
2957 2957 >>> rs = addset(baseset([]), baseset([]))
2958 2958 >>> bool(rs), 0 in rs, rs.first(), rs.last()
2959 2959 (False, False, None, None)
2960 2960
2961 2961 iterate unsorted:
2962 2962 >>> rs = addset(xs, ys)
2963 2963 >>> [x for x in rs] # without _genlist
2964 2964 [0, 3, 2, 5, 4]
2965 2965 >>> assert not rs._genlist
2966 2966 >>> len(rs)
2967 2967 5
2968 2968 >>> [x for x in rs] # with _genlist
2969 2969 [0, 3, 2, 5, 4]
2970 2970 >>> assert rs._genlist
2971 2971
2972 2972 iterate ascending:
2973 2973 >>> rs = addset(xs, ys, ascending=True)
2974 2974 >>> [x for x in rs], [x for x in rs.fastasc()] # without _asclist
2975 2975 ([0, 2, 3, 4, 5], [0, 2, 3, 4, 5])
2976 2976 >>> assert not rs._asclist
2977 >>> len(rs) # BROKEN
2978 6
2979 >>> [x for x in rs], [x for x in rs.fastasc()] # BROKEN with _asclist
2980 ([0, 2, 2, 3, 4, 5], [0, 2, 2, 3, 4, 5])
2977 >>> len(rs)
2978 5
2979 >>> [x for x in rs], [x for x in rs.fastasc()]
2980 ([0, 2, 3, 4, 5], [0, 2, 3, 4, 5])
2981 2981 >>> assert rs._asclist
2982 2982
2983 2983 iterate descending:
2984 2984 >>> rs = addset(xs, ys, ascending=False)
2985 2985 >>> [x for x in rs], [x for x in rs.fastdesc()] # without _asclist
2986 2986 ([5, 4, 3, 2, 0], [5, 4, 3, 2, 0])
2987 2987 >>> assert not rs._asclist
2988 >>> len(rs) # BROKEN
2989 6
2990 >>> [x for x in rs], [x for x in rs.fastdesc()] # BROKEN with _asclist
2991 ([5, 4, 3, 2, 2, 0], [5, 4, 3, 2, 2, 0])
2988 >>> len(rs)
2989 5
2990 >>> [x for x in rs], [x for x in rs.fastdesc()]
2991 ([5, 4, 3, 2, 0], [5, 4, 3, 2, 0])
2992 2992 >>> assert rs._asclist
2993 2993
2994 2994 iterate ascending without fastasc:
2995 2995 >>> rs = addset(xs, generatorset(ys), ascending=True)
2996 2996 >>> assert rs.fastasc is None
2997 >>> [x for x in rs] # BROKEN
2998 [0, 2, 2, 3, 4, 5]
2997 >>> [x for x in rs]
2998 [0, 2, 3, 4, 5]
2999 2999
3000 3000 iterate descending without fastdesc:
3001 3001 >>> rs = addset(generatorset(xs), ys, ascending=False)
3002 3002 >>> assert rs.fastdesc is None
3003 >>> [x for x in rs] # BROKEN
3004 [5, 4, 3, 2, 2, 0]
3003 >>> [x for x in rs]
3004 [5, 4, 3, 2, 0]
3005 3005 """
3006 3006 def __init__(self, revs1, revs2, ascending=None):
3007 3007 self._r1 = revs1
3008 3008 self._r2 = revs2
3009 3009 self._iter = None
3010 3010 self._ascending = ascending
3011 3011 self._genlist = None
3012 3012 self._asclist = None
3013 3013
3014 3014 def __len__(self):
3015 3015 return len(self._list)
3016 3016
3017 3017 def __nonzero__(self):
3018 3018 return bool(self._r1) or bool(self._r2)
3019 3019
3020 3020 @util.propertycache
3021 3021 def _list(self):
3022 3022 if not self._genlist:
3023 self._genlist = baseset(self._iterator())
3023 self._genlist = baseset(iter(self))
3024 3024 return self._genlist
3025 3025
3026 def _iterator(self):
3026 def __iter__(self):
3027 3027 """Iterate over both collections without repeating elements
3028 3028
3029 3029 If the ascending attribute is not set, iterate over the first one and
3030 3030 then over the second one checking for membership on the first one so we
3031 3031 dont yield any duplicates.
3032 3032
3033 3033 If the ascending attribute is set, iterate over both collections at the
3034 3034 same time, yielding only one value at a time in the given order.
3035 3035 """
3036 3036 if self._ascending is None:
3037 def gen():
3037 if self._genlist:
3038 return iter(self._genlist)
3039 def arbitraryordergen():
3038 3040 for r in self._r1:
3039 3041 yield r
3040 3042 inr1 = self._r1.__contains__
3041 3043 for r in self._r2:
3042 3044 if not inr1(r):
3043 3045 yield r
3044 gen = gen()
3045 else:
3046 iter1 = iter(self._r1)
3047 iter2 = iter(self._r2)
3048 gen = self._iterordered(self._ascending, iter1, iter2)
3049 return gen
3050
3051 def __iter__(self):
3052 if self._ascending is None:
3053 if self._genlist:
3054 return iter(self._genlist)
3055 return iter(self._iterator())
3046 return arbitraryordergen()
3047 # try to use our own fast iterator if it exists
3056 3048 self._trysetasclist()
3057 3049 if self._ascending:
3058 3050 it = self.fastasc
3059 3051 else:
3060 3052 it = self.fastdesc
3061 if it is None:
3062 # consume the gen and try again
3063 self._list
3064 return iter(self)
3065 return it()
3053 if it is not None:
3054 return it()
3055 # maybe half of the component supports fast
3056 attr = 'fastdesc'
3057 if self._ascending:
3058 attr = 'fastasc'
3059 # get iterator for _r1
3060 iter1 = getattr(self._r1, attr)
3061 if iter1 is None:
3062 # let's avoid side effect (not sure it matters)
3063 iter1 = iter(sorted(self._r1, reverse=not self._ascending))
3064 else:
3065 iter1 = iter1()
3066 # get iterator for _r2
3067 iter2 = getattr(self._r2, attr)
3068 if iter2 is None:
3069 # let's avoid side effect (not sure it matters)
3070 iter2 = iter(sorted(self._r2, reverse=not self._ascending))
3071 else:
3072 iter2 = iter2()
3073 return self._iterordered(self._ascending, iter1, iter2)
3066 3074
3067 3075 def _trysetasclist(self):
3068 3076 """populate the _asclist attribute if possible and necessary"""
3069 3077 if self._genlist is not None and self._asclist is None:
3070 3078 self._asclist = sorted(self._genlist)
3071 3079
3072 3080 @property
3073 3081 def fastasc(self):
3074 3082 self._trysetasclist()
3075 3083 if self._asclist is not None:
3076 3084 return self._asclist.__iter__
3077 3085 iter1 = self._r1.fastasc
3078 3086 iter2 = self._r2.fastasc
3079 3087 if None in (iter1, iter2):
3080 3088 return None
3081 3089 return lambda: self._iterordered(True, iter1(), iter2())
3082 3090
3083 3091 @property
3084 3092 def fastdesc(self):
3085 3093 self._trysetasclist()
3086 3094 if self._asclist is not None:
3087 3095 return self._asclist.__reversed__
3088 3096 iter1 = self._r1.fastdesc
3089 3097 iter2 = self._r2.fastdesc
3090 3098 if None in (iter1, iter2):
3091 3099 return None
3092 3100 return lambda: self._iterordered(False, iter1(), iter2())
3093 3101
3094 3102 def _iterordered(self, ascending, iter1, iter2):
3095 3103 """produce an ordered iteration from two iterators with the same order
3096 3104
3097 3105 The ascending is used to indicated the iteration direction.
3098 3106 """
3099 3107 choice = max
3100 3108 if ascending:
3101 3109 choice = min
3102 3110
3103 3111 val1 = None
3104 3112 val2 = None
3105 3113 try:
3106 3114 # Consume both iterators in an ordered way until one is
3107 3115 # empty
3108 3116 while True:
3109 3117 if val1 is None:
3110 3118 val1 = iter1.next()
3111 3119 if val2 is None:
3112 3120 val2 = iter2.next()
3113 3121 next = choice(val1, val2)
3114 3122 yield next
3115 3123 if val1 == next:
3116 3124 val1 = None
3117 3125 if val2 == next:
3118 3126 val2 = None
3119 3127 except StopIteration:
3120 3128 # Flush any remaining values and consume the other one
3121 3129 it = iter2
3122 3130 if val1 is not None:
3123 3131 yield val1
3124 3132 it = iter1
3125 3133 elif val2 is not None:
3126 3134 # might have been equality and both are empty
3127 3135 yield val2
3128 3136 for val in it:
3129 3137 yield val
3130 3138
3131 3139 def __contains__(self, x):
3132 3140 return x in self._r1 or x in self._r2
3133 3141
3134 3142 def sort(self, reverse=False):
3135 3143 """Sort the added set
3136 3144
3137 3145 For this we use the cached list with all the generated values and if we
3138 3146 know they are ascending or descending we can sort them in a smart way.
3139 3147 """
3140 3148 self._ascending = not reverse
3141 3149
3142 3150 def isascending(self):
3143 3151 return self._ascending is not None and self._ascending
3144 3152
3145 3153 def isdescending(self):
3146 3154 return self._ascending is not None and not self._ascending
3147 3155
3148 3156 def reverse(self):
3149 3157 if self._ascending is None:
3150 3158 self._list.reverse()
3151 3159 else:
3152 3160 self._ascending = not self._ascending
3153 3161
3154 3162 def first(self):
3155 3163 for x in self:
3156 3164 return x
3157 3165 return None
3158 3166
3159 3167 def last(self):
3160 3168 self.reverse()
3161 3169 val = self.first()
3162 3170 self.reverse()
3163 3171 return val
3164 3172
3165 3173 def __repr__(self):
3166 3174 d = {None: '', False: '-', True: '+'}[self._ascending]
3167 3175 return '<%s%s %r, %r>' % (type(self).__name__, d, self._r1, self._r2)
3168 3176
3169 3177 class generatorset(abstractsmartset):
3170 3178 """Wrap a generator for lazy iteration
3171 3179
3172 3180 Wrapper structure for generators that provides lazy membership and can
3173 3181 be iterated more than once.
3174 3182 When asked for membership it generates values until either it finds the
3175 3183 requested one or has gone through all the elements in the generator
3176 3184 """
3177 3185 def __init__(self, gen, iterasc=None):
3178 3186 """
3179 3187 gen: a generator producing the values for the generatorset.
3180 3188 """
3181 3189 self._gen = gen
3182 3190 self._asclist = None
3183 3191 self._cache = {}
3184 3192 self._genlist = []
3185 3193 self._finished = False
3186 3194 self._ascending = True
3187 3195 if iterasc is not None:
3188 3196 if iterasc:
3189 3197 self.fastasc = self._iterator
3190 3198 self.__contains__ = self._asccontains
3191 3199 else:
3192 3200 self.fastdesc = self._iterator
3193 3201 self.__contains__ = self._desccontains
3194 3202
3195 3203 def __nonzero__(self):
3196 3204 # Do not use 'for r in self' because it will enforce the iteration
3197 3205 # order (default ascending), possibly unrolling a whole descending
3198 3206 # iterator.
3199 3207 if self._genlist:
3200 3208 return True
3201 3209 for r in self._consumegen():
3202 3210 return True
3203 3211 return False
3204 3212
3205 3213 def __contains__(self, x):
3206 3214 if x in self._cache:
3207 3215 return self._cache[x]
3208 3216
3209 3217 # Use new values only, as existing values would be cached.
3210 3218 for l in self._consumegen():
3211 3219 if l == x:
3212 3220 return True
3213 3221
3214 3222 self._cache[x] = False
3215 3223 return False
3216 3224
3217 3225 def _asccontains(self, x):
3218 3226 """version of contains optimised for ascending generator"""
3219 3227 if x in self._cache:
3220 3228 return self._cache[x]
3221 3229
3222 3230 # Use new values only, as existing values would be cached.
3223 3231 for l in self._consumegen():
3224 3232 if l == x:
3225 3233 return True
3226 3234 if l > x:
3227 3235 break
3228 3236
3229 3237 self._cache[x] = False
3230 3238 return False
3231 3239
3232 3240 def _desccontains(self, x):
3233 3241 """version of contains optimised for descending generator"""
3234 3242 if x in self._cache:
3235 3243 return self._cache[x]
3236 3244
3237 3245 # Use new values only, as existing values would be cached.
3238 3246 for l in self._consumegen():
3239 3247 if l == x:
3240 3248 return True
3241 3249 if l < x:
3242 3250 break
3243 3251
3244 3252 self._cache[x] = False
3245 3253 return False
3246 3254
3247 3255 def __iter__(self):
3248 3256 if self._ascending:
3249 3257 it = self.fastasc
3250 3258 else:
3251 3259 it = self.fastdesc
3252 3260 if it is not None:
3253 3261 return it()
3254 3262 # we need to consume the iterator
3255 3263 for x in self._consumegen():
3256 3264 pass
3257 3265 # recall the same code
3258 3266 return iter(self)
3259 3267
3260 3268 def _iterator(self):
3261 3269 if self._finished:
3262 3270 return iter(self._genlist)
3263 3271
3264 3272 # We have to use this complex iteration strategy to allow multiple
3265 3273 # iterations at the same time. We need to be able to catch revision
3266 3274 # removed from _consumegen and added to genlist in another instance.
3267 3275 #
3268 3276 # Getting rid of it would provide an about 15% speed up on this
3269 3277 # iteration.
3270 3278 genlist = self._genlist
3271 3279 nextrev = self._consumegen().next
3272 3280 _len = len # cache global lookup
3273 3281 def gen():
3274 3282 i = 0
3275 3283 while True:
3276 3284 if i < _len(genlist):
3277 3285 yield genlist[i]
3278 3286 else:
3279 3287 yield nextrev()
3280 3288 i += 1
3281 3289 return gen()
3282 3290
3283 3291 def _consumegen(self):
3284 3292 cache = self._cache
3285 3293 genlist = self._genlist.append
3286 3294 for item in self._gen:
3287 3295 cache[item] = True
3288 3296 genlist(item)
3289 3297 yield item
3290 3298 if not self._finished:
3291 3299 self._finished = True
3292 3300 asc = self._genlist[:]
3293 3301 asc.sort()
3294 3302 self._asclist = asc
3295 3303 self.fastasc = asc.__iter__
3296 3304 self.fastdesc = asc.__reversed__
3297 3305
3298 3306 def __len__(self):
3299 3307 for x in self._consumegen():
3300 3308 pass
3301 3309 return len(self._genlist)
3302 3310
3303 3311 def sort(self, reverse=False):
3304 3312 self._ascending = not reverse
3305 3313
3306 3314 def reverse(self):
3307 3315 self._ascending = not self._ascending
3308 3316
3309 3317 def isascending(self):
3310 3318 return self._ascending
3311 3319
3312 3320 def isdescending(self):
3313 3321 return not self._ascending
3314 3322
3315 3323 def first(self):
3316 3324 if self._ascending:
3317 3325 it = self.fastasc
3318 3326 else:
3319 3327 it = self.fastdesc
3320 3328 if it is None:
3321 3329 # we need to consume all and try again
3322 3330 for x in self._consumegen():
3323 3331 pass
3324 3332 return self.first()
3325 3333 if self:
3326 3334 return it().next()
3327 3335 return None
3328 3336
3329 3337 def last(self):
3330 3338 if self._ascending:
3331 3339 it = self.fastdesc
3332 3340 else:
3333 3341 it = self.fastasc
3334 3342 if it is None:
3335 3343 # we need to consume all and try again
3336 3344 for x in self._consumegen():
3337 3345 pass
3338 3346 return self.first()
3339 3347 if self:
3340 3348 return it().next()
3341 3349 return None
3342 3350
3343 3351 def __repr__(self):
3344 3352 d = {False: '-', True: '+'}[self._ascending]
3345 3353 return '<%s%s>' % (type(self).__name__, d)
3346 3354
3347 3355 class spanset(abstractsmartset):
3348 3356 """Duck type for baseset class which represents a range of revisions and
3349 3357 can work lazily and without having all the range in memory
3350 3358
3351 3359 Note that spanset(x, y) behave almost like xrange(x, y) except for two
3352 3360 notable points:
3353 3361 - when x < y it will be automatically descending,
3354 3362 - revision filtered with this repoview will be skipped.
3355 3363
3356 3364 """
3357 3365 def __init__(self, repo, start=0, end=None):
3358 3366 """
3359 3367 start: first revision included the set
3360 3368 (default to 0)
3361 3369 end: first revision excluded (last+1)
3362 3370 (default to len(repo)
3363 3371
3364 3372 Spanset will be descending if `end` < `start`.
3365 3373 """
3366 3374 if end is None:
3367 3375 end = len(repo)
3368 3376 self._ascending = start <= end
3369 3377 if not self._ascending:
3370 3378 start, end = end + 1, start +1
3371 3379 self._start = start
3372 3380 self._end = end
3373 3381 self._hiddenrevs = repo.changelog.filteredrevs
3374 3382
3375 3383 def sort(self, reverse=False):
3376 3384 self._ascending = not reverse
3377 3385
3378 3386 def reverse(self):
3379 3387 self._ascending = not self._ascending
3380 3388
3381 3389 def _iterfilter(self, iterrange):
3382 3390 s = self._hiddenrevs
3383 3391 for r in iterrange:
3384 3392 if r not in s:
3385 3393 yield r
3386 3394
3387 3395 def __iter__(self):
3388 3396 if self._ascending:
3389 3397 return self.fastasc()
3390 3398 else:
3391 3399 return self.fastdesc()
3392 3400
3393 3401 def fastasc(self):
3394 3402 iterrange = xrange(self._start, self._end)
3395 3403 if self._hiddenrevs:
3396 3404 return self._iterfilter(iterrange)
3397 3405 return iter(iterrange)
3398 3406
3399 3407 def fastdesc(self):
3400 3408 iterrange = xrange(self._end - 1, self._start - 1, -1)
3401 3409 if self._hiddenrevs:
3402 3410 return self._iterfilter(iterrange)
3403 3411 return iter(iterrange)
3404 3412
3405 3413 def __contains__(self, rev):
3406 3414 hidden = self._hiddenrevs
3407 3415 return ((self._start <= rev < self._end)
3408 3416 and not (hidden and rev in hidden))
3409 3417
3410 3418 def __nonzero__(self):
3411 3419 for r in self:
3412 3420 return True
3413 3421 return False
3414 3422
3415 3423 def __len__(self):
3416 3424 if not self._hiddenrevs:
3417 3425 return abs(self._end - self._start)
3418 3426 else:
3419 3427 count = 0
3420 3428 start = self._start
3421 3429 end = self._end
3422 3430 for rev in self._hiddenrevs:
3423 3431 if (end < rev <= start) or (start <= rev < end):
3424 3432 count += 1
3425 3433 return abs(self._end - self._start) - count
3426 3434
3427 3435 def isascending(self):
3428 3436 return self._ascending
3429 3437
3430 3438 def isdescending(self):
3431 3439 return not self._ascending
3432 3440
3433 3441 def first(self):
3434 3442 if self._ascending:
3435 3443 it = self.fastasc
3436 3444 else:
3437 3445 it = self.fastdesc
3438 3446 for x in it():
3439 3447 return x
3440 3448 return None
3441 3449
3442 3450 def last(self):
3443 3451 if self._ascending:
3444 3452 it = self.fastdesc
3445 3453 else:
3446 3454 it = self.fastasc
3447 3455 for x in it():
3448 3456 return x
3449 3457 return None
3450 3458
3451 3459 def __repr__(self):
3452 3460 d = {False: '-', True: '+'}[self._ascending]
3453 3461 return '<%s%s %d:%d>' % (type(self).__name__, d,
3454 3462 self._start, self._end - 1)
3455 3463
3456 3464 class fullreposet(spanset):
3457 3465 """a set containing all revisions in the repo
3458 3466
3459 3467 This class exists to host special optimization and magic to handle virtual
3460 3468 revisions such as "null".
3461 3469 """
3462 3470
3463 3471 def __init__(self, repo):
3464 3472 super(fullreposet, self).__init__(repo)
3465 3473
3466 3474 def __contains__(self, rev):
3467 3475 # assumes the given rev is valid
3468 3476 hidden = self._hiddenrevs
3469 3477 return not (hidden and rev in hidden)
3470 3478
3471 3479 def __and__(self, other):
3472 3480 """As self contains the whole repo, all of the other set should also be
3473 3481 in self. Therefore `self & other = other`.
3474 3482
3475 3483 This boldly assumes the other contains valid revs only.
3476 3484 """
3477 3485 # other not a smartset, make is so
3478 3486 if not util.safehasattr(other, 'isascending'):
3479 3487 # filter out hidden revision
3480 3488 # (this boldly assumes all smartset are pure)
3481 3489 #
3482 3490 # `other` was used with "&", let's assume this is a set like
3483 3491 # object.
3484 3492 other = baseset(other - self._hiddenrevs)
3485 3493
3486 3494 other.sort(reverse=self.isdescending())
3487 3495 return other
3488 3496
3489 3497 def prettyformatset(revs):
3490 3498 lines = []
3491 3499 rs = repr(revs)
3492 3500 p = 0
3493 3501 while p < len(rs):
3494 3502 q = rs.find('<', p + 1)
3495 3503 if q < 0:
3496 3504 q = len(rs)
3497 3505 l = rs.count('<', 0, p) - rs.count('>', 0, p)
3498 3506 assert l >= 0
3499 3507 lines.append((l, rs[p:q].rstrip()))
3500 3508 p = q
3501 3509 return '\n'.join(' ' * l + s for l, s in lines)
3502 3510
3503 3511 # tell hggettext to extract docstrings from these functions:
3504 3512 i18nfunctions = symbols.values()
General Comments 0
You need to be logged in to leave comments. Login now