##// END OF EJS Templates
revset: remove nullrev from the bookmark computation...
Pierre-Yves David -
r22500:171015d3 default
parent child Browse files
Show More
@@ -1,2850 +1,2851
1 1 # revset.py - revision set queries for mercurial
2 2 #
3 3 # Copyright 2010 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 import re
9 9 import parser, util, error, discovery, hbisect, phases
10 10 import node
11 11 import heapq
12 12 import match as matchmod
13 13 import ancestor as ancestormod
14 14 from i18n import _
15 15 import encoding
16 16 import obsolete as obsmod
17 17 import pathutil
18 18 import repoview
19 19
20 20 def _revancestors(repo, revs, followfirst):
21 21 """Like revlog.ancestors(), but supports followfirst."""
22 22 cut = followfirst and 1 or None
23 23 cl = repo.changelog
24 24
25 25 def iterate():
26 26 revqueue, revsnode = None, None
27 27 h = []
28 28
29 29 revs.descending()
30 30 revqueue = util.deque(revs)
31 31 if revqueue:
32 32 revsnode = revqueue.popleft()
33 33 heapq.heappush(h, -revsnode)
34 34
35 35 seen = set([node.nullrev])
36 36 while h:
37 37 current = -heapq.heappop(h)
38 38 if current not in seen:
39 39 if revsnode and current == revsnode:
40 40 if revqueue:
41 41 revsnode = revqueue.popleft()
42 42 heapq.heappush(h, -revsnode)
43 43 seen.add(current)
44 44 yield current
45 45 for parent in cl.parentrevs(current)[:cut]:
46 46 if parent != node.nullrev:
47 47 heapq.heappush(h, -parent)
48 48
49 49 return _descgeneratorset(iterate())
50 50
51 51 def _revdescendants(repo, revs, followfirst):
52 52 """Like revlog.descendants() but supports followfirst."""
53 53 cut = followfirst and 1 or None
54 54
55 55 def iterate():
56 56 cl = repo.changelog
57 57 first = min(revs)
58 58 nullrev = node.nullrev
59 59 if first == nullrev:
60 60 # Are there nodes with a null first parent and a non-null
61 61 # second one? Maybe. Do we care? Probably not.
62 62 for i in cl:
63 63 yield i
64 64 else:
65 65 seen = set(revs)
66 66 for i in cl.revs(first + 1):
67 67 for x in cl.parentrevs(i)[:cut]:
68 68 if x != nullrev and x in seen:
69 69 seen.add(i)
70 70 yield i
71 71 break
72 72
73 73 return _ascgeneratorset(iterate())
74 74
75 75 def _revsbetween(repo, roots, heads):
76 76 """Return all paths between roots and heads, inclusive of both endpoint
77 77 sets."""
78 78 if not roots:
79 79 return baseset([])
80 80 parentrevs = repo.changelog.parentrevs
81 81 visit = list(heads)
82 82 reachable = set()
83 83 seen = {}
84 84 minroot = min(roots)
85 85 roots = set(roots)
86 86 # open-code the post-order traversal due to the tiny size of
87 87 # sys.getrecursionlimit()
88 88 while visit:
89 89 rev = visit.pop()
90 90 if rev in roots:
91 91 reachable.add(rev)
92 92 parents = parentrevs(rev)
93 93 seen[rev] = parents
94 94 for parent in parents:
95 95 if parent >= minroot and parent not in seen:
96 96 visit.append(parent)
97 97 if not reachable:
98 98 return baseset([])
99 99 for rev in sorted(seen):
100 100 for parent in seen[rev]:
101 101 if parent in reachable:
102 102 reachable.add(rev)
103 103 return baseset(sorted(reachable))
104 104
105 105 elements = {
106 106 "(": (20, ("group", 1, ")"), ("func", 1, ")")),
107 107 "~": (18, None, ("ancestor", 18)),
108 108 "^": (18, None, ("parent", 18), ("parentpost", 18)),
109 109 "-": (5, ("negate", 19), ("minus", 5)),
110 110 "::": (17, ("dagrangepre", 17), ("dagrange", 17),
111 111 ("dagrangepost", 17)),
112 112 "..": (17, ("dagrangepre", 17), ("dagrange", 17),
113 113 ("dagrangepost", 17)),
114 114 ":": (15, ("rangepre", 15), ("range", 15), ("rangepost", 15)),
115 115 "not": (10, ("not", 10)),
116 116 "!": (10, ("not", 10)),
117 117 "and": (5, None, ("and", 5)),
118 118 "&": (5, None, ("and", 5)),
119 119 "or": (4, None, ("or", 4)),
120 120 "|": (4, None, ("or", 4)),
121 121 "+": (4, None, ("or", 4)),
122 122 ",": (2, None, ("list", 2)),
123 123 ")": (0, None, None),
124 124 "symbol": (0, ("symbol",), None),
125 125 "string": (0, ("string",), None),
126 126 "end": (0, None, None),
127 127 }
128 128
129 129 keywords = set(['and', 'or', 'not'])
130 130
131 131 def tokenize(program, lookup=None):
132 132 '''
133 133 Parse a revset statement into a stream of tokens
134 134
135 135 Check that @ is a valid unquoted token character (issue3686):
136 136 >>> list(tokenize("@::"))
137 137 [('symbol', '@', 0), ('::', None, 1), ('end', None, 3)]
138 138
139 139 '''
140 140
141 141 pos, l = 0, len(program)
142 142 while pos < l:
143 143 c = program[pos]
144 144 if c.isspace(): # skip inter-token whitespace
145 145 pass
146 146 elif c == ':' and program[pos:pos + 2] == '::': # look ahead carefully
147 147 yield ('::', None, pos)
148 148 pos += 1 # skip ahead
149 149 elif c == '.' and program[pos:pos + 2] == '..': # look ahead carefully
150 150 yield ('..', None, pos)
151 151 pos += 1 # skip ahead
152 152 elif c in "():,-|&+!~^": # handle simple operators
153 153 yield (c, None, pos)
154 154 elif (c in '"\'' or c == 'r' and
155 155 program[pos:pos + 2] in ("r'", 'r"')): # handle quoted strings
156 156 if c == 'r':
157 157 pos += 1
158 158 c = program[pos]
159 159 decode = lambda x: x
160 160 else:
161 161 decode = lambda x: x.decode('string-escape')
162 162 pos += 1
163 163 s = pos
164 164 while pos < l: # find closing quote
165 165 d = program[pos]
166 166 if d == '\\': # skip over escaped characters
167 167 pos += 2
168 168 continue
169 169 if d == c:
170 170 yield ('string', decode(program[s:pos]), s)
171 171 break
172 172 pos += 1
173 173 else:
174 174 raise error.ParseError(_("unterminated string"), s)
175 175 # gather up a symbol/keyword
176 176 elif c.isalnum() or c in '._@' or ord(c) > 127:
177 177 s = pos
178 178 pos += 1
179 179 while pos < l: # find end of symbol
180 180 d = program[pos]
181 181 if not (d.isalnum() or d in "-._/@" or ord(d) > 127):
182 182 break
183 183 if d == '.' and program[pos - 1] == '.': # special case for ..
184 184 pos -= 1
185 185 break
186 186 pos += 1
187 187 sym = program[s:pos]
188 188 if sym in keywords: # operator keywords
189 189 yield (sym, None, s)
190 190 elif '-' in sym:
191 191 # some jerk gave us foo-bar-baz, try to check if it's a symbol
192 192 if lookup and lookup(sym):
193 193 # looks like a real symbol
194 194 yield ('symbol', sym, s)
195 195 else:
196 196 # looks like an expression
197 197 parts = sym.split('-')
198 198 for p in parts[:-1]:
199 199 if p: # possible consecutive -
200 200 yield ('symbol', p, s)
201 201 s += len(p)
202 202 yield ('-', None, pos)
203 203 s += 1
204 204 if parts[-1]: # possible trailing -
205 205 yield ('symbol', parts[-1], s)
206 206 else:
207 207 yield ('symbol', sym, s)
208 208 pos -= 1
209 209 else:
210 210 raise error.ParseError(_("syntax error"), pos)
211 211 pos += 1
212 212 yield ('end', None, pos)
213 213
214 214 # helpers
215 215
216 216 def getstring(x, err):
217 217 if x and (x[0] == 'string' or x[0] == 'symbol'):
218 218 return x[1]
219 219 raise error.ParseError(err)
220 220
221 221 def getlist(x):
222 222 if not x:
223 223 return []
224 224 if x[0] == 'list':
225 225 return getlist(x[1]) + [x[2]]
226 226 return [x]
227 227
228 228 def getargs(x, min, max, err):
229 229 l = getlist(x)
230 230 if len(l) < min or (max >= 0 and len(l) > max):
231 231 raise error.ParseError(err)
232 232 return l
233 233
234 234 def getset(repo, subset, x):
235 235 if not x:
236 236 raise error.ParseError(_("missing argument"))
237 237 s = methods[x[0]](repo, subset, *x[1:])
238 238 if util.safehasattr(s, 'set'):
239 239 return s
240 240 return baseset(s)
241 241
242 242 def _getrevsource(repo, r):
243 243 extra = repo[r].extra()
244 244 for label in ('source', 'transplant_source', 'rebase_source'):
245 245 if label in extra:
246 246 try:
247 247 return repo[extra[label]].rev()
248 248 except error.RepoLookupError:
249 249 pass
250 250 return None
251 251
252 252 # operator methods
253 253
254 254 def stringset(repo, subset, x):
255 255 x = repo[x].rev()
256 256 if x == -1 and len(subset) == len(repo):
257 257 return baseset([-1])
258 258 if len(subset) == len(repo) or x in subset:
259 259 return baseset([x])
260 260 return baseset([])
261 261
262 262 def symbolset(repo, subset, x):
263 263 if x in symbols:
264 264 raise error.ParseError(_("can't use %s here") % x)
265 265 return stringset(repo, subset, x)
266 266
267 267 def rangeset(repo, subset, x, y):
268 268 cl = baseset(repo.changelog)
269 269 m = getset(repo, cl, x)
270 270 n = getset(repo, cl, y)
271 271
272 272 if not m or not n:
273 273 return baseset([])
274 274 m, n = m[0], n[-1]
275 275
276 276 if m < n:
277 277 r = spanset(repo, m, n + 1)
278 278 else:
279 279 r = spanset(repo, m, n - 1)
280 280 return r & subset
281 281
282 282 def dagrange(repo, subset, x, y):
283 283 r = spanset(repo)
284 284 xs = _revsbetween(repo, getset(repo, r, x), getset(repo, r, y))
285 285 s = subset.set()
286 286 return xs.filter(s.__contains__)
287 287
288 288 def andset(repo, subset, x, y):
289 289 return getset(repo, getset(repo, subset, x), y)
290 290
291 291 def orset(repo, subset, x, y):
292 292 xl = getset(repo, subset, x)
293 293 yl = getset(repo, subset - xl, y)
294 294 return xl + yl
295 295
296 296 def notset(repo, subset, x):
297 297 return subset - getset(repo, subset, x)
298 298
299 299 def listset(repo, subset, a, b):
300 300 raise error.ParseError(_("can't use a list in this context"))
301 301
302 302 def func(repo, subset, a, b):
303 303 if a[0] == 'symbol' and a[1] in symbols:
304 304 return symbols[a[1]](repo, subset, b)
305 305 raise error.ParseError(_("not a function: %s") % a[1])
306 306
307 307 # functions
308 308
309 309 def adds(repo, subset, x):
310 310 """``adds(pattern)``
311 311 Changesets that add a file matching pattern.
312 312
313 313 The pattern without explicit kind like ``glob:`` is expected to be
314 314 relative to the current directory and match against a file or a
315 315 directory.
316 316 """
317 317 # i18n: "adds" is a keyword
318 318 pat = getstring(x, _("adds requires a pattern"))
319 319 return checkstatus(repo, subset, pat, 1)
320 320
321 321 def ancestor(repo, subset, x):
322 322 """``ancestor(*changeset)``
323 323 A greatest common ancestor of the changesets.
324 324
325 325 Accepts 0 or more changesets.
326 326 Will return empty list when passed no args.
327 327 Greatest common ancestor of a single changeset is that changeset.
328 328 """
329 329 # i18n: "ancestor" is a keyword
330 330 l = getlist(x)
331 331 rl = spanset(repo)
332 332 anc = None
333 333
334 334 # (getset(repo, rl, i) for i in l) generates a list of lists
335 335 for revs in (getset(repo, rl, i) for i in l):
336 336 for r in revs:
337 337 if anc is None:
338 338 anc = repo[r]
339 339 else:
340 340 anc = anc.ancestor(repo[r])
341 341
342 342 if anc is not None and anc.rev() in subset:
343 343 return baseset([anc.rev()])
344 344 return baseset([])
345 345
346 346 def _ancestors(repo, subset, x, followfirst=False):
347 347 args = getset(repo, spanset(repo), x)
348 348 if not args:
349 349 return baseset([])
350 350 s = _revancestors(repo, args, followfirst)
351 351 return subset.filter(s.__contains__)
352 352
353 353 def ancestors(repo, subset, x):
354 354 """``ancestors(set)``
355 355 Changesets that are ancestors of a changeset in set.
356 356 """
357 357 return _ancestors(repo, subset, x)
358 358
359 359 def _firstancestors(repo, subset, x):
360 360 # ``_firstancestors(set)``
361 361 # Like ``ancestors(set)`` but follows only the first parents.
362 362 return _ancestors(repo, subset, x, followfirst=True)
363 363
364 364 def ancestorspec(repo, subset, x, n):
365 365 """``set~n``
366 366 Changesets that are the Nth ancestor (first parents only) of a changeset
367 367 in set.
368 368 """
369 369 try:
370 370 n = int(n[1])
371 371 except (TypeError, ValueError):
372 372 raise error.ParseError(_("~ expects a number"))
373 373 ps = set()
374 374 cl = repo.changelog
375 375 for r in getset(repo, baseset(cl), x):
376 376 for i in range(n):
377 377 r = cl.parentrevs(r)[0]
378 378 ps.add(r)
379 379 return subset.filter(ps.__contains__)
380 380
381 381 def author(repo, subset, x):
382 382 """``author(string)``
383 383 Alias for ``user(string)``.
384 384 """
385 385 # i18n: "author" is a keyword
386 386 n = encoding.lower(getstring(x, _("author requires a string")))
387 387 kind, pattern, matcher = _substringmatcher(n)
388 388 return subset.filter(lambda x: matcher(encoding.lower(repo[x].user())))
389 389
390 390 def only(repo, subset, x):
391 391 """``only(set, [set])``
392 392 Changesets that are ancestors of the first set that are not ancestors
393 393 of any other head in the repo. If a second set is specified, the result
394 394 is ancestors of the first set that are not ancestors of the second set
395 395 (i.e. ::<set1> - ::<set2>).
396 396 """
397 397 cl = repo.changelog
398 398 # i18n: "only" is a keyword
399 399 args = getargs(x, 1, 2, _('only takes one or two arguments'))
400 400 include = getset(repo, spanset(repo), args[0]).set()
401 401 if len(args) == 1:
402 402 if len(include) == 0:
403 403 return baseset([])
404 404
405 405 descendants = set(_revdescendants(repo, include, False))
406 406 exclude = [rev for rev in cl.headrevs()
407 407 if not rev in descendants and not rev in include]
408 408 else:
409 409 exclude = getset(repo, spanset(repo), args[1])
410 410
411 411 results = set(ancestormod.missingancestors(include, exclude, cl.parentrevs))
412 412 return lazyset(subset, results.__contains__)
413 413
414 414 def bisect(repo, subset, x):
415 415 """``bisect(string)``
416 416 Changesets marked in the specified bisect status:
417 417
418 418 - ``good``, ``bad``, ``skip``: csets explicitly marked as good/bad/skip
419 419 - ``goods``, ``bads`` : csets topologically good/bad
420 420 - ``range`` : csets taking part in the bisection
421 421 - ``pruned`` : csets that are goods, bads or skipped
422 422 - ``untested`` : csets whose fate is yet unknown
423 423 - ``ignored`` : csets ignored due to DAG topology
424 424 - ``current`` : the cset currently being bisected
425 425 """
426 426 # i18n: "bisect" is a keyword
427 427 status = getstring(x, _("bisect requires a string")).lower()
428 428 state = set(hbisect.get(repo, status))
429 429 return subset.filter(state.__contains__)
430 430
431 431 # Backward-compatibility
432 432 # - no help entry so that we do not advertise it any more
433 433 def bisected(repo, subset, x):
434 434 return bisect(repo, subset, x)
435 435
436 436 def bookmark(repo, subset, x):
437 437 """``bookmark([name])``
438 438 The named bookmark or all bookmarks.
439 439
440 440 If `name` starts with `re:`, the remainder of the name is treated as
441 441 a regular expression. To match a bookmark that actually starts with `re:`,
442 442 use the prefix `literal:`.
443 443 """
444 444 # i18n: "bookmark" is a keyword
445 445 args = getargs(x, 0, 1, _('bookmark takes one or no arguments'))
446 446 if args:
447 447 bm = getstring(args[0],
448 448 # i18n: "bookmark" is a keyword
449 449 _('the argument to bookmark must be a string'))
450 450 kind, pattern, matcher = _stringmatcher(bm)
451 451 bms = set()
452 452 if kind == 'literal':
453 453 bmrev = repo._bookmarks.get(pattern, None)
454 454 if not bmrev:
455 455 raise util.Abort(_("bookmark '%s' does not exist") % bm)
456 456 bms.add(repo[bmrev].rev())
457 457 else:
458 458 matchrevs = set()
459 459 for name, bmrev in repo._bookmarks.iteritems():
460 460 if matcher(name):
461 461 matchrevs.add(bmrev)
462 462 if not matchrevs:
463 463 raise util.Abort(_("no bookmarks exist that match '%s'")
464 464 % pattern)
465 465 for bmrev in matchrevs:
466 466 bms.add(repo[bmrev].rev())
467 467 else:
468 468 bms = set([repo[r].rev()
469 469 for r in repo._bookmarks.values()])
470 bms -= set([node.nullrev])
470 471 return subset.filter(bms.__contains__)
471 472
472 473 def branch(repo, subset, x):
473 474 """``branch(string or set)``
474 475 All changesets belonging to the given branch or the branches of the given
475 476 changesets.
476 477
477 478 If `string` starts with `re:`, the remainder of the name is treated as
478 479 a regular expression. To match a branch that actually starts with `re:`,
479 480 use the prefix `literal:`.
480 481 """
481 482 try:
482 483 b = getstring(x, '')
483 484 except error.ParseError:
484 485 # not a string, but another revspec, e.g. tip()
485 486 pass
486 487 else:
487 488 kind, pattern, matcher = _stringmatcher(b)
488 489 if kind == 'literal':
489 490 # note: falls through to the revspec case if no branch with
490 491 # this name exists
491 492 if pattern in repo.branchmap():
492 493 return subset.filter(lambda r: matcher(repo[r].branch()))
493 494 else:
494 495 return subset.filter(lambda r: matcher(repo[r].branch()))
495 496
496 497 s = getset(repo, spanset(repo), x)
497 498 b = set()
498 499 for r in s:
499 500 b.add(repo[r].branch())
500 501 s = s.set()
501 502 return subset.filter(lambda r: r in s or repo[r].branch() in b)
502 503
503 504 def bumped(repo, subset, x):
504 505 """``bumped()``
505 506 Mutable changesets marked as successors of public changesets.
506 507
507 508 Only non-public and non-obsolete changesets can be `bumped`.
508 509 """
509 510 # i18n: "bumped" is a keyword
510 511 getargs(x, 0, 0, _("bumped takes no arguments"))
511 512 bumped = obsmod.getrevs(repo, 'bumped')
512 513 return subset & bumped
513 514
514 515 def bundle(repo, subset, x):
515 516 """``bundle()``
516 517 Changesets in the bundle.
517 518
518 519 Bundle must be specified by the -R option."""
519 520
520 521 try:
521 522 bundlerevs = repo.changelog.bundlerevs
522 523 except AttributeError:
523 524 raise util.Abort(_("no bundle provided - specify with -R"))
524 525 return subset & bundlerevs
525 526
526 527 def checkstatus(repo, subset, pat, field):
527 528 hasset = matchmod.patkind(pat) == 'set'
528 529
529 530 def matches(x):
530 531 m = None
531 532 fname = None
532 533 c = repo[x]
533 534 if not m or hasset:
534 535 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
535 536 if not m.anypats() and len(m.files()) == 1:
536 537 fname = m.files()[0]
537 538 if fname is not None:
538 539 if fname not in c.files():
539 540 return False
540 541 else:
541 542 for f in c.files():
542 543 if m(f):
543 544 break
544 545 else:
545 546 return False
546 547 files = repo.status(c.p1().node(), c.node())[field]
547 548 if fname is not None:
548 549 if fname in files:
549 550 return True
550 551 else:
551 552 for f in files:
552 553 if m(f):
553 554 return True
554 555
555 556 return subset.filter(matches)
556 557
557 558 def _children(repo, narrow, parentset):
558 559 cs = set()
559 560 if not parentset:
560 561 return baseset(cs)
561 562 pr = repo.changelog.parentrevs
562 563 minrev = min(parentset)
563 564 for r in narrow:
564 565 if r <= minrev:
565 566 continue
566 567 for p in pr(r):
567 568 if p in parentset:
568 569 cs.add(r)
569 570 return baseset(cs)
570 571
571 572 def children(repo, subset, x):
572 573 """``children(set)``
573 574 Child changesets of changesets in set.
574 575 """
575 576 s = getset(repo, baseset(repo), x).set()
576 577 cs = _children(repo, subset, s)
577 578 return subset & cs
578 579
579 580 def closed(repo, subset, x):
580 581 """``closed()``
581 582 Changeset is closed.
582 583 """
583 584 # i18n: "closed" is a keyword
584 585 getargs(x, 0, 0, _("closed takes no arguments"))
585 586 return subset.filter(lambda r: repo[r].closesbranch())
586 587
587 588 def contains(repo, subset, x):
588 589 """``contains(pattern)``
589 590 The revision's manifest contains a file matching pattern (but might not
590 591 modify it). See :hg:`help patterns` for information about file patterns.
591 592
592 593 The pattern without explicit kind like ``glob:`` is expected to be
593 594 relative to the current directory and match against a file exactly
594 595 for efficiency.
595 596 """
596 597 # i18n: "contains" is a keyword
597 598 pat = getstring(x, _("contains requires a pattern"))
598 599
599 600 def matches(x):
600 601 if not matchmod.patkind(pat):
601 602 pats = pathutil.canonpath(repo.root, repo.getcwd(), pat)
602 603 if pats in repo[x]:
603 604 return True
604 605 else:
605 606 c = repo[x]
606 607 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
607 608 for f in c.manifest():
608 609 if m(f):
609 610 return True
610 611 return False
611 612
612 613 return subset.filter(matches)
613 614
614 615 def converted(repo, subset, x):
615 616 """``converted([id])``
616 617 Changesets converted from the given identifier in the old repository if
617 618 present, or all converted changesets if no identifier is specified.
618 619 """
619 620
620 621 # There is exactly no chance of resolving the revision, so do a simple
621 622 # string compare and hope for the best
622 623
623 624 rev = None
624 625 # i18n: "converted" is a keyword
625 626 l = getargs(x, 0, 1, _('converted takes one or no arguments'))
626 627 if l:
627 628 # i18n: "converted" is a keyword
628 629 rev = getstring(l[0], _('converted requires a revision'))
629 630
630 631 def _matchvalue(r):
631 632 source = repo[r].extra().get('convert_revision', None)
632 633 return source is not None and (rev is None or source.startswith(rev))
633 634
634 635 return subset.filter(lambda r: _matchvalue(r))
635 636
636 637 def date(repo, subset, x):
637 638 """``date(interval)``
638 639 Changesets within the interval, see :hg:`help dates`.
639 640 """
640 641 # i18n: "date" is a keyword
641 642 ds = getstring(x, _("date requires a string"))
642 643 dm = util.matchdate(ds)
643 644 return subset.filter(lambda x: dm(repo[x].date()[0]))
644 645
645 646 def desc(repo, subset, x):
646 647 """``desc(string)``
647 648 Search commit message for string. The match is case-insensitive.
648 649 """
649 650 # i18n: "desc" is a keyword
650 651 ds = encoding.lower(getstring(x, _("desc requires a string")))
651 652
652 653 def matches(x):
653 654 c = repo[x]
654 655 return ds in encoding.lower(c.description())
655 656
656 657 return subset.filter(matches)
657 658
658 659 def _descendants(repo, subset, x, followfirst=False):
659 660 args = getset(repo, spanset(repo), x)
660 661 if not args:
661 662 return baseset([])
662 663 s = _revdescendants(repo, args, followfirst)
663 664
664 665 # Both sets need to be ascending in order to lazily return the union
665 666 # in the correct order.
666 667 args.ascending()
667 668 result = (orderedlazyset(s, subset.__contains__, ascending=True) +
668 669 orderedlazyset(args, subset.__contains__, ascending=True))
669 670
670 671 # Wrap result in a lazyset since it's an _addset, which doesn't implement
671 672 # all the necessary functions to be consumed by callers.
672 673 return orderedlazyset(result, lambda r: True, ascending=True)
673 674
674 675 def descendants(repo, subset, x):
675 676 """``descendants(set)``
676 677 Changesets which are descendants of changesets in set.
677 678 """
678 679 return _descendants(repo, subset, x)
679 680
680 681 def _firstdescendants(repo, subset, x):
681 682 # ``_firstdescendants(set)``
682 683 # Like ``descendants(set)`` but follows only the first parents.
683 684 return _descendants(repo, subset, x, followfirst=True)
684 685
685 686 def destination(repo, subset, x):
686 687 """``destination([set])``
687 688 Changesets that were created by a graft, transplant or rebase operation,
688 689 with the given revisions specified as the source. Omitting the optional set
689 690 is the same as passing all().
690 691 """
691 692 if x is not None:
692 693 args = getset(repo, spanset(repo), x).set()
693 694 else:
694 695 args = getall(repo, spanset(repo), x).set()
695 696
696 697 dests = set()
697 698
698 699 # subset contains all of the possible destinations that can be returned, so
699 700 # iterate over them and see if their source(s) were provided in the args.
700 701 # Even if the immediate src of r is not in the args, src's source (or
701 702 # further back) may be. Scanning back further than the immediate src allows
702 703 # transitive transplants and rebases to yield the same results as transitive
703 704 # grafts.
704 705 for r in subset:
705 706 src = _getrevsource(repo, r)
706 707 lineage = None
707 708
708 709 while src is not None:
709 710 if lineage is None:
710 711 lineage = list()
711 712
712 713 lineage.append(r)
713 714
714 715 # The visited lineage is a match if the current source is in the arg
715 716 # set. Since every candidate dest is visited by way of iterating
716 717 # subset, any dests further back in the lineage will be tested by a
717 718 # different iteration over subset. Likewise, if the src was already
718 719 # selected, the current lineage can be selected without going back
719 720 # further.
720 721 if src in args or src in dests:
721 722 dests.update(lineage)
722 723 break
723 724
724 725 r = src
725 726 src = _getrevsource(repo, r)
726 727
727 728 return subset.filter(dests.__contains__)
728 729
729 730 def divergent(repo, subset, x):
730 731 """``divergent()``
731 732 Final successors of changesets with an alternative set of final successors.
732 733 """
733 734 # i18n: "divergent" is a keyword
734 735 getargs(x, 0, 0, _("divergent takes no arguments"))
735 736 divergent = obsmod.getrevs(repo, 'divergent')
736 737 return subset.filter(divergent.__contains__)
737 738
738 739 def draft(repo, subset, x):
739 740 """``draft()``
740 741 Changeset in draft phase."""
741 742 # i18n: "draft" is a keyword
742 743 getargs(x, 0, 0, _("draft takes no arguments"))
743 744 pc = repo._phasecache
744 745 return subset.filter(lambda r: pc.phase(repo, r) == phases.draft)
745 746
746 747 def extinct(repo, subset, x):
747 748 """``extinct()``
748 749 Obsolete changesets with obsolete descendants only.
749 750 """
750 751 # i18n: "extinct" is a keyword
751 752 getargs(x, 0, 0, _("extinct takes no arguments"))
752 753 extincts = obsmod.getrevs(repo, 'extinct')
753 754 return subset & extincts
754 755
755 756 def extra(repo, subset, x):
756 757 """``extra(label, [value])``
757 758 Changesets with the given label in the extra metadata, with the given
758 759 optional value.
759 760
760 761 If `value` starts with `re:`, the remainder of the value is treated as
761 762 a regular expression. To match a value that actually starts with `re:`,
762 763 use the prefix `literal:`.
763 764 """
764 765
765 766 # i18n: "extra" is a keyword
766 767 l = getargs(x, 1, 2, _('extra takes at least 1 and at most 2 arguments'))
767 768 # i18n: "extra" is a keyword
768 769 label = getstring(l[0], _('first argument to extra must be a string'))
769 770 value = None
770 771
771 772 if len(l) > 1:
772 773 # i18n: "extra" is a keyword
773 774 value = getstring(l[1], _('second argument to extra must be a string'))
774 775 kind, value, matcher = _stringmatcher(value)
775 776
776 777 def _matchvalue(r):
777 778 extra = repo[r].extra()
778 779 return label in extra and (value is None or matcher(extra[label]))
779 780
780 781 return subset.filter(lambda r: _matchvalue(r))
781 782
782 783 def filelog(repo, subset, x):
783 784 """``filelog(pattern)``
784 785 Changesets connected to the specified filelog.
785 786
786 787 For performance reasons, visits only revisions mentioned in the file-level
787 788 filelog, rather than filtering through all changesets (much faster, but
788 789 doesn't include deletes or duplicate changes). For a slower, more accurate
789 790 result, use ``file()``.
790 791
791 792 The pattern without explicit kind like ``glob:`` is expected to be
792 793 relative to the current directory and match against a file exactly
793 794 for efficiency.
794 795 """
795 796
796 797 # i18n: "filelog" is a keyword
797 798 pat = getstring(x, _("filelog requires a pattern"))
798 799 s = set()
799 800
800 801 if not matchmod.patkind(pat):
801 802 f = pathutil.canonpath(repo.root, repo.getcwd(), pat)
802 803 fl = repo.file(f)
803 804 for fr in fl:
804 805 s.add(fl.linkrev(fr))
805 806 else:
806 807 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=repo[None])
807 808 for f in repo[None]:
808 809 if m(f):
809 810 fl = repo.file(f)
810 811 for fr in fl:
811 812 s.add(fl.linkrev(fr))
812 813
813 814 return subset.filter(s.__contains__)
814 815
815 816 def first(repo, subset, x):
816 817 """``first(set, [n])``
817 818 An alias for limit().
818 819 """
819 820 return limit(repo, subset, x)
820 821
821 822 def _follow(repo, subset, x, name, followfirst=False):
822 823 l = getargs(x, 0, 1, _("%s takes no arguments or a filename") % name)
823 824 c = repo['.']
824 825 if l:
825 826 x = getstring(l[0], _("%s expected a filename") % name)
826 827 if x in c:
827 828 cx = c[x]
828 829 s = set(ctx.rev() for ctx in cx.ancestors(followfirst=followfirst))
829 830 # include the revision responsible for the most recent version
830 831 s.add(cx.linkrev())
831 832 else:
832 833 return baseset([])
833 834 else:
834 835 s = _revancestors(repo, baseset([c.rev()]), followfirst)
835 836
836 837 return subset.filter(s.__contains__)
837 838
838 839 def follow(repo, subset, x):
839 840 """``follow([file])``
840 841 An alias for ``::.`` (ancestors of the working copy's first parent).
841 842 If a filename is specified, the history of the given file is followed,
842 843 including copies.
843 844 """
844 845 return _follow(repo, subset, x, 'follow')
845 846
846 847 def _followfirst(repo, subset, x):
847 848 # ``followfirst([file])``
848 849 # Like ``follow([file])`` but follows only the first parent of
849 850 # every revision or file revision.
850 851 return _follow(repo, subset, x, '_followfirst', followfirst=True)
851 852
852 853 def getall(repo, subset, x):
853 854 """``all()``
854 855 All changesets, the same as ``0:tip``.
855 856 """
856 857 # i18n: "all" is a keyword
857 858 getargs(x, 0, 0, _("all takes no arguments"))
858 859 return subset
859 860
860 861 def grep(repo, subset, x):
861 862 """``grep(regex)``
862 863 Like ``keyword(string)`` but accepts a regex. Use ``grep(r'...')``
863 864 to ensure special escape characters are handled correctly. Unlike
864 865 ``keyword(string)``, the match is case-sensitive.
865 866 """
866 867 try:
867 868 # i18n: "grep" is a keyword
868 869 gr = re.compile(getstring(x, _("grep requires a string")))
869 870 except re.error, e:
870 871 raise error.ParseError(_('invalid match pattern: %s') % e)
871 872
872 873 def matches(x):
873 874 c = repo[x]
874 875 for e in c.files() + [c.user(), c.description()]:
875 876 if gr.search(e):
876 877 return True
877 878 return False
878 879
879 880 return subset.filter(matches)
880 881
881 882 def _matchfiles(repo, subset, x):
882 883 # _matchfiles takes a revset list of prefixed arguments:
883 884 #
884 885 # [p:foo, i:bar, x:baz]
885 886 #
886 887 # builds a match object from them and filters subset. Allowed
887 888 # prefixes are 'p:' for regular patterns, 'i:' for include
888 889 # patterns and 'x:' for exclude patterns. Use 'r:' prefix to pass
889 890 # a revision identifier, or the empty string to reference the
890 891 # working directory, from which the match object is
891 892 # initialized. Use 'd:' to set the default matching mode, default
892 893 # to 'glob'. At most one 'r:' and 'd:' argument can be passed.
893 894
894 895 # i18n: "_matchfiles" is a keyword
895 896 l = getargs(x, 1, -1, _("_matchfiles requires at least one argument"))
896 897 pats, inc, exc = [], [], []
897 898 hasset = False
898 899 rev, default = None, None
899 900 for arg in l:
900 901 # i18n: "_matchfiles" is a keyword
901 902 s = getstring(arg, _("_matchfiles requires string arguments"))
902 903 prefix, value = s[:2], s[2:]
903 904 if prefix == 'p:':
904 905 pats.append(value)
905 906 elif prefix == 'i:':
906 907 inc.append(value)
907 908 elif prefix == 'x:':
908 909 exc.append(value)
909 910 elif prefix == 'r:':
910 911 if rev is not None:
911 912 # i18n: "_matchfiles" is a keyword
912 913 raise error.ParseError(_('_matchfiles expected at most one '
913 914 'revision'))
914 915 rev = value
915 916 elif prefix == 'd:':
916 917 if default is not None:
917 918 # i18n: "_matchfiles" is a keyword
918 919 raise error.ParseError(_('_matchfiles expected at most one '
919 920 'default mode'))
920 921 default = value
921 922 else:
922 923 # i18n: "_matchfiles" is a keyword
923 924 raise error.ParseError(_('invalid _matchfiles prefix: %s') % prefix)
924 925 if not hasset and matchmod.patkind(value) == 'set':
925 926 hasset = True
926 927 if not default:
927 928 default = 'glob'
928 929
929 930 def matches(x):
930 931 m = None
931 932 c = repo[x]
932 933 if not m or (hasset and rev is None):
933 934 ctx = c
934 935 if rev is not None:
935 936 ctx = repo[rev or None]
936 937 m = matchmod.match(repo.root, repo.getcwd(), pats, include=inc,
937 938 exclude=exc, ctx=ctx, default=default)
938 939 for f in c.files():
939 940 if m(f):
940 941 return True
941 942 return False
942 943
943 944 return subset.filter(matches)
944 945
945 946 def hasfile(repo, subset, x):
946 947 """``file(pattern)``
947 948 Changesets affecting files matched by pattern.
948 949
949 950 For a faster but less accurate result, consider using ``filelog()``
950 951 instead.
951 952
952 953 This predicate uses ``glob:`` as the default kind of pattern.
953 954 """
954 955 # i18n: "file" is a keyword
955 956 pat = getstring(x, _("file requires a pattern"))
956 957 return _matchfiles(repo, subset, ('string', 'p:' + pat))
957 958
958 959 def head(repo, subset, x):
959 960 """``head()``
960 961 Changeset is a named branch head.
961 962 """
962 963 # i18n: "head" is a keyword
963 964 getargs(x, 0, 0, _("head takes no arguments"))
964 965 hs = set()
965 966 for b, ls in repo.branchmap().iteritems():
966 967 hs.update(repo[h].rev() for h in ls)
967 968 return baseset(hs).filter(subset.__contains__)
968 969
969 970 def heads(repo, subset, x):
970 971 """``heads(set)``
971 972 Members of set with no children in set.
972 973 """
973 974 s = getset(repo, subset, x)
974 975 ps = parents(repo, subset, x)
975 976 return s - ps
976 977
977 978 def hidden(repo, subset, x):
978 979 """``hidden()``
979 980 Hidden changesets.
980 981 """
981 982 # i18n: "hidden" is a keyword
982 983 getargs(x, 0, 0, _("hidden takes no arguments"))
983 984 hiddenrevs = repoview.filterrevs(repo, 'visible')
984 985 return subset & hiddenrevs
985 986
986 987 def keyword(repo, subset, x):
987 988 """``keyword(string)``
988 989 Search commit message, user name, and names of changed files for
989 990 string. The match is case-insensitive.
990 991 """
991 992 # i18n: "keyword" is a keyword
992 993 kw = encoding.lower(getstring(x, _("keyword requires a string")))
993 994
994 995 def matches(r):
995 996 c = repo[r]
996 997 return util.any(kw in encoding.lower(t) for t in c.files() + [c.user(),
997 998 c.description()])
998 999
999 1000 return subset.filter(matches)
1000 1001
1001 1002 def limit(repo, subset, x):
1002 1003 """``limit(set, [n])``
1003 1004 First n members of set, defaulting to 1.
1004 1005 """
1005 1006 # i18n: "limit" is a keyword
1006 1007 l = getargs(x, 1, 2, _("limit requires one or two arguments"))
1007 1008 try:
1008 1009 lim = 1
1009 1010 if len(l) == 2:
1010 1011 # i18n: "limit" is a keyword
1011 1012 lim = int(getstring(l[1], _("limit requires a number")))
1012 1013 except (TypeError, ValueError):
1013 1014 # i18n: "limit" is a keyword
1014 1015 raise error.ParseError(_("limit expects a number"))
1015 1016 ss = subset.set()
1016 1017 os = getset(repo, spanset(repo), l[0])
1017 1018 bs = baseset([])
1018 1019 it = iter(os)
1019 1020 for x in xrange(lim):
1020 1021 try:
1021 1022 y = it.next()
1022 1023 if y in ss:
1023 1024 bs.append(y)
1024 1025 except (StopIteration):
1025 1026 break
1026 1027 return bs
1027 1028
1028 1029 def last(repo, subset, x):
1029 1030 """``last(set, [n])``
1030 1031 Last n members of set, defaulting to 1.
1031 1032 """
1032 1033 # i18n: "last" is a keyword
1033 1034 l = getargs(x, 1, 2, _("last requires one or two arguments"))
1034 1035 try:
1035 1036 lim = 1
1036 1037 if len(l) == 2:
1037 1038 # i18n: "last" is a keyword
1038 1039 lim = int(getstring(l[1], _("last requires a number")))
1039 1040 except (TypeError, ValueError):
1040 1041 # i18n: "last" is a keyword
1041 1042 raise error.ParseError(_("last expects a number"))
1042 1043 ss = subset.set()
1043 1044 os = getset(repo, spanset(repo), l[0])
1044 1045 os.reverse()
1045 1046 bs = baseset([])
1046 1047 it = iter(os)
1047 1048 for x in xrange(lim):
1048 1049 try:
1049 1050 y = it.next()
1050 1051 if y in ss:
1051 1052 bs.append(y)
1052 1053 except (StopIteration):
1053 1054 break
1054 1055 return bs
1055 1056
1056 1057 def maxrev(repo, subset, x):
1057 1058 """``max(set)``
1058 1059 Changeset with highest revision number in set.
1059 1060 """
1060 1061 os = getset(repo, spanset(repo), x)
1061 1062 if os:
1062 1063 m = os.max()
1063 1064 if m in subset:
1064 1065 return baseset([m])
1065 1066 return baseset([])
1066 1067
1067 1068 def merge(repo, subset, x):
1068 1069 """``merge()``
1069 1070 Changeset is a merge changeset.
1070 1071 """
1071 1072 # i18n: "merge" is a keyword
1072 1073 getargs(x, 0, 0, _("merge takes no arguments"))
1073 1074 cl = repo.changelog
1074 1075 return subset.filter(lambda r: cl.parentrevs(r)[1] != -1)
1075 1076
1076 1077 def branchpoint(repo, subset, x):
1077 1078 """``branchpoint()``
1078 1079 Changesets with more than one child.
1079 1080 """
1080 1081 # i18n: "branchpoint" is a keyword
1081 1082 getargs(x, 0, 0, _("branchpoint takes no arguments"))
1082 1083 cl = repo.changelog
1083 1084 if not subset:
1084 1085 return baseset([])
1085 1086 baserev = min(subset)
1086 1087 parentscount = [0]*(len(repo) - baserev)
1087 1088 for r in cl.revs(start=baserev + 1):
1088 1089 for p in cl.parentrevs(r):
1089 1090 if p >= baserev:
1090 1091 parentscount[p - baserev] += 1
1091 1092 return subset.filter(lambda r: parentscount[r - baserev] > 1)
1092 1093
1093 1094 def minrev(repo, subset, x):
1094 1095 """``min(set)``
1095 1096 Changeset with lowest revision number in set.
1096 1097 """
1097 1098 os = getset(repo, spanset(repo), x)
1098 1099 if os:
1099 1100 m = os.min()
1100 1101 if m in subset:
1101 1102 return baseset([m])
1102 1103 return baseset([])
1103 1104
1104 1105 def modifies(repo, subset, x):
1105 1106 """``modifies(pattern)``
1106 1107 Changesets modifying files matched by pattern.
1107 1108
1108 1109 The pattern without explicit kind like ``glob:`` is expected to be
1109 1110 relative to the current directory and match against a file or a
1110 1111 directory.
1111 1112 """
1112 1113 # i18n: "modifies" is a keyword
1113 1114 pat = getstring(x, _("modifies requires a pattern"))
1114 1115 return checkstatus(repo, subset, pat, 0)
1115 1116
1116 1117 def node_(repo, subset, x):
1117 1118 """``id(string)``
1118 1119 Revision non-ambiguously specified by the given hex string prefix.
1119 1120 """
1120 1121 # i18n: "id" is a keyword
1121 1122 l = getargs(x, 1, 1, _("id requires one argument"))
1122 1123 # i18n: "id" is a keyword
1123 1124 n = getstring(l[0], _("id requires a string"))
1124 1125 if len(n) == 40:
1125 1126 rn = repo[n].rev()
1126 1127 else:
1127 1128 rn = None
1128 1129 pm = repo.changelog._partialmatch(n)
1129 1130 if pm is not None:
1130 1131 rn = repo.changelog.rev(pm)
1131 1132
1132 1133 return subset.filter(lambda r: r == rn)
1133 1134
1134 1135 def obsolete(repo, subset, x):
1135 1136 """``obsolete()``
1136 1137 Mutable changeset with a newer version."""
1137 1138 # i18n: "obsolete" is a keyword
1138 1139 getargs(x, 0, 0, _("obsolete takes no arguments"))
1139 1140 obsoletes = obsmod.getrevs(repo, 'obsolete')
1140 1141 return subset & obsoletes
1141 1142
1142 1143 def origin(repo, subset, x):
1143 1144 """``origin([set])``
1144 1145 Changesets that were specified as a source for the grafts, transplants or
1145 1146 rebases that created the given revisions. Omitting the optional set is the
1146 1147 same as passing all(). If a changeset created by these operations is itself
1147 1148 specified as a source for one of these operations, only the source changeset
1148 1149 for the first operation is selected.
1149 1150 """
1150 1151 if x is not None:
1151 1152 args = getset(repo, spanset(repo), x).set()
1152 1153 else:
1153 1154 args = getall(repo, spanset(repo), x).set()
1154 1155
1155 1156 def _firstsrc(rev):
1156 1157 src = _getrevsource(repo, rev)
1157 1158 if src is None:
1158 1159 return None
1159 1160
1160 1161 while True:
1161 1162 prev = _getrevsource(repo, src)
1162 1163
1163 1164 if prev is None:
1164 1165 return src
1165 1166 src = prev
1166 1167
1167 1168 o = set([_firstsrc(r) for r in args])
1168 1169 o -= set([None])
1169 1170 return subset.filter(o.__contains__)
1170 1171
1171 1172 def outgoing(repo, subset, x):
1172 1173 """``outgoing([path])``
1173 1174 Changesets not found in the specified destination repository, or the
1174 1175 default push location.
1175 1176 """
1176 1177 import hg # avoid start-up nasties
1177 1178 # i18n: "outgoing" is a keyword
1178 1179 l = getargs(x, 0, 1, _("outgoing takes one or no arguments"))
1179 1180 # i18n: "outgoing" is a keyword
1180 1181 dest = l and getstring(l[0], _("outgoing requires a repository path")) or ''
1181 1182 dest = repo.ui.expandpath(dest or 'default-push', dest or 'default')
1182 1183 dest, branches = hg.parseurl(dest)
1183 1184 revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
1184 1185 if revs:
1185 1186 revs = [repo.lookup(rev) for rev in revs]
1186 1187 other = hg.peer(repo, {}, dest)
1187 1188 repo.ui.pushbuffer()
1188 1189 outgoing = discovery.findcommonoutgoing(repo, other, onlyheads=revs)
1189 1190 repo.ui.popbuffer()
1190 1191 cl = repo.changelog
1191 1192 o = set([cl.rev(r) for r in outgoing.missing])
1192 1193 return subset.filter(o.__contains__)
1193 1194
1194 1195 def p1(repo, subset, x):
1195 1196 """``p1([set])``
1196 1197 First parent of changesets in set, or the working directory.
1197 1198 """
1198 1199 if x is None:
1199 1200 p = repo[x].p1().rev()
1200 1201 return subset.filter(lambda r: r == p)
1201 1202
1202 1203 ps = set()
1203 1204 cl = repo.changelog
1204 1205 for r in getset(repo, spanset(repo), x):
1205 1206 ps.add(cl.parentrevs(r)[0])
1206 1207 ps -= set([node.nullrev])
1207 1208 return subset & ps
1208 1209
1209 1210 def p2(repo, subset, x):
1210 1211 """``p2([set])``
1211 1212 Second parent of changesets in set, or the working directory.
1212 1213 """
1213 1214 if x is None:
1214 1215 ps = repo[x].parents()
1215 1216 try:
1216 1217 p = ps[1].rev()
1217 1218 return subset.filter(lambda r: r == p)
1218 1219 except IndexError:
1219 1220 return baseset([])
1220 1221
1221 1222 ps = set()
1222 1223 cl = repo.changelog
1223 1224 for r in getset(repo, spanset(repo), x):
1224 1225 ps.add(cl.parentrevs(r)[1])
1225 1226 ps -= set([node.nullrev])
1226 1227 return subset & ps
1227 1228
1228 1229 def parents(repo, subset, x):
1229 1230 """``parents([set])``
1230 1231 The set of all parents for all changesets in set, or the working directory.
1231 1232 """
1232 1233 if x is None:
1233 1234 ps = set(p.rev() for p in repo[x].parents())
1234 1235 else:
1235 1236 ps = set()
1236 1237 cl = repo.changelog
1237 1238 for r in getset(repo, spanset(repo), x):
1238 1239 ps.update(cl.parentrevs(r))
1239 1240 ps -= set([node.nullrev])
1240 1241 return baseset(ps) & subset
1241 1242
1242 1243 def parentspec(repo, subset, x, n):
1243 1244 """``set^0``
1244 1245 The set.
1245 1246 ``set^1`` (or ``set^``), ``set^2``
1246 1247 First or second parent, respectively, of all changesets in set.
1247 1248 """
1248 1249 try:
1249 1250 n = int(n[1])
1250 1251 if n not in (0, 1, 2):
1251 1252 raise ValueError
1252 1253 except (TypeError, ValueError):
1253 1254 raise error.ParseError(_("^ expects a number 0, 1, or 2"))
1254 1255 ps = set()
1255 1256 cl = repo.changelog
1256 1257 for r in getset(repo, baseset(cl), x):
1257 1258 if n == 0:
1258 1259 ps.add(r)
1259 1260 elif n == 1:
1260 1261 ps.add(cl.parentrevs(r)[0])
1261 1262 elif n == 2:
1262 1263 parents = cl.parentrevs(r)
1263 1264 if len(parents) > 1:
1264 1265 ps.add(parents[1])
1265 1266 return subset & ps
1266 1267
1267 1268 def present(repo, subset, x):
1268 1269 """``present(set)``
1269 1270 An empty set, if any revision in set isn't found; otherwise,
1270 1271 all revisions in set.
1271 1272
1272 1273 If any of specified revisions is not present in the local repository,
1273 1274 the query is normally aborted. But this predicate allows the query
1274 1275 to continue even in such cases.
1275 1276 """
1276 1277 try:
1277 1278 return getset(repo, subset, x)
1278 1279 except error.RepoLookupError:
1279 1280 return baseset([])
1280 1281
1281 1282 def public(repo, subset, x):
1282 1283 """``public()``
1283 1284 Changeset in public phase."""
1284 1285 # i18n: "public" is a keyword
1285 1286 getargs(x, 0, 0, _("public takes no arguments"))
1286 1287 pc = repo._phasecache
1287 1288 return subset.filter(lambda r: pc.phase(repo, r) == phases.public)
1288 1289
1289 1290 def remote(repo, subset, x):
1290 1291 """``remote([id [,path]])``
1291 1292 Local revision that corresponds to the given identifier in a
1292 1293 remote repository, if present. Here, the '.' identifier is a
1293 1294 synonym for the current local branch.
1294 1295 """
1295 1296
1296 1297 import hg # avoid start-up nasties
1297 1298 # i18n: "remote" is a keyword
1298 1299 l = getargs(x, 0, 2, _("remote takes one, two or no arguments"))
1299 1300
1300 1301 q = '.'
1301 1302 if len(l) > 0:
1302 1303 # i18n: "remote" is a keyword
1303 1304 q = getstring(l[0], _("remote requires a string id"))
1304 1305 if q == '.':
1305 1306 q = repo['.'].branch()
1306 1307
1307 1308 dest = ''
1308 1309 if len(l) > 1:
1309 1310 # i18n: "remote" is a keyword
1310 1311 dest = getstring(l[1], _("remote requires a repository path"))
1311 1312 dest = repo.ui.expandpath(dest or 'default')
1312 1313 dest, branches = hg.parseurl(dest)
1313 1314 revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
1314 1315 if revs:
1315 1316 revs = [repo.lookup(rev) for rev in revs]
1316 1317 other = hg.peer(repo, {}, dest)
1317 1318 n = other.lookup(q)
1318 1319 if n in repo:
1319 1320 r = repo[n].rev()
1320 1321 if r in subset:
1321 1322 return baseset([r])
1322 1323 return baseset([])
1323 1324
1324 1325 def removes(repo, subset, x):
1325 1326 """``removes(pattern)``
1326 1327 Changesets which remove files matching pattern.
1327 1328
1328 1329 The pattern without explicit kind like ``glob:`` is expected to be
1329 1330 relative to the current directory and match against a file or a
1330 1331 directory.
1331 1332 """
1332 1333 # i18n: "removes" is a keyword
1333 1334 pat = getstring(x, _("removes requires a pattern"))
1334 1335 return checkstatus(repo, subset, pat, 2)
1335 1336
1336 1337 def rev(repo, subset, x):
1337 1338 """``rev(number)``
1338 1339 Revision with the given numeric identifier.
1339 1340 """
1340 1341 # i18n: "rev" is a keyword
1341 1342 l = getargs(x, 1, 1, _("rev requires one argument"))
1342 1343 try:
1343 1344 # i18n: "rev" is a keyword
1344 1345 l = int(getstring(l[0], _("rev requires a number")))
1345 1346 except (TypeError, ValueError):
1346 1347 # i18n: "rev" is a keyword
1347 1348 raise error.ParseError(_("rev expects a number"))
1348 1349 return subset.filter(lambda r: r == l)
1349 1350
1350 1351 def matching(repo, subset, x):
1351 1352 """``matching(revision [, field])``
1352 1353 Changesets in which a given set of fields match the set of fields in the
1353 1354 selected revision or set.
1354 1355
1355 1356 To match more than one field pass the list of fields to match separated
1356 1357 by spaces (e.g. ``author description``).
1357 1358
1358 1359 Valid fields are most regular revision fields and some special fields.
1359 1360
1360 1361 Regular revision fields are ``description``, ``author``, ``branch``,
1361 1362 ``date``, ``files``, ``phase``, ``parents``, ``substate``, ``user``
1362 1363 and ``diff``.
1363 1364 Note that ``author`` and ``user`` are synonyms. ``diff`` refers to the
1364 1365 contents of the revision. Two revisions matching their ``diff`` will
1365 1366 also match their ``files``.
1366 1367
1367 1368 Special fields are ``summary`` and ``metadata``:
1368 1369 ``summary`` matches the first line of the description.
1369 1370 ``metadata`` is equivalent to matching ``description user date``
1370 1371 (i.e. it matches the main metadata fields).
1371 1372
1372 1373 ``metadata`` is the default field which is used when no fields are
1373 1374 specified. You can match more than one field at a time.
1374 1375 """
1375 1376 # i18n: "matching" is a keyword
1376 1377 l = getargs(x, 1, 2, _("matching takes 1 or 2 arguments"))
1377 1378
1378 1379 revs = getset(repo, baseset(repo.changelog), l[0])
1379 1380
1380 1381 fieldlist = ['metadata']
1381 1382 if len(l) > 1:
1382 1383 fieldlist = getstring(l[1],
1383 1384 # i18n: "matching" is a keyword
1384 1385 _("matching requires a string "
1385 1386 "as its second argument")).split()
1386 1387
1387 1388 # Make sure that there are no repeated fields,
1388 1389 # expand the 'special' 'metadata' field type
1389 1390 # and check the 'files' whenever we check the 'diff'
1390 1391 fields = []
1391 1392 for field in fieldlist:
1392 1393 if field == 'metadata':
1393 1394 fields += ['user', 'description', 'date']
1394 1395 elif field == 'diff':
1395 1396 # a revision matching the diff must also match the files
1396 1397 # since matching the diff is very costly, make sure to
1397 1398 # also match the files first
1398 1399 fields += ['files', 'diff']
1399 1400 else:
1400 1401 if field == 'author':
1401 1402 field = 'user'
1402 1403 fields.append(field)
1403 1404 fields = set(fields)
1404 1405 if 'summary' in fields and 'description' in fields:
1405 1406 # If a revision matches its description it also matches its summary
1406 1407 fields.discard('summary')
1407 1408
1408 1409 # We may want to match more than one field
1409 1410 # Not all fields take the same amount of time to be matched
1410 1411 # Sort the selected fields in order of increasing matching cost
1411 1412 fieldorder = ['phase', 'parents', 'user', 'date', 'branch', 'summary',
1412 1413 'files', 'description', 'substate', 'diff']
1413 1414 def fieldkeyfunc(f):
1414 1415 try:
1415 1416 return fieldorder.index(f)
1416 1417 except ValueError:
1417 1418 # assume an unknown field is very costly
1418 1419 return len(fieldorder)
1419 1420 fields = list(fields)
1420 1421 fields.sort(key=fieldkeyfunc)
1421 1422
1422 1423 # Each field will be matched with its own "getfield" function
1423 1424 # which will be added to the getfieldfuncs array of functions
1424 1425 getfieldfuncs = []
1425 1426 _funcs = {
1426 1427 'user': lambda r: repo[r].user(),
1427 1428 'branch': lambda r: repo[r].branch(),
1428 1429 'date': lambda r: repo[r].date(),
1429 1430 'description': lambda r: repo[r].description(),
1430 1431 'files': lambda r: repo[r].files(),
1431 1432 'parents': lambda r: repo[r].parents(),
1432 1433 'phase': lambda r: repo[r].phase(),
1433 1434 'substate': lambda r: repo[r].substate,
1434 1435 'summary': lambda r: repo[r].description().splitlines()[0],
1435 1436 'diff': lambda r: list(repo[r].diff(git=True),)
1436 1437 }
1437 1438 for info in fields:
1438 1439 getfield = _funcs.get(info, None)
1439 1440 if getfield is None:
1440 1441 raise error.ParseError(
1441 1442 # i18n: "matching" is a keyword
1442 1443 _("unexpected field name passed to matching: %s") % info)
1443 1444 getfieldfuncs.append(getfield)
1444 1445 # convert the getfield array of functions into a "getinfo" function
1445 1446 # which returns an array of field values (or a single value if there
1446 1447 # is only one field to match)
1447 1448 getinfo = lambda r: [f(r) for f in getfieldfuncs]
1448 1449
1449 1450 def matches(x):
1450 1451 for rev in revs:
1451 1452 target = getinfo(rev)
1452 1453 match = True
1453 1454 for n, f in enumerate(getfieldfuncs):
1454 1455 if target[n] != f(x):
1455 1456 match = False
1456 1457 if match:
1457 1458 return True
1458 1459 return False
1459 1460
1460 1461 return subset.filter(matches)
1461 1462
1462 1463 def reverse(repo, subset, x):
1463 1464 """``reverse(set)``
1464 1465 Reverse order of set.
1465 1466 """
1466 1467 l = getset(repo, subset, x)
1467 1468 l.reverse()
1468 1469 return l
1469 1470
1470 1471 def roots(repo, subset, x):
1471 1472 """``roots(set)``
1472 1473 Changesets in set with no parent changeset in set.
1473 1474 """
1474 1475 s = getset(repo, spanset(repo), x).set()
1475 1476 subset = baseset([r for r in s if r in subset.set()])
1476 1477 cs = _children(repo, subset, s)
1477 1478 return subset - cs
1478 1479
1479 1480 def secret(repo, subset, x):
1480 1481 """``secret()``
1481 1482 Changeset in secret phase."""
1482 1483 # i18n: "secret" is a keyword
1483 1484 getargs(x, 0, 0, _("secret takes no arguments"))
1484 1485 pc = repo._phasecache
1485 1486 return subset.filter(lambda x: pc.phase(repo, x) == phases.secret)
1486 1487
1487 1488 def sort(repo, subset, x):
1488 1489 """``sort(set[, [-]key...])``
1489 1490 Sort set by keys. The default sort order is ascending, specify a key
1490 1491 as ``-key`` to sort in descending order.
1491 1492
1492 1493 The keys can be:
1493 1494
1494 1495 - ``rev`` for the revision number,
1495 1496 - ``branch`` for the branch name,
1496 1497 - ``desc`` for the commit message (description),
1497 1498 - ``user`` for user name (``author`` can be used as an alias),
1498 1499 - ``date`` for the commit date
1499 1500 """
1500 1501 # i18n: "sort" is a keyword
1501 1502 l = getargs(x, 1, 2, _("sort requires one or two arguments"))
1502 1503 keys = "rev"
1503 1504 if len(l) == 2:
1504 1505 # i18n: "sort" is a keyword
1505 1506 keys = getstring(l[1], _("sort spec must be a string"))
1506 1507
1507 1508 s = l[0]
1508 1509 keys = keys.split()
1509 1510 l = []
1510 1511 def invert(s):
1511 1512 return "".join(chr(255 - ord(c)) for c in s)
1512 1513 revs = getset(repo, subset, s)
1513 1514 if keys == ["rev"]:
1514 1515 revs.sort()
1515 1516 return revs
1516 1517 elif keys == ["-rev"]:
1517 1518 revs.sort(reverse=True)
1518 1519 return revs
1519 1520 for r in revs:
1520 1521 c = repo[r]
1521 1522 e = []
1522 1523 for k in keys:
1523 1524 if k == 'rev':
1524 1525 e.append(r)
1525 1526 elif k == '-rev':
1526 1527 e.append(-r)
1527 1528 elif k == 'branch':
1528 1529 e.append(c.branch())
1529 1530 elif k == '-branch':
1530 1531 e.append(invert(c.branch()))
1531 1532 elif k == 'desc':
1532 1533 e.append(c.description())
1533 1534 elif k == '-desc':
1534 1535 e.append(invert(c.description()))
1535 1536 elif k in 'user author':
1536 1537 e.append(c.user())
1537 1538 elif k in '-user -author':
1538 1539 e.append(invert(c.user()))
1539 1540 elif k == 'date':
1540 1541 e.append(c.date()[0])
1541 1542 elif k == '-date':
1542 1543 e.append(-c.date()[0])
1543 1544 else:
1544 1545 raise error.ParseError(_("unknown sort key %r") % k)
1545 1546 e.append(r)
1546 1547 l.append(e)
1547 1548 l.sort()
1548 1549 return baseset([e[-1] for e in l])
1549 1550
1550 1551 def _stringmatcher(pattern):
1551 1552 """
1552 1553 accepts a string, possibly starting with 're:' or 'literal:' prefix.
1553 1554 returns the matcher name, pattern, and matcher function.
1554 1555 missing or unknown prefixes are treated as literal matches.
1555 1556
1556 1557 helper for tests:
1557 1558 >>> def test(pattern, *tests):
1558 1559 ... kind, pattern, matcher = _stringmatcher(pattern)
1559 1560 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
1560 1561
1561 1562 exact matching (no prefix):
1562 1563 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
1563 1564 ('literal', 'abcdefg', [False, False, True])
1564 1565
1565 1566 regex matching ('re:' prefix)
1566 1567 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
1567 1568 ('re', 'a.+b', [False, False, True])
1568 1569
1569 1570 force exact matches ('literal:' prefix)
1570 1571 >>> test('literal:re:foobar', 'foobar', 're:foobar')
1571 1572 ('literal', 're:foobar', [False, True])
1572 1573
1573 1574 unknown prefixes are ignored and treated as literals
1574 1575 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
1575 1576 ('literal', 'foo:bar', [False, False, True])
1576 1577 """
1577 1578 if pattern.startswith('re:'):
1578 1579 pattern = pattern[3:]
1579 1580 try:
1580 1581 regex = re.compile(pattern)
1581 1582 except re.error, e:
1582 1583 raise error.ParseError(_('invalid regular expression: %s')
1583 1584 % e)
1584 1585 return 're', pattern, regex.search
1585 1586 elif pattern.startswith('literal:'):
1586 1587 pattern = pattern[8:]
1587 1588 return 'literal', pattern, pattern.__eq__
1588 1589
1589 1590 def _substringmatcher(pattern):
1590 1591 kind, pattern, matcher = _stringmatcher(pattern)
1591 1592 if kind == 'literal':
1592 1593 matcher = lambda s: pattern in s
1593 1594 return kind, pattern, matcher
1594 1595
1595 1596 def tag(repo, subset, x):
1596 1597 """``tag([name])``
1597 1598 The specified tag by name, or all tagged revisions if no name is given.
1598 1599
1599 1600 If `name` starts with `re:`, the remainder of the name is treated as
1600 1601 a regular expression. To match a tag that actually starts with `re:`,
1601 1602 use the prefix `literal:`.
1602 1603 """
1603 1604 # i18n: "tag" is a keyword
1604 1605 args = getargs(x, 0, 1, _("tag takes one or no arguments"))
1605 1606 cl = repo.changelog
1606 1607 if args:
1607 1608 pattern = getstring(args[0],
1608 1609 # i18n: "tag" is a keyword
1609 1610 _('the argument to tag must be a string'))
1610 1611 kind, pattern, matcher = _stringmatcher(pattern)
1611 1612 if kind == 'literal':
1612 1613 # avoid resolving all tags
1613 1614 tn = repo._tagscache.tags.get(pattern, None)
1614 1615 if tn is None:
1615 1616 raise util.Abort(_("tag '%s' does not exist") % pattern)
1616 1617 s = set([repo[tn].rev()])
1617 1618 else:
1618 1619 s = set([cl.rev(n) for t, n in repo.tagslist() if matcher(t)])
1619 1620 else:
1620 1621 s = set([cl.rev(n) for t, n in repo.tagslist() if t != 'tip'])
1621 1622 return subset & s
1622 1623
1623 1624 def tagged(repo, subset, x):
1624 1625 return tag(repo, subset, x)
1625 1626
1626 1627 def unstable(repo, subset, x):
1627 1628 """``unstable()``
1628 1629 Non-obsolete changesets with obsolete ancestors.
1629 1630 """
1630 1631 # i18n: "unstable" is a keyword
1631 1632 getargs(x, 0, 0, _("unstable takes no arguments"))
1632 1633 unstables = obsmod.getrevs(repo, 'unstable')
1633 1634 return subset & unstables
1634 1635
1635 1636
1636 1637 def user(repo, subset, x):
1637 1638 """``user(string)``
1638 1639 User name contains string. The match is case-insensitive.
1639 1640
1640 1641 If `string` starts with `re:`, the remainder of the string is treated as
1641 1642 a regular expression. To match a user that actually contains `re:`, use
1642 1643 the prefix `literal:`.
1643 1644 """
1644 1645 return author(repo, subset, x)
1645 1646
1646 1647 # for internal use
1647 1648 def _list(repo, subset, x):
1648 1649 s = getstring(x, "internal error")
1649 1650 if not s:
1650 1651 return baseset([])
1651 1652 ls = [repo[r].rev() for r in s.split('\0')]
1652 1653 s = subset.set()
1653 1654 return baseset([r for r in ls if r in s])
1654 1655
1655 1656 # for internal use
1656 1657 def _intlist(repo, subset, x):
1657 1658 s = getstring(x, "internal error")
1658 1659 if not s:
1659 1660 return baseset([])
1660 1661 ls = [int(r) for r in s.split('\0')]
1661 1662 s = subset.set()
1662 1663 return baseset([r for r in ls if r in s])
1663 1664
1664 1665 # for internal use
1665 1666 def _hexlist(repo, subset, x):
1666 1667 s = getstring(x, "internal error")
1667 1668 if not s:
1668 1669 return baseset([])
1669 1670 cl = repo.changelog
1670 1671 ls = [cl.rev(node.bin(r)) for r in s.split('\0')]
1671 1672 s = subset.set()
1672 1673 return baseset([r for r in ls if r in s])
1673 1674
1674 1675 symbols = {
1675 1676 "adds": adds,
1676 1677 "all": getall,
1677 1678 "ancestor": ancestor,
1678 1679 "ancestors": ancestors,
1679 1680 "_firstancestors": _firstancestors,
1680 1681 "author": author,
1681 1682 "only": only,
1682 1683 "bisect": bisect,
1683 1684 "bisected": bisected,
1684 1685 "bookmark": bookmark,
1685 1686 "branch": branch,
1686 1687 "branchpoint": branchpoint,
1687 1688 "bumped": bumped,
1688 1689 "bundle": bundle,
1689 1690 "children": children,
1690 1691 "closed": closed,
1691 1692 "contains": contains,
1692 1693 "converted": converted,
1693 1694 "date": date,
1694 1695 "desc": desc,
1695 1696 "descendants": descendants,
1696 1697 "_firstdescendants": _firstdescendants,
1697 1698 "destination": destination,
1698 1699 "divergent": divergent,
1699 1700 "draft": draft,
1700 1701 "extinct": extinct,
1701 1702 "extra": extra,
1702 1703 "file": hasfile,
1703 1704 "filelog": filelog,
1704 1705 "first": first,
1705 1706 "follow": follow,
1706 1707 "_followfirst": _followfirst,
1707 1708 "grep": grep,
1708 1709 "head": head,
1709 1710 "heads": heads,
1710 1711 "hidden": hidden,
1711 1712 "id": node_,
1712 1713 "keyword": keyword,
1713 1714 "last": last,
1714 1715 "limit": limit,
1715 1716 "_matchfiles": _matchfiles,
1716 1717 "max": maxrev,
1717 1718 "merge": merge,
1718 1719 "min": minrev,
1719 1720 "modifies": modifies,
1720 1721 "obsolete": obsolete,
1721 1722 "origin": origin,
1722 1723 "outgoing": outgoing,
1723 1724 "p1": p1,
1724 1725 "p2": p2,
1725 1726 "parents": parents,
1726 1727 "present": present,
1727 1728 "public": public,
1728 1729 "remote": remote,
1729 1730 "removes": removes,
1730 1731 "rev": rev,
1731 1732 "reverse": reverse,
1732 1733 "roots": roots,
1733 1734 "sort": sort,
1734 1735 "secret": secret,
1735 1736 "matching": matching,
1736 1737 "tag": tag,
1737 1738 "tagged": tagged,
1738 1739 "user": user,
1739 1740 "unstable": unstable,
1740 1741 "_list": _list,
1741 1742 "_intlist": _intlist,
1742 1743 "_hexlist": _hexlist,
1743 1744 }
1744 1745
1745 1746 # symbols which can't be used for a DoS attack for any given input
1746 1747 # (e.g. those which accept regexes as plain strings shouldn't be included)
1747 1748 # functions that just return a lot of changesets (like all) don't count here
1748 1749 safesymbols = set([
1749 1750 "adds",
1750 1751 "all",
1751 1752 "ancestor",
1752 1753 "ancestors",
1753 1754 "_firstancestors",
1754 1755 "author",
1755 1756 "bisect",
1756 1757 "bisected",
1757 1758 "bookmark",
1758 1759 "branch",
1759 1760 "branchpoint",
1760 1761 "bumped",
1761 1762 "bundle",
1762 1763 "children",
1763 1764 "closed",
1764 1765 "converted",
1765 1766 "date",
1766 1767 "desc",
1767 1768 "descendants",
1768 1769 "_firstdescendants",
1769 1770 "destination",
1770 1771 "divergent",
1771 1772 "draft",
1772 1773 "extinct",
1773 1774 "extra",
1774 1775 "file",
1775 1776 "filelog",
1776 1777 "first",
1777 1778 "follow",
1778 1779 "_followfirst",
1779 1780 "head",
1780 1781 "heads",
1781 1782 "hidden",
1782 1783 "id",
1783 1784 "keyword",
1784 1785 "last",
1785 1786 "limit",
1786 1787 "_matchfiles",
1787 1788 "max",
1788 1789 "merge",
1789 1790 "min",
1790 1791 "modifies",
1791 1792 "obsolete",
1792 1793 "origin",
1793 1794 "outgoing",
1794 1795 "p1",
1795 1796 "p2",
1796 1797 "parents",
1797 1798 "present",
1798 1799 "public",
1799 1800 "remote",
1800 1801 "removes",
1801 1802 "rev",
1802 1803 "reverse",
1803 1804 "roots",
1804 1805 "sort",
1805 1806 "secret",
1806 1807 "matching",
1807 1808 "tag",
1808 1809 "tagged",
1809 1810 "user",
1810 1811 "unstable",
1811 1812 "_list",
1812 1813 "_intlist",
1813 1814 "_hexlist",
1814 1815 ])
1815 1816
1816 1817 methods = {
1817 1818 "range": rangeset,
1818 1819 "dagrange": dagrange,
1819 1820 "string": stringset,
1820 1821 "symbol": symbolset,
1821 1822 "and": andset,
1822 1823 "or": orset,
1823 1824 "not": notset,
1824 1825 "list": listset,
1825 1826 "func": func,
1826 1827 "ancestor": ancestorspec,
1827 1828 "parent": parentspec,
1828 1829 "parentpost": p1,
1829 1830 }
1830 1831
1831 1832 def optimize(x, small):
1832 1833 if x is None:
1833 1834 return 0, x
1834 1835
1835 1836 smallbonus = 1
1836 1837 if small:
1837 1838 smallbonus = .5
1838 1839
1839 1840 op = x[0]
1840 1841 if op == 'minus':
1841 1842 return optimize(('and', x[1], ('not', x[2])), small)
1842 1843 elif op == 'dagrangepre':
1843 1844 return optimize(('func', ('symbol', 'ancestors'), x[1]), small)
1844 1845 elif op == 'dagrangepost':
1845 1846 return optimize(('func', ('symbol', 'descendants'), x[1]), small)
1846 1847 elif op == 'rangepre':
1847 1848 return optimize(('range', ('string', '0'), x[1]), small)
1848 1849 elif op == 'rangepost':
1849 1850 return optimize(('range', x[1], ('string', 'tip')), small)
1850 1851 elif op == 'negate':
1851 1852 return optimize(('string',
1852 1853 '-' + getstring(x[1], _("can't negate that"))), small)
1853 1854 elif op in 'string symbol negate':
1854 1855 return smallbonus, x # single revisions are small
1855 1856 elif op == 'and':
1856 1857 wa, ta = optimize(x[1], True)
1857 1858 wb, tb = optimize(x[2], True)
1858 1859
1859 1860 # (::x and not ::y)/(not ::y and ::x) have a fast path
1860 1861 def isonly(revs, bases):
1861 1862 return (
1862 1863 revs[0] == 'func'
1863 1864 and getstring(revs[1], _('not a symbol')) == 'ancestors'
1864 1865 and bases[0] == 'not'
1865 1866 and bases[1][0] == 'func'
1866 1867 and getstring(bases[1][1], _('not a symbol')) == 'ancestors')
1867 1868
1868 1869 w = min(wa, wb)
1869 1870 if isonly(ta, tb):
1870 1871 return w, ('func', ('symbol', 'only'), ('list', ta[2], tb[1][2]))
1871 1872 if isonly(tb, ta):
1872 1873 return w, ('func', ('symbol', 'only'), ('list', tb[2], ta[1][2]))
1873 1874
1874 1875 if wa > wb:
1875 1876 return w, (op, tb, ta)
1876 1877 return w, (op, ta, tb)
1877 1878 elif op == 'or':
1878 1879 wa, ta = optimize(x[1], False)
1879 1880 wb, tb = optimize(x[2], False)
1880 1881 if wb < wa:
1881 1882 wb, wa = wa, wb
1882 1883 return max(wa, wb), (op, ta, tb)
1883 1884 elif op == 'not':
1884 1885 o = optimize(x[1], not small)
1885 1886 return o[0], (op, o[1])
1886 1887 elif op == 'parentpost':
1887 1888 o = optimize(x[1], small)
1888 1889 return o[0], (op, o[1])
1889 1890 elif op == 'group':
1890 1891 return optimize(x[1], small)
1891 1892 elif op in 'dagrange range list parent ancestorspec':
1892 1893 if op == 'parent':
1893 1894 # x^:y means (x^) : y, not x ^ (:y)
1894 1895 post = ('parentpost', x[1])
1895 1896 if x[2][0] == 'dagrangepre':
1896 1897 return optimize(('dagrange', post, x[2][1]), small)
1897 1898 elif x[2][0] == 'rangepre':
1898 1899 return optimize(('range', post, x[2][1]), small)
1899 1900
1900 1901 wa, ta = optimize(x[1], small)
1901 1902 wb, tb = optimize(x[2], small)
1902 1903 return wa + wb, (op, ta, tb)
1903 1904 elif op == 'func':
1904 1905 f = getstring(x[1], _("not a symbol"))
1905 1906 wa, ta = optimize(x[2], small)
1906 1907 if f in ("author branch closed date desc file grep keyword "
1907 1908 "outgoing user"):
1908 1909 w = 10 # slow
1909 1910 elif f in "modifies adds removes":
1910 1911 w = 30 # slower
1911 1912 elif f == "contains":
1912 1913 w = 100 # very slow
1913 1914 elif f == "ancestor":
1914 1915 w = 1 * smallbonus
1915 1916 elif f in "reverse limit first _intlist":
1916 1917 w = 0
1917 1918 elif f in "sort":
1918 1919 w = 10 # assume most sorts look at changelog
1919 1920 else:
1920 1921 w = 1
1921 1922 return w + wa, (op, x[1], ta)
1922 1923 return 1, x
1923 1924
1924 1925 _aliasarg = ('func', ('symbol', '_aliasarg'))
1925 1926 def _getaliasarg(tree):
1926 1927 """If tree matches ('func', ('symbol', '_aliasarg'), ('string', X))
1927 1928 return X, None otherwise.
1928 1929 """
1929 1930 if (len(tree) == 3 and tree[:2] == _aliasarg
1930 1931 and tree[2][0] == 'string'):
1931 1932 return tree[2][1]
1932 1933 return None
1933 1934
1934 1935 def _checkaliasarg(tree, known=None):
1935 1936 """Check tree contains no _aliasarg construct or only ones which
1936 1937 value is in known. Used to avoid alias placeholders injection.
1937 1938 """
1938 1939 if isinstance(tree, tuple):
1939 1940 arg = _getaliasarg(tree)
1940 1941 if arg is not None and (not known or arg not in known):
1941 1942 raise error.ParseError(_("not a function: %s") % '_aliasarg')
1942 1943 for t in tree:
1943 1944 _checkaliasarg(t, known)
1944 1945
1945 1946 class revsetalias(object):
1946 1947 funcre = re.compile('^([^(]+)\(([^)]+)\)$')
1947 1948 args = None
1948 1949
1949 1950 def __init__(self, name, value):
1950 1951 '''Aliases like:
1951 1952
1952 1953 h = heads(default)
1953 1954 b($1) = ancestors($1) - ancestors(default)
1954 1955 '''
1955 1956 m = self.funcre.search(name)
1956 1957 if m:
1957 1958 self.name = m.group(1)
1958 1959 self.tree = ('func', ('symbol', m.group(1)))
1959 1960 self.args = [x.strip() for x in m.group(2).split(',')]
1960 1961 for arg in self.args:
1961 1962 # _aliasarg() is an unknown symbol only used separate
1962 1963 # alias argument placeholders from regular strings.
1963 1964 value = value.replace(arg, '_aliasarg(%r)' % (arg,))
1964 1965 else:
1965 1966 self.name = name
1966 1967 self.tree = ('symbol', name)
1967 1968
1968 1969 self.replacement, pos = parse(value)
1969 1970 if pos != len(value):
1970 1971 raise error.ParseError(_('invalid token'), pos)
1971 1972 # Check for placeholder injection
1972 1973 _checkaliasarg(self.replacement, self.args)
1973 1974
1974 1975 def _getalias(aliases, tree):
1975 1976 """If tree looks like an unexpanded alias, return it. Return None
1976 1977 otherwise.
1977 1978 """
1978 1979 if isinstance(tree, tuple) and tree:
1979 1980 if tree[0] == 'symbol' and len(tree) == 2:
1980 1981 name = tree[1]
1981 1982 alias = aliases.get(name)
1982 1983 if alias and alias.args is None and alias.tree == tree:
1983 1984 return alias
1984 1985 if tree[0] == 'func' and len(tree) > 1:
1985 1986 if tree[1][0] == 'symbol' and len(tree[1]) == 2:
1986 1987 name = tree[1][1]
1987 1988 alias = aliases.get(name)
1988 1989 if alias and alias.args is not None and alias.tree == tree[:2]:
1989 1990 return alias
1990 1991 return None
1991 1992
1992 1993 def _expandargs(tree, args):
1993 1994 """Replace _aliasarg instances with the substitution value of the
1994 1995 same name in args, recursively.
1995 1996 """
1996 1997 if not tree or not isinstance(tree, tuple):
1997 1998 return tree
1998 1999 arg = _getaliasarg(tree)
1999 2000 if arg is not None:
2000 2001 return args[arg]
2001 2002 return tuple(_expandargs(t, args) for t in tree)
2002 2003
2003 2004 def _expandaliases(aliases, tree, expanding, cache):
2004 2005 """Expand aliases in tree, recursively.
2005 2006
2006 2007 'aliases' is a dictionary mapping user defined aliases to
2007 2008 revsetalias objects.
2008 2009 """
2009 2010 if not isinstance(tree, tuple):
2010 2011 # Do not expand raw strings
2011 2012 return tree
2012 2013 alias = _getalias(aliases, tree)
2013 2014 if alias is not None:
2014 2015 if alias in expanding:
2015 2016 raise error.ParseError(_('infinite expansion of revset alias "%s" '
2016 2017 'detected') % alias.name)
2017 2018 expanding.append(alias)
2018 2019 if alias.name not in cache:
2019 2020 cache[alias.name] = _expandaliases(aliases, alias.replacement,
2020 2021 expanding, cache)
2021 2022 result = cache[alias.name]
2022 2023 expanding.pop()
2023 2024 if alias.args is not None:
2024 2025 l = getlist(tree[2])
2025 2026 if len(l) != len(alias.args):
2026 2027 raise error.ParseError(
2027 2028 _('invalid number of arguments: %s') % len(l))
2028 2029 l = [_expandaliases(aliases, a, [], cache) for a in l]
2029 2030 result = _expandargs(result, dict(zip(alias.args, l)))
2030 2031 else:
2031 2032 result = tuple(_expandaliases(aliases, t, expanding, cache)
2032 2033 for t in tree)
2033 2034 return result
2034 2035
2035 2036 def findaliases(ui, tree):
2036 2037 _checkaliasarg(tree)
2037 2038 aliases = {}
2038 2039 for k, v in ui.configitems('revsetalias'):
2039 2040 alias = revsetalias(k, v)
2040 2041 aliases[alias.name] = alias
2041 2042 return _expandaliases(aliases, tree, [], {})
2042 2043
2043 2044 def parse(spec, lookup=None):
2044 2045 p = parser.parser(tokenize, elements)
2045 2046 return p.parse(spec, lookup=lookup)
2046 2047
2047 2048 def match(ui, spec, repo=None):
2048 2049 if not spec:
2049 2050 raise error.ParseError(_("empty query"))
2050 2051 lookup = None
2051 2052 if repo:
2052 2053 lookup = repo.__contains__
2053 2054 tree, pos = parse(spec, lookup)
2054 2055 if (pos != len(spec)):
2055 2056 raise error.ParseError(_("invalid token"), pos)
2056 2057 if ui:
2057 2058 tree = findaliases(ui, tree)
2058 2059 weight, tree = optimize(tree, True)
2059 2060 def mfunc(repo, subset):
2060 2061 if util.safehasattr(subset, 'set'):
2061 2062 return getset(repo, subset, tree)
2062 2063 return getset(repo, baseset(subset), tree)
2063 2064 return mfunc
2064 2065
2065 2066 def formatspec(expr, *args):
2066 2067 '''
2067 2068 This is a convenience function for using revsets internally, and
2068 2069 escapes arguments appropriately. Aliases are intentionally ignored
2069 2070 so that intended expression behavior isn't accidentally subverted.
2070 2071
2071 2072 Supported arguments:
2072 2073
2073 2074 %r = revset expression, parenthesized
2074 2075 %d = int(arg), no quoting
2075 2076 %s = string(arg), escaped and single-quoted
2076 2077 %b = arg.branch(), escaped and single-quoted
2077 2078 %n = hex(arg), single-quoted
2078 2079 %% = a literal '%'
2079 2080
2080 2081 Prefixing the type with 'l' specifies a parenthesized list of that type.
2081 2082
2082 2083 >>> formatspec('%r:: and %lr', '10 or 11', ("this()", "that()"))
2083 2084 '(10 or 11):: and ((this()) or (that()))'
2084 2085 >>> formatspec('%d:: and not %d::', 10, 20)
2085 2086 '10:: and not 20::'
2086 2087 >>> formatspec('%ld or %ld', [], [1])
2087 2088 "_list('') or 1"
2088 2089 >>> formatspec('keyword(%s)', 'foo\\xe9')
2089 2090 "keyword('foo\\\\xe9')"
2090 2091 >>> b = lambda: 'default'
2091 2092 >>> b.branch = b
2092 2093 >>> formatspec('branch(%b)', b)
2093 2094 "branch('default')"
2094 2095 >>> formatspec('root(%ls)', ['a', 'b', 'c', 'd'])
2095 2096 "root(_list('a\\x00b\\x00c\\x00d'))"
2096 2097 '''
2097 2098
2098 2099 def quote(s):
2099 2100 return repr(str(s))
2100 2101
2101 2102 def argtype(c, arg):
2102 2103 if c == 'd':
2103 2104 return str(int(arg))
2104 2105 elif c == 's':
2105 2106 return quote(arg)
2106 2107 elif c == 'r':
2107 2108 parse(arg) # make sure syntax errors are confined
2108 2109 return '(%s)' % arg
2109 2110 elif c == 'n':
2110 2111 return quote(node.hex(arg))
2111 2112 elif c == 'b':
2112 2113 return quote(arg.branch())
2113 2114
2114 2115 def listexp(s, t):
2115 2116 l = len(s)
2116 2117 if l == 0:
2117 2118 return "_list('')"
2118 2119 elif l == 1:
2119 2120 return argtype(t, s[0])
2120 2121 elif t == 'd':
2121 2122 return "_intlist('%s')" % "\0".join(str(int(a)) for a in s)
2122 2123 elif t == 's':
2123 2124 return "_list('%s')" % "\0".join(s)
2124 2125 elif t == 'n':
2125 2126 return "_hexlist('%s')" % "\0".join(node.hex(a) for a in s)
2126 2127 elif t == 'b':
2127 2128 return "_list('%s')" % "\0".join(a.branch() for a in s)
2128 2129
2129 2130 m = l // 2
2130 2131 return '(%s or %s)' % (listexp(s[:m], t), listexp(s[m:], t))
2131 2132
2132 2133 ret = ''
2133 2134 pos = 0
2134 2135 arg = 0
2135 2136 while pos < len(expr):
2136 2137 c = expr[pos]
2137 2138 if c == '%':
2138 2139 pos += 1
2139 2140 d = expr[pos]
2140 2141 if d == '%':
2141 2142 ret += d
2142 2143 elif d in 'dsnbr':
2143 2144 ret += argtype(d, args[arg])
2144 2145 arg += 1
2145 2146 elif d == 'l':
2146 2147 # a list of some type
2147 2148 pos += 1
2148 2149 d = expr[pos]
2149 2150 ret += listexp(list(args[arg]), d)
2150 2151 arg += 1
2151 2152 else:
2152 2153 raise util.Abort('unexpected revspec format character %s' % d)
2153 2154 else:
2154 2155 ret += c
2155 2156 pos += 1
2156 2157
2157 2158 return ret
2158 2159
2159 2160 def prettyformat(tree):
2160 2161 def _prettyformat(tree, level, lines):
2161 2162 if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
2162 2163 lines.append((level, str(tree)))
2163 2164 else:
2164 2165 lines.append((level, '(%s' % tree[0]))
2165 2166 for s in tree[1:]:
2166 2167 _prettyformat(s, level + 1, lines)
2167 2168 lines[-1:] = [(lines[-1][0], lines[-1][1] + ')')]
2168 2169
2169 2170 lines = []
2170 2171 _prettyformat(tree, 0, lines)
2171 2172 output = '\n'.join((' '*l + s) for l, s in lines)
2172 2173 return output
2173 2174
2174 2175 def depth(tree):
2175 2176 if isinstance(tree, tuple):
2176 2177 return max(map(depth, tree)) + 1
2177 2178 else:
2178 2179 return 0
2179 2180
2180 2181 def funcsused(tree):
2181 2182 if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
2182 2183 return set()
2183 2184 else:
2184 2185 funcs = set()
2185 2186 for s in tree[1:]:
2186 2187 funcs |= funcsused(s)
2187 2188 if tree[0] == 'func':
2188 2189 funcs.add(tree[1][1])
2189 2190 return funcs
2190 2191
2191 2192 class baseset(list):
2192 2193 """Basic data structure that represents a revset and contains the basic
2193 2194 operation that it should be able to perform.
2194 2195
2195 2196 Every method in this class should be implemented by any smartset class.
2196 2197 """
2197 2198 def __init__(self, data=()):
2198 2199 super(baseset, self).__init__(data)
2199 2200 self._set = None
2200 2201
2201 2202 def ascending(self):
2202 2203 """Sorts the set in ascending order (in place).
2203 2204
2204 2205 This is part of the mandatory API for smartset."""
2205 2206 self.sort()
2206 2207
2207 2208 def descending(self):
2208 2209 """Sorts the set in descending order (in place).
2209 2210
2210 2211 This is part of the mandatory API for smartset."""
2211 2212 self.sort(reverse=True)
2212 2213
2213 2214 def min(self):
2214 2215 return min(self)
2215 2216
2216 2217 def max(self):
2217 2218 return max(self)
2218 2219
2219 2220 def set(self):
2220 2221 """Returns a set or a smartset containing all the elements.
2221 2222
2222 2223 The returned structure should be the fastest option for membership
2223 2224 testing.
2224 2225
2225 2226 This is part of the mandatory API for smartset."""
2226 2227 if not self._set:
2227 2228 self._set = set(self)
2228 2229 return self._set
2229 2230
2230 2231 def __sub__(self, other):
2231 2232 """Returns a new object with the substraction of the two collections.
2232 2233
2233 2234 This is part of the mandatory API for smartset."""
2234 2235 # If we are operating on 2 baseset, do the computation now since all
2235 2236 # data is available. The alternative is to involve a lazyset, which
2236 2237 # may be slow.
2237 2238 if isinstance(other, baseset):
2238 2239 other = other.set()
2239 2240 return baseset([x for x in self if x not in other])
2240 2241
2241 2242 return self.filter(lambda x: x not in other)
2242 2243
2243 2244 def __and__(self, other):
2244 2245 """Returns a new object with the intersection of the two collections.
2245 2246
2246 2247 This is part of the mandatory API for smartset."""
2247 2248 if isinstance(other, baseset):
2248 2249 other = other.set()
2249 2250 return baseset([y for y in self if y in other])
2250 2251
2251 2252 def __add__(self, other):
2252 2253 """Returns a new object with the union of the two collections.
2253 2254
2254 2255 This is part of the mandatory API for smartset."""
2255 2256 s = self.set()
2256 2257 l = [r for r in other if r not in s]
2257 2258 return baseset(list(self) + l)
2258 2259
2259 2260 def isascending(self):
2260 2261 """Returns True if the collection is ascending order, False if not.
2261 2262
2262 2263 This is part of the mandatory API for smartset."""
2263 2264 return False
2264 2265
2265 2266 def isdescending(self):
2266 2267 """Returns True if the collection is descending order, False if not.
2267 2268
2268 2269 This is part of the mandatory API for smartset."""
2269 2270 return False
2270 2271
2271 2272 def filter(self, condition):
2272 2273 """Returns this smartset filtered by condition as a new smartset.
2273 2274
2274 2275 `condition` is a callable which takes a revision number and returns a
2275 2276 boolean.
2276 2277
2277 2278 This is part of the mandatory API for smartset."""
2278 2279 return lazyset(self, condition)
2279 2280
2280 2281 class _orderedsetmixin(object):
2281 2282 """Mixin class with utility methods for smartsets
2282 2283
2283 2284 This should be extended by smartsets which have the isascending(),
2284 2285 isdescending() and reverse() methods"""
2285 2286
2286 2287 def _first(self):
2287 2288 """return the first revision in the set"""
2288 2289 for r in self:
2289 2290 return r
2290 2291 raise ValueError('arg is an empty sequence')
2291 2292
2292 2293 def _last(self):
2293 2294 """return the last revision in the set"""
2294 2295 self.reverse()
2295 2296 m = self._first()
2296 2297 self.reverse()
2297 2298 return m
2298 2299
2299 2300 def min(self):
2300 2301 """return the smallest element in the set"""
2301 2302 if self.isascending():
2302 2303 return self._first()
2303 2304 return self._last()
2304 2305
2305 2306 def max(self):
2306 2307 """return the largest element in the set"""
2307 2308 if self.isascending():
2308 2309 return self._last()
2309 2310 return self._first()
2310 2311
2311 2312 class lazyset(object):
2312 2313 """Duck type for baseset class which iterates lazily over the revisions in
2313 2314 the subset and contains a function which tests for membership in the
2314 2315 revset
2315 2316 """
2316 2317 def __init__(self, subset, condition=lambda x: True):
2317 2318 """
2318 2319 condition: a function that decide whether a revision in the subset
2319 2320 belongs to the revset or not.
2320 2321 """
2321 2322 self._subset = subset
2322 2323 self._condition = condition
2323 2324 self._cache = {}
2324 2325
2325 2326 def ascending(self):
2326 2327 self._subset.sort()
2327 2328
2328 2329 def descending(self):
2329 2330 self._subset.sort(reverse=True)
2330 2331
2331 2332 def min(self):
2332 2333 return min(self)
2333 2334
2334 2335 def max(self):
2335 2336 return max(self)
2336 2337
2337 2338 def __contains__(self, x):
2338 2339 c = self._cache
2339 2340 if x not in c:
2340 2341 c[x] = x in self._subset and self._condition(x)
2341 2342 return c[x]
2342 2343
2343 2344 def __iter__(self):
2344 2345 cond = self._condition
2345 2346 for x in self._subset:
2346 2347 if cond(x):
2347 2348 yield x
2348 2349
2349 2350 def __and__(self, x):
2350 2351 return lazyset(self, x.__contains__)
2351 2352
2352 2353 def __sub__(self, x):
2353 2354 return lazyset(self, lambda r: r not in x)
2354 2355
2355 2356 def __add__(self, x):
2356 2357 return _addset(self, x)
2357 2358
2358 2359 def __nonzero__(self):
2359 2360 for r in self:
2360 2361 return True
2361 2362 return False
2362 2363
2363 2364 def __len__(self):
2364 2365 # Basic implementation to be changed in future patches.
2365 2366 l = baseset([r for r in self])
2366 2367 return len(l)
2367 2368
2368 2369 def __getitem__(self, x):
2369 2370 # Basic implementation to be changed in future patches.
2370 2371 l = baseset([r for r in self])
2371 2372 return l[x]
2372 2373
2373 2374 def sort(self, reverse=False):
2374 2375 if not util.safehasattr(self._subset, 'sort'):
2375 2376 self._subset = baseset(self._subset)
2376 2377 self._subset.sort(reverse=reverse)
2377 2378
2378 2379 def reverse(self):
2379 2380 self._subset.reverse()
2380 2381
2381 2382 def set(self):
2382 2383 return set([r for r in self])
2383 2384
2384 2385 def isascending(self):
2385 2386 return False
2386 2387
2387 2388 def isdescending(self):
2388 2389 return False
2389 2390
2390 2391 def filter(self, l):
2391 2392 return lazyset(self, l)
2392 2393
2393 2394 class orderedlazyset(_orderedsetmixin, lazyset):
2394 2395 """Subclass of lazyset which subset can be ordered either ascending or
2395 2396 descendingly
2396 2397 """
2397 2398 def __init__(self, subset, condition, ascending=True):
2398 2399 super(orderedlazyset, self).__init__(subset, condition)
2399 2400 self._ascending = ascending
2400 2401
2401 2402 def filter(self, l):
2402 2403 return orderedlazyset(self, l, ascending=self._ascending)
2403 2404
2404 2405 def ascending(self):
2405 2406 if not self._ascending:
2406 2407 self.reverse()
2407 2408
2408 2409 def descending(self):
2409 2410 if self._ascending:
2410 2411 self.reverse()
2411 2412
2412 2413 def __and__(self, x):
2413 2414 return orderedlazyset(self, x.__contains__,
2414 2415 ascending=self._ascending)
2415 2416
2416 2417 def __sub__(self, x):
2417 2418 return orderedlazyset(self, lambda r: r not in x,
2418 2419 ascending=self._ascending)
2419 2420
2420 2421 def __add__(self, x):
2421 2422 kwargs = {}
2422 2423 if self.isascending() and x.isascending():
2423 2424 kwargs['ascending'] = True
2424 2425 if self.isdescending() and x.isdescending():
2425 2426 kwargs['ascending'] = False
2426 2427 return _addset(self, x, **kwargs)
2427 2428
2428 2429 def sort(self, reverse=False):
2429 2430 if reverse:
2430 2431 if self._ascending:
2431 2432 self._subset.sort(reverse=reverse)
2432 2433 else:
2433 2434 if not self._ascending:
2434 2435 self._subset.sort(reverse=reverse)
2435 2436 self._ascending = not reverse
2436 2437
2437 2438 def isascending(self):
2438 2439 return self._ascending
2439 2440
2440 2441 def isdescending(self):
2441 2442 return not self._ascending
2442 2443
2443 2444 def reverse(self):
2444 2445 self._subset.reverse()
2445 2446 self._ascending = not self._ascending
2446 2447
2447 2448 class _addset(_orderedsetmixin):
2448 2449 """Represent the addition of two sets
2449 2450
2450 2451 Wrapper structure for lazily adding two structures without losing much
2451 2452 performance on the __contains__ method
2452 2453
2453 2454 If the ascending attribute is set, that means the two structures are
2454 2455 ordered in either an ascending or descending way. Therefore, we can add
2455 2456 them maintaining the order by iterating over both at the same time
2456 2457
2457 2458 This class does not duck-type baseset and it's only supposed to be used
2458 2459 internally
2459 2460 """
2460 2461 def __init__(self, revs1, revs2, ascending=None):
2461 2462 self._r1 = revs1
2462 2463 self._r2 = revs2
2463 2464 self._iter = None
2464 2465 self._ascending = ascending
2465 2466 self._genlist = None
2466 2467
2467 2468 def __len__(self):
2468 2469 return len(self._list)
2469 2470
2470 2471 @util.propertycache
2471 2472 def _list(self):
2472 2473 if not self._genlist:
2473 2474 self._genlist = baseset(self._iterator())
2474 2475 return self._genlist
2475 2476
2476 2477 def filter(self, condition):
2477 2478 if self._ascending is not None:
2478 2479 return orderedlazyset(self, condition, ascending=self._ascending)
2479 2480 return lazyset(self, condition)
2480 2481
2481 2482 def ascending(self):
2482 2483 if self._ascending is None:
2483 2484 self.sort()
2484 2485 self._ascending = True
2485 2486 else:
2486 2487 if not self._ascending:
2487 2488 self.reverse()
2488 2489
2489 2490 def descending(self):
2490 2491 if self._ascending is None:
2491 2492 self.sort(reverse=True)
2492 2493 self._ascending = False
2493 2494 else:
2494 2495 if self._ascending:
2495 2496 self.reverse()
2496 2497
2497 2498 def __and__(self, other):
2498 2499 filterfunc = other.__contains__
2499 2500 if self._ascending is not None:
2500 2501 return orderedlazyset(self, filterfunc, ascending=self._ascending)
2501 2502 return lazyset(self, filterfunc)
2502 2503
2503 2504 def __sub__(self, other):
2504 2505 filterfunc = lambda r: r not in other
2505 2506 if self._ascending is not None:
2506 2507 return orderedlazyset(self, filterfunc, ascending=self._ascending)
2507 2508 return lazyset(self, filterfunc)
2508 2509
2509 2510 def __add__(self, other):
2510 2511 """When both collections are ascending or descending, preserve the order
2511 2512 """
2512 2513 kwargs = {}
2513 2514 if self._ascending is not None:
2514 2515 if self.isascending() and other.isascending():
2515 2516 kwargs['ascending'] = True
2516 2517 if self.isdescending() and other.isdescending():
2517 2518 kwargs['ascending'] = False
2518 2519 return _addset(self, other, **kwargs)
2519 2520
2520 2521 def _iterator(self):
2521 2522 """Iterate over both collections without repeating elements
2522 2523
2523 2524 If the ascending attribute is not set, iterate over the first one and
2524 2525 then over the second one checking for membership on the first one so we
2525 2526 dont yield any duplicates.
2526 2527
2527 2528 If the ascending attribute is set, iterate over both collections at the
2528 2529 same time, yielding only one value at a time in the given order.
2529 2530 """
2530 2531 if not self._iter:
2531 2532 def gen():
2532 2533 if self._ascending is None:
2533 2534 for r in self._r1:
2534 2535 yield r
2535 2536 s = self._r1.set()
2536 2537 for r in self._r2:
2537 2538 if r not in s:
2538 2539 yield r
2539 2540 else:
2540 2541 iter1 = iter(self._r1)
2541 2542 iter2 = iter(self._r2)
2542 2543
2543 2544 val1 = None
2544 2545 val2 = None
2545 2546
2546 2547 choice = max
2547 2548 if self._ascending:
2548 2549 choice = min
2549 2550 try:
2550 2551 # Consume both iterators in an ordered way until one is
2551 2552 # empty
2552 2553 while True:
2553 2554 if val1 is None:
2554 2555 val1 = iter1.next()
2555 2556 if val2 is None:
2556 2557 val2 = iter2.next()
2557 2558 next = choice(val1, val2)
2558 2559 yield next
2559 2560 if val1 == next:
2560 2561 val1 = None
2561 2562 if val2 == next:
2562 2563 val2 = None
2563 2564 except StopIteration:
2564 2565 # Flush any remaining values and consume the other one
2565 2566 it = iter2
2566 2567 if val1 is not None:
2567 2568 yield val1
2568 2569 it = iter1
2569 2570 elif val2 is not None:
2570 2571 # might have been equality and both are empty
2571 2572 yield val2
2572 2573 for val in it:
2573 2574 yield val
2574 2575
2575 2576 self._iter = _generatorset(gen())
2576 2577
2577 2578 return self._iter
2578 2579
2579 2580 def __iter__(self):
2580 2581 if self._genlist:
2581 2582 return iter(self._genlist)
2582 2583 return iter(self._iterator())
2583 2584
2584 2585 def __contains__(self, x):
2585 2586 return x in self._r1 or x in self._r2
2586 2587
2587 2588 def set(self):
2588 2589 return self
2589 2590
2590 2591 def sort(self, reverse=False):
2591 2592 """Sort the added set
2592 2593
2593 2594 For this we use the cached list with all the generated values and if we
2594 2595 know they are ascending or descending we can sort them in a smart way.
2595 2596 """
2596 2597 if self._ascending is None:
2597 2598 self._list.sort(reverse=reverse)
2598 2599 self._ascending = not reverse
2599 2600 else:
2600 2601 if bool(self._ascending) == bool(reverse):
2601 2602 self.reverse()
2602 2603
2603 2604 def isascending(self):
2604 2605 return self._ascending is not None and self._ascending
2605 2606
2606 2607 def isdescending(self):
2607 2608 return self._ascending is not None and not self._ascending
2608 2609
2609 2610 def reverse(self):
2610 2611 self._list.reverse()
2611 2612 if self._ascending is not None:
2612 2613 self._ascending = not self._ascending
2613 2614
2614 2615 class _generatorset(object):
2615 2616 """Wrap a generator for lazy iteration
2616 2617
2617 2618 Wrapper structure for generators that provides lazy membership and can
2618 2619 be iterated more than once.
2619 2620 When asked for membership it generates values until either it finds the
2620 2621 requested one or has gone through all the elements in the generator
2621 2622
2622 2623 This class does not duck-type baseset and it's only supposed to be used
2623 2624 internally
2624 2625 """
2625 2626 def __init__(self, gen):
2626 2627 """
2627 2628 gen: a generator producing the values for the generatorset.
2628 2629 """
2629 2630 self._gen = gen
2630 2631 self._cache = {}
2631 2632 self._genlist = baseset([])
2632 2633 self._finished = False
2633 2634
2634 2635 def __contains__(self, x):
2635 2636 if x in self._cache:
2636 2637 return self._cache[x]
2637 2638
2638 2639 # Use new values only, as existing values would be cached.
2639 2640 for l in self._consumegen():
2640 2641 if l == x:
2641 2642 return True
2642 2643
2643 2644 self._cache[x] = False
2644 2645 return False
2645 2646
2646 2647 def __iter__(self):
2647 2648 if self._finished:
2648 2649 for x in self._genlist:
2649 2650 yield x
2650 2651 return
2651 2652
2652 2653 # We have to use this complex iteration strategy to allow multiple
2653 2654 # iterations at the same time. We need to be able to catch revision
2654 2655 # removed from `consumegen` and added to genlist in another instance.
2655 2656 #
2656 2657 # Getting rid of it would provide an about 15% speed up on this
2657 2658 # iteration.
2658 2659 i = 0
2659 2660 genlist = self._genlist
2660 2661 consume = self._consumegen()
2661 2662 while True:
2662 2663 if i < len(genlist):
2663 2664 yield genlist[i]
2664 2665 else:
2665 2666 yield consume.next()
2666 2667 i += 1
2667 2668
2668 2669 def _consumegen(self):
2669 2670 for item in self._gen:
2670 2671 self._cache[item] = True
2671 2672 self._genlist.append(item)
2672 2673 yield item
2673 2674 self._finished = True
2674 2675
2675 2676 def set(self):
2676 2677 return self
2677 2678
2678 2679 def sort(self, reverse=False):
2679 2680 if not self._finished:
2680 2681 for i in self:
2681 2682 continue
2682 2683 self._genlist.sort(reverse=reverse)
2683 2684
2684 2685 class _ascgeneratorset(_generatorset):
2685 2686 """Wrap a generator of ascending elements for lazy iteration
2686 2687
2687 2688 Same structure as _generatorset but stops iterating after it goes past
2688 2689 the value when asked for membership and the element is not contained
2689 2690
2690 2691 This class does not duck-type baseset and it's only supposed to be used
2691 2692 internally
2692 2693 """
2693 2694 def __contains__(self, x):
2694 2695 if x in self._cache:
2695 2696 return self._cache[x]
2696 2697
2697 2698 # Use new values only, as existing values would be cached.
2698 2699 for l in self._consumegen():
2699 2700 if l == x:
2700 2701 return True
2701 2702 if l > x:
2702 2703 break
2703 2704
2704 2705 self._cache[x] = False
2705 2706 return False
2706 2707
2707 2708 class _descgeneratorset(_generatorset):
2708 2709 """Wrap a generator of descending elements for lazy iteration
2709 2710
2710 2711 Same structure as _generatorset but stops iterating after it goes past
2711 2712 the value when asked for membership and the element is not contained
2712 2713
2713 2714 This class does not duck-type baseset and it's only supposed to be used
2714 2715 internally
2715 2716 """
2716 2717 def __contains__(self, x):
2717 2718 if x in self._cache:
2718 2719 return self._cache[x]
2719 2720
2720 2721 # Use new values only, as existing values would be cached.
2721 2722 for l in self._consumegen():
2722 2723 if l == x:
2723 2724 return True
2724 2725 if l < x:
2725 2726 break
2726 2727
2727 2728 self._cache[x] = False
2728 2729 return False
2729 2730
2730 2731 class spanset(_orderedsetmixin):
2731 2732 """Duck type for baseset class which represents a range of revisions and
2732 2733 can work lazily and without having all the range in memory
2733 2734
2734 2735 Note that spanset(x, y) behave almost like xrange(x, y) except for two
2735 2736 notable points:
2736 2737 - when x < y it will be automatically descending,
2737 2738 - revision filtered with this repoview will be skipped.
2738 2739
2739 2740 """
2740 2741 def __init__(self, repo, start=0, end=None):
2741 2742 """
2742 2743 start: first revision included the set
2743 2744 (default to 0)
2744 2745 end: first revision excluded (last+1)
2745 2746 (default to len(repo)
2746 2747
2747 2748 Spanset will be descending if `end` < `start`.
2748 2749 """
2749 2750 self._start = start
2750 2751 if end is not None:
2751 2752 self._end = end
2752 2753 else:
2753 2754 self._end = len(repo)
2754 2755 self._hiddenrevs = repo.changelog.filteredrevs
2755 2756
2756 2757 def ascending(self):
2757 2758 if not self.isascending():
2758 2759 self.reverse()
2759 2760
2760 2761 def descending(self):
2761 2762 if not self.isdescending():
2762 2763 self.reverse()
2763 2764
2764 2765 def __iter__(self):
2765 2766 if self.isascending():
2766 2767 iterrange = xrange(self._start, self._end)
2767 2768 else:
2768 2769 iterrange = xrange(self._start, self._end, -1)
2769 2770
2770 2771 if self._hiddenrevs:
2771 2772 s = self._hiddenrevs
2772 2773 for r in iterrange:
2773 2774 if r not in s:
2774 2775 yield r
2775 2776 else:
2776 2777 for r in iterrange:
2777 2778 yield r
2778 2779
2779 2780 def __contains__(self, rev):
2780 2781 return (((self._end < rev <= self._start)
2781 2782 or (self._start <= rev < self._end))
2782 2783 and not (self._hiddenrevs and rev in self._hiddenrevs))
2783 2784
2784 2785 def __nonzero__(self):
2785 2786 for r in self:
2786 2787 return True
2787 2788 return False
2788 2789
2789 2790 def __and__(self, x):
2790 2791 if isinstance(x, baseset):
2791 2792 x = x.set()
2792 2793 return orderedlazyset(self, x.__contains__,
2793 2794 ascending=self.isascending())
2794 2795
2795 2796 def __sub__(self, x):
2796 2797 if isinstance(x, baseset):
2797 2798 x = x.set()
2798 2799 return orderedlazyset(self, lambda r: r not in x,
2799 2800 ascending=self.isascending())
2800 2801
2801 2802 def __add__(self, x):
2802 2803 kwargs = {}
2803 2804 if self.isascending() and x.isascending():
2804 2805 kwargs['ascending'] = True
2805 2806 if self.isdescending() and x.isdescending():
2806 2807 kwargs['ascending'] = False
2807 2808 return _addset(self, x, **kwargs)
2808 2809
2809 2810 def __len__(self):
2810 2811 if not self._hiddenrevs:
2811 2812 return abs(self._end - self._start)
2812 2813 else:
2813 2814 count = 0
2814 2815 start = self._start
2815 2816 end = self._end
2816 2817 for rev in self._hiddenrevs:
2817 2818 if (end < rev <= start) or (start <= rev < end):
2818 2819 count += 1
2819 2820 return abs(self._end - self._start) - count
2820 2821
2821 2822 def __getitem__(self, x):
2822 2823 # Basic implementation to be changed in future patches.
2823 2824 l = baseset([r for r in self])
2824 2825 return l[x]
2825 2826
2826 2827 def sort(self, reverse=False):
2827 2828 if bool(reverse) != (self._start > self._end):
2828 2829 self.reverse()
2829 2830
2830 2831 def reverse(self):
2831 2832 # Just switch the _start and _end parameters
2832 2833 if self.isascending():
2833 2834 self._start, self._end = self._end - 1, self._start - 1
2834 2835 else:
2835 2836 self._start, self._end = self._end + 1, self._start + 1
2836 2837
2837 2838 def set(self):
2838 2839 return self
2839 2840
2840 2841 def isascending(self):
2841 2842 return self._start <= self._end
2842 2843
2843 2844 def isdescending(self):
2844 2845 return self._start >= self._end
2845 2846
2846 2847 def filter(self, l):
2847 2848 return orderedlazyset(self, l, ascending=self.isascending())
2848 2849
2849 2850 # tell hggettext to extract docstrings from these functions:
2850 2851 i18nfunctions = symbols.values()
General Comments 0
You need to be logged in to leave comments. Login now