##// END OF EJS Templates
revset: fix generatorset race condition...
Durham Goode -
r20833:47d43e23 default
parent child Browse files
Show More
@@ -1,2849 +1,2849
1 1 # revset.py - revision set queries for mercurial
2 2 #
3 3 # Copyright 2010 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 import re
9 9 import parser, util, error, discovery, hbisect, phases
10 10 import node
11 11 import heapq
12 12 import match as matchmod
13 13 import ancestor as ancestormod
14 14 from i18n import _
15 15 import encoding
16 16 import obsolete as obsmod
17 17 import pathutil
18 18 import repoview
19 19
20 20 def _revancestors(repo, revs, followfirst):
21 21 """Like revlog.ancestors(), but supports followfirst."""
22 22 cut = followfirst and 1 or None
23 23 cl = repo.changelog
24 24
25 25 def iterate():
26 26 revqueue, revsnode = None, None
27 27 h = []
28 28
29 29 revs.descending()
30 30 revqueue = util.deque(revs)
31 31 if revqueue:
32 32 revsnode = revqueue.popleft()
33 33 heapq.heappush(h, -revsnode)
34 34
35 35 seen = set([node.nullrev])
36 36 while h:
37 37 current = -heapq.heappop(h)
38 38 if current not in seen:
39 39 if revsnode and current == revsnode:
40 40 if revqueue:
41 41 revsnode = revqueue.popleft()
42 42 heapq.heappush(h, -revsnode)
43 43 seen.add(current)
44 44 yield current
45 45 for parent in cl.parentrevs(current)[:cut]:
46 46 if parent != node.nullrev:
47 47 heapq.heappush(h, -parent)
48 48
49 49 return _descgeneratorset(iterate())
50 50
51 51 def _revdescendants(repo, revs, followfirst):
52 52 """Like revlog.descendants() but supports followfirst."""
53 53 cut = followfirst and 1 or None
54 54
55 55 def iterate():
56 56 cl = repo.changelog
57 57 first = min(revs)
58 58 nullrev = node.nullrev
59 59 if first == nullrev:
60 60 # Are there nodes with a null first parent and a non-null
61 61 # second one? Maybe. Do we care? Probably not.
62 62 for i in cl:
63 63 yield i
64 64 else:
65 65 seen = set(revs)
66 66 for i in cl.revs(first + 1):
67 67 for x in cl.parentrevs(i)[:cut]:
68 68 if x != nullrev and x in seen:
69 69 seen.add(i)
70 70 yield i
71 71 break
72 72
73 73 return _ascgeneratorset(iterate())
74 74
75 75 def _revsbetween(repo, roots, heads):
76 76 """Return all paths between roots and heads, inclusive of both endpoint
77 77 sets."""
78 78 if not roots:
79 79 return baseset([])
80 80 parentrevs = repo.changelog.parentrevs
81 81 visit = baseset(heads)
82 82 reachable = set()
83 83 seen = {}
84 84 minroot = min(roots)
85 85 roots = set(roots)
86 86 # open-code the post-order traversal due to the tiny size of
87 87 # sys.getrecursionlimit()
88 88 while visit:
89 89 rev = visit.pop()
90 90 if rev in roots:
91 91 reachable.add(rev)
92 92 parents = parentrevs(rev)
93 93 seen[rev] = parents
94 94 for parent in parents:
95 95 if parent >= minroot and parent not in seen:
96 96 visit.append(parent)
97 97 if not reachable:
98 98 return baseset([])
99 99 for rev in sorted(seen):
100 100 for parent in seen[rev]:
101 101 if parent in reachable:
102 102 reachable.add(rev)
103 103 return baseset(sorted(reachable))
104 104
105 105 elements = {
106 106 "(": (20, ("group", 1, ")"), ("func", 1, ")")),
107 107 "~": (18, None, ("ancestor", 18)),
108 108 "^": (18, None, ("parent", 18), ("parentpost", 18)),
109 109 "-": (5, ("negate", 19), ("minus", 5)),
110 110 "::": (17, ("dagrangepre", 17), ("dagrange", 17),
111 111 ("dagrangepost", 17)),
112 112 "..": (17, ("dagrangepre", 17), ("dagrange", 17),
113 113 ("dagrangepost", 17)),
114 114 ":": (15, ("rangepre", 15), ("range", 15), ("rangepost", 15)),
115 115 "not": (10, ("not", 10)),
116 116 "!": (10, ("not", 10)),
117 117 "and": (5, None, ("and", 5)),
118 118 "&": (5, None, ("and", 5)),
119 119 "or": (4, None, ("or", 4)),
120 120 "|": (4, None, ("or", 4)),
121 121 "+": (4, None, ("or", 4)),
122 122 ",": (2, None, ("list", 2)),
123 123 ")": (0, None, None),
124 124 "symbol": (0, ("symbol",), None),
125 125 "string": (0, ("string",), None),
126 126 "end": (0, None, None),
127 127 }
128 128
129 129 keywords = set(['and', 'or', 'not'])
130 130
131 131 def tokenize(program, lookup=None):
132 132 '''
133 133 Parse a revset statement into a stream of tokens
134 134
135 135 Check that @ is a valid unquoted token character (issue3686):
136 136 >>> list(tokenize("@::"))
137 137 [('symbol', '@', 0), ('::', None, 1), ('end', None, 3)]
138 138
139 139 '''
140 140
141 141 pos, l = 0, len(program)
142 142 while pos < l:
143 143 c = program[pos]
144 144 if c.isspace(): # skip inter-token whitespace
145 145 pass
146 146 elif c == ':' and program[pos:pos + 2] == '::': # look ahead carefully
147 147 yield ('::', None, pos)
148 148 pos += 1 # skip ahead
149 149 elif c == '.' and program[pos:pos + 2] == '..': # look ahead carefully
150 150 yield ('..', None, pos)
151 151 pos += 1 # skip ahead
152 152 elif c in "():,-|&+!~^": # handle simple operators
153 153 yield (c, None, pos)
154 154 elif (c in '"\'' or c == 'r' and
155 155 program[pos:pos + 2] in ("r'", 'r"')): # handle quoted strings
156 156 if c == 'r':
157 157 pos += 1
158 158 c = program[pos]
159 159 decode = lambda x: x
160 160 else:
161 161 decode = lambda x: x.decode('string-escape')
162 162 pos += 1
163 163 s = pos
164 164 while pos < l: # find closing quote
165 165 d = program[pos]
166 166 if d == '\\': # skip over escaped characters
167 167 pos += 2
168 168 continue
169 169 if d == c:
170 170 yield ('string', decode(program[s:pos]), s)
171 171 break
172 172 pos += 1
173 173 else:
174 174 raise error.ParseError(_("unterminated string"), s)
175 175 # gather up a symbol/keyword
176 176 elif c.isalnum() or c in '._@' or ord(c) > 127:
177 177 s = pos
178 178 pos += 1
179 179 while pos < l: # find end of symbol
180 180 d = program[pos]
181 181 if not (d.isalnum() or d in "-._/@" or ord(d) > 127):
182 182 break
183 183 if d == '.' and program[pos - 1] == '.': # special case for ..
184 184 pos -= 1
185 185 break
186 186 pos += 1
187 187 sym = program[s:pos]
188 188 if sym in keywords: # operator keywords
189 189 yield (sym, None, s)
190 190 elif '-' in sym:
191 191 # some jerk gave us foo-bar-baz, try to check if it's a symbol
192 192 if lookup and lookup(sym):
193 193 # looks like a real symbol
194 194 yield ('symbol', sym, s)
195 195 else:
196 196 # looks like an expression
197 197 parts = sym.split('-')
198 198 for p in parts[:-1]:
199 199 if p: # possible consecutive -
200 200 yield ('symbol', p, s)
201 201 s += len(p)
202 202 yield ('-', None, pos)
203 203 s += 1
204 204 if parts[-1]: # possible trailing -
205 205 yield ('symbol', parts[-1], s)
206 206 else:
207 207 yield ('symbol', sym, s)
208 208 pos -= 1
209 209 else:
210 210 raise error.ParseError(_("syntax error"), pos)
211 211 pos += 1
212 212 yield ('end', None, pos)
213 213
214 214 # helpers
215 215
216 216 def getstring(x, err):
217 217 if x and (x[0] == 'string' or x[0] == 'symbol'):
218 218 return x[1]
219 219 raise error.ParseError(err)
220 220
221 221 def getlist(x):
222 222 if not x:
223 223 return []
224 224 if x[0] == 'list':
225 225 return getlist(x[1]) + [x[2]]
226 226 return [x]
227 227
228 228 def getargs(x, min, max, err):
229 229 l = getlist(x)
230 230 if len(l) < min or (max >= 0 and len(l) > max):
231 231 raise error.ParseError(err)
232 232 return l
233 233
234 234 def getset(repo, subset, x):
235 235 if not x:
236 236 raise error.ParseError(_("missing argument"))
237 237 s = methods[x[0]](repo, subset, *x[1:])
238 238 if util.safehasattr(s, 'set'):
239 239 return s
240 240 return baseset(s)
241 241
242 242 def _getrevsource(repo, r):
243 243 extra = repo[r].extra()
244 244 for label in ('source', 'transplant_source', 'rebase_source'):
245 245 if label in extra:
246 246 try:
247 247 return repo[extra[label]].rev()
248 248 except error.RepoLookupError:
249 249 pass
250 250 return None
251 251
252 252 # operator methods
253 253
254 254 def stringset(repo, subset, x):
255 255 x = repo[x].rev()
256 256 if x == -1 and len(subset) == len(repo):
257 257 return baseset([-1])
258 258 if len(subset) == len(repo) or x in subset:
259 259 return baseset([x])
260 260 return baseset([])
261 261
262 262 def symbolset(repo, subset, x):
263 263 if x in symbols:
264 264 raise error.ParseError(_("can't use %s here") % x)
265 265 return stringset(repo, subset, x)
266 266
267 267 def rangeset(repo, subset, x, y):
268 268 cl = baseset(repo.changelog)
269 269 m = getset(repo, cl, x)
270 270 n = getset(repo, cl, y)
271 271
272 272 if not m or not n:
273 273 return baseset([])
274 274 m, n = m[0], n[-1]
275 275
276 276 if m < n:
277 277 r = spanset(repo, m, n + 1)
278 278 else:
279 279 r = spanset(repo, m, n - 1)
280 280 return r & subset
281 281
282 282 def dagrange(repo, subset, x, y):
283 283 r = spanset(repo)
284 284 xs = _revsbetween(repo, getset(repo, r, x), getset(repo, r, y))
285 285 s = subset.set()
286 286 return xs.filter(lambda r: r in s)
287 287
288 288 def andset(repo, subset, x, y):
289 289 return getset(repo, getset(repo, subset, x), y)
290 290
291 291 def orset(repo, subset, x, y):
292 292 xl = getset(repo, subset, x)
293 293 yl = getset(repo, subset - xl, y)
294 294 return xl + yl
295 295
296 296 def notset(repo, subset, x):
297 297 return subset - getset(repo, subset, x)
298 298
299 299 def listset(repo, subset, a, b):
300 300 raise error.ParseError(_("can't use a list in this context"))
301 301
302 302 def func(repo, subset, a, b):
303 303 if a[0] == 'symbol' and a[1] in symbols:
304 304 return symbols[a[1]](repo, subset, b)
305 305 raise error.ParseError(_("not a function: %s") % a[1])
306 306
307 307 # functions
308 308
309 309 def adds(repo, subset, x):
310 310 """``adds(pattern)``
311 311 Changesets that add a file matching pattern.
312 312
313 313 The pattern without explicit kind like ``glob:`` is expected to be
314 314 relative to the current directory and match against a file or a
315 315 directory.
316 316 """
317 317 # i18n: "adds" is a keyword
318 318 pat = getstring(x, _("adds requires a pattern"))
319 319 return checkstatus(repo, subset, pat, 1)
320 320
321 321 def ancestor(repo, subset, x):
322 322 """``ancestor(*changeset)``
323 323 Greatest common ancestor of the changesets.
324 324
325 325 Accepts 0 or more changesets.
326 326 Will return empty list when passed no args.
327 327 Greatest common ancestor of a single changeset is that changeset.
328 328 """
329 329 # i18n: "ancestor" is a keyword
330 330 l = getlist(x)
331 331 rl = spanset(repo)
332 332 anc = None
333 333
334 334 # (getset(repo, rl, i) for i in l) generates a list of lists
335 335 rev = repo.changelog.rev
336 336 ancestor = repo.changelog.ancestor
337 337 node = repo.changelog.node
338 338 for revs in (getset(repo, rl, i) for i in l):
339 339 for r in revs:
340 340 if anc is None:
341 341 anc = r
342 342 else:
343 343 anc = rev(ancestor(node(anc), node(r)))
344 344
345 345 if anc is not None and anc in subset:
346 346 return baseset([anc])
347 347 return baseset([])
348 348
349 349 def _ancestors(repo, subset, x, followfirst=False):
350 350 args = getset(repo, spanset(repo), x)
351 351 if not args:
352 352 return baseset([])
353 353 s = _revancestors(repo, args, followfirst)
354 354 return subset.filter(lambda r: r in s)
355 355
356 356 def ancestors(repo, subset, x):
357 357 """``ancestors(set)``
358 358 Changesets that are ancestors of a changeset in set.
359 359 """
360 360 return _ancestors(repo, subset, x)
361 361
362 362 def _firstancestors(repo, subset, x):
363 363 # ``_firstancestors(set)``
364 364 # Like ``ancestors(set)`` but follows only the first parents.
365 365 return _ancestors(repo, subset, x, followfirst=True)
366 366
367 367 def ancestorspec(repo, subset, x, n):
368 368 """``set~n``
369 369 Changesets that are the Nth ancestor (first parents only) of a changeset
370 370 in set.
371 371 """
372 372 try:
373 373 n = int(n[1])
374 374 except (TypeError, ValueError):
375 375 raise error.ParseError(_("~ expects a number"))
376 376 ps = set()
377 377 cl = repo.changelog
378 378 for r in getset(repo, baseset(cl), x):
379 379 for i in range(n):
380 380 r = cl.parentrevs(r)[0]
381 381 ps.add(r)
382 382 return subset.filter(lambda r: r in ps)
383 383
384 384 def author(repo, subset, x):
385 385 """``author(string)``
386 386 Alias for ``user(string)``.
387 387 """
388 388 # i18n: "author" is a keyword
389 389 n = encoding.lower(getstring(x, _("author requires a string")))
390 390 kind, pattern, matcher = _substringmatcher(n)
391 391 return subset.filter(lambda x: matcher(encoding.lower(repo[x].user())))
392 392
393 393 def only(repo, subset, x):
394 394 """``only(set, [set])``
395 395 Changesets that are ancestors of the first set that are not ancestors
396 396 of any other head in the repo. If a second set is specified, the result
397 397 is ancestors of the first set that are not ancestors of the second set
398 398 (i.e. ::<set1> - ::<set2>).
399 399 """
400 400 cl = repo.changelog
401 401 args = getargs(x, 1, 2, _('only takes one or two arguments'))
402 402 include = getset(repo, spanset(repo), args[0]).set()
403 403 if len(args) == 1:
404 404 descendants = set(_revdescendants(repo, include, False))
405 405 exclude = [rev for rev in cl.headrevs()
406 406 if not rev in descendants and not rev in include]
407 407 else:
408 408 exclude = getset(repo, spanset(repo), args[1])
409 409
410 410 results = set(ancestormod.missingancestors(include, exclude, cl.parentrevs))
411 411 return lazyset(subset, lambda x: x in results)
412 412
413 413 def bisect(repo, subset, x):
414 414 """``bisect(string)``
415 415 Changesets marked in the specified bisect status:
416 416
417 417 - ``good``, ``bad``, ``skip``: csets explicitly marked as good/bad/skip
418 418 - ``goods``, ``bads`` : csets topologically good/bad
419 419 - ``range`` : csets taking part in the bisection
420 420 - ``pruned`` : csets that are goods, bads or skipped
421 421 - ``untested`` : csets whose fate is yet unknown
422 422 - ``ignored`` : csets ignored due to DAG topology
423 423 - ``current`` : the cset currently being bisected
424 424 """
425 425 # i18n: "bisect" is a keyword
426 426 status = getstring(x, _("bisect requires a string")).lower()
427 427 state = set(hbisect.get(repo, status))
428 428 return subset.filter(lambda r: r in state)
429 429
430 430 # Backward-compatibility
431 431 # - no help entry so that we do not advertise it any more
432 432 def bisected(repo, subset, x):
433 433 return bisect(repo, subset, x)
434 434
435 435 def bookmark(repo, subset, x):
436 436 """``bookmark([name])``
437 437 The named bookmark or all bookmarks.
438 438
439 439 If `name` starts with `re:`, the remainder of the name is treated as
440 440 a regular expression. To match a bookmark that actually starts with `re:`,
441 441 use the prefix `literal:`.
442 442 """
443 443 # i18n: "bookmark" is a keyword
444 444 args = getargs(x, 0, 1, _('bookmark takes one or no arguments'))
445 445 if args:
446 446 bm = getstring(args[0],
447 447 # i18n: "bookmark" is a keyword
448 448 _('the argument to bookmark must be a string'))
449 449 kind, pattern, matcher = _stringmatcher(bm)
450 450 if kind == 'literal':
451 451 bmrev = repo._bookmarks.get(bm, None)
452 452 if not bmrev:
453 453 raise util.Abort(_("bookmark '%s' does not exist") % bm)
454 454 bmrev = repo[bmrev].rev()
455 455 return subset.filter(lambda r: r == bmrev)
456 456 else:
457 457 matchrevs = set()
458 458 for name, bmrev in repo._bookmarks.iteritems():
459 459 if matcher(name):
460 460 matchrevs.add(bmrev)
461 461 if not matchrevs:
462 462 raise util.Abort(_("no bookmarks exist that match '%s'")
463 463 % pattern)
464 464 bmrevs = set()
465 465 for bmrev in matchrevs:
466 466 bmrevs.add(repo[bmrev].rev())
467 467 return subset & bmrevs
468 468
469 469 bms = set([repo[r].rev()
470 470 for r in repo._bookmarks.values()])
471 471 return subset.filter(lambda r: r in bms)
472 472
473 473 def branch(repo, subset, x):
474 474 """``branch(string or set)``
475 475 All changesets belonging to the given branch or the branches of the given
476 476 changesets.
477 477
478 478 If `string` starts with `re:`, the remainder of the name is treated as
479 479 a regular expression. To match a branch that actually starts with `re:`,
480 480 use the prefix `literal:`.
481 481 """
482 482 try:
483 483 b = getstring(x, '')
484 484 except error.ParseError:
485 485 # not a string, but another revspec, e.g. tip()
486 486 pass
487 487 else:
488 488 kind, pattern, matcher = _stringmatcher(b)
489 489 if kind == 'literal':
490 490 # note: falls through to the revspec case if no branch with
491 491 # this name exists
492 492 if pattern in repo.branchmap():
493 493 return subset.filter(lambda r: matcher(repo[r].branch()))
494 494 else:
495 495 return subset.filter(lambda r: matcher(repo[r].branch()))
496 496
497 497 s = getset(repo, spanset(repo), x)
498 498 b = set()
499 499 for r in s:
500 500 b.add(repo[r].branch())
501 501 s = s.set()
502 502 return subset.filter(lambda r: r in s or repo[r].branch() in b)
503 503
504 504 def bumped(repo, subset, x):
505 505 """``bumped()``
506 506 Mutable changesets marked as successors of public changesets.
507 507
508 508 Only non-public and non-obsolete changesets can be `bumped`.
509 509 """
510 510 # i18n: "bumped" is a keyword
511 511 getargs(x, 0, 0, _("bumped takes no arguments"))
512 512 bumped = obsmod.getrevs(repo, 'bumped')
513 513 return subset & bumped
514 514
515 515 def bundle(repo, subset, x):
516 516 """``bundle()``
517 517 Changesets in the bundle.
518 518
519 519 Bundle must be specified by the -R option."""
520 520
521 521 try:
522 522 bundlerevs = repo.changelog.bundlerevs
523 523 except AttributeError:
524 524 raise util.Abort(_("no bundle provided - specify with -R"))
525 525 return subset & bundlerevs
526 526
527 527 def checkstatus(repo, subset, pat, field):
528 528 hasset = matchmod.patkind(pat) == 'set'
529 529
530 530 def matches(x):
531 531 m = None
532 532 fname = None
533 533 c = repo[x]
534 534 if not m or hasset:
535 535 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
536 536 if not m.anypats() and len(m.files()) == 1:
537 537 fname = m.files()[0]
538 538 if fname is not None:
539 539 if fname not in c.files():
540 540 return False
541 541 else:
542 542 for f in c.files():
543 543 if m(f):
544 544 break
545 545 else:
546 546 return False
547 547 files = repo.status(c.p1().node(), c.node())[field]
548 548 if fname is not None:
549 549 if fname in files:
550 550 return True
551 551 else:
552 552 for f in files:
553 553 if m(f):
554 554 return True
555 555
556 556 return subset.filter(matches)
557 557
558 558 def _children(repo, narrow, parentset):
559 559 cs = set()
560 560 if not parentset:
561 561 return baseset(cs)
562 562 pr = repo.changelog.parentrevs
563 563 minrev = min(parentset)
564 564 for r in narrow:
565 565 if r <= minrev:
566 566 continue
567 567 for p in pr(r):
568 568 if p in parentset:
569 569 cs.add(r)
570 570 return baseset(cs)
571 571
572 572 def children(repo, subset, x):
573 573 """``children(set)``
574 574 Child changesets of changesets in set.
575 575 """
576 576 s = getset(repo, baseset(repo), x).set()
577 577 cs = _children(repo, subset, s)
578 578 return subset & cs
579 579
580 580 def closed(repo, subset, x):
581 581 """``closed()``
582 582 Changeset is closed.
583 583 """
584 584 # i18n: "closed" is a keyword
585 585 getargs(x, 0, 0, _("closed takes no arguments"))
586 586 return subset.filter(lambda r: repo[r].closesbranch())
587 587
588 588 def contains(repo, subset, x):
589 589 """``contains(pattern)``
590 590 Revision contains a file matching pattern. See :hg:`help patterns`
591 591 for information about file patterns.
592 592
593 593 The pattern without explicit kind like ``glob:`` is expected to be
594 594 relative to the current directory and match against a file exactly
595 595 for efficiency.
596 596 """
597 597 # i18n: "contains" is a keyword
598 598 pat = getstring(x, _("contains requires a pattern"))
599 599
600 600 def matches(x):
601 601 if not matchmod.patkind(pat):
602 602 pats = pathutil.canonpath(repo.root, repo.getcwd(), pat)
603 603 if pats in repo[x]:
604 604 return True
605 605 else:
606 606 c = repo[x]
607 607 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
608 608 for f in c.manifest():
609 609 if m(f):
610 610 return True
611 611 return False
612 612
613 613 return subset.filter(matches)
614 614
615 615 def converted(repo, subset, x):
616 616 """``converted([id])``
617 617 Changesets converted from the given identifier in the old repository if
618 618 present, or all converted changesets if no identifier is specified.
619 619 """
620 620
621 621 # There is exactly no chance of resolving the revision, so do a simple
622 622 # string compare and hope for the best
623 623
624 624 rev = None
625 625 # i18n: "converted" is a keyword
626 626 l = getargs(x, 0, 1, _('converted takes one or no arguments'))
627 627 if l:
628 628 # i18n: "converted" is a keyword
629 629 rev = getstring(l[0], _('converted requires a revision'))
630 630
631 631 def _matchvalue(r):
632 632 source = repo[r].extra().get('convert_revision', None)
633 633 return source is not None and (rev is None or source.startswith(rev))
634 634
635 635 return subset.filter(lambda r: _matchvalue(r))
636 636
637 637 def date(repo, subset, x):
638 638 """``date(interval)``
639 639 Changesets within the interval, see :hg:`help dates`.
640 640 """
641 641 # i18n: "date" is a keyword
642 642 ds = getstring(x, _("date requires a string"))
643 643 dm = util.matchdate(ds)
644 644 return subset.filter(lambda x: dm(repo[x].date()[0]))
645 645
646 646 def desc(repo, subset, x):
647 647 """``desc(string)``
648 648 Search commit message for string. The match is case-insensitive.
649 649 """
650 650 # i18n: "desc" is a keyword
651 651 ds = encoding.lower(getstring(x, _("desc requires a string")))
652 652
653 653 def matches(x):
654 654 c = repo[x]
655 655 return ds in encoding.lower(c.description())
656 656
657 657 return subset.filter(matches)
658 658
659 659 def _descendants(repo, subset, x, followfirst=False):
660 660 args = getset(repo, spanset(repo), x)
661 661 if not args:
662 662 return baseset([])
663 663 s = _revdescendants(repo, args, followfirst)
664 664 a = set(args)
665 665 return subset.filter(lambda r: r in s or r in a)
666 666
667 667 def descendants(repo, subset, x):
668 668 """``descendants(set)``
669 669 Changesets which are descendants of changesets in set.
670 670 """
671 671 return _descendants(repo, subset, x)
672 672
673 673 def _firstdescendants(repo, subset, x):
674 674 # ``_firstdescendants(set)``
675 675 # Like ``descendants(set)`` but follows only the first parents.
676 676 return _descendants(repo, subset, x, followfirst=True)
677 677
678 678 def destination(repo, subset, x):
679 679 """``destination([set])``
680 680 Changesets that were created by a graft, transplant or rebase operation,
681 681 with the given revisions specified as the source. Omitting the optional set
682 682 is the same as passing all().
683 683 """
684 684 if x is not None:
685 685 args = getset(repo, spanset(repo), x).set()
686 686 else:
687 687 args = getall(repo, spanset(repo), x).set()
688 688
689 689 dests = set()
690 690
691 691 # subset contains all of the possible destinations that can be returned, so
692 692 # iterate over them and see if their source(s) were provided in the args.
693 693 # Even if the immediate src of r is not in the args, src's source (or
694 694 # further back) may be. Scanning back further than the immediate src allows
695 695 # transitive transplants and rebases to yield the same results as transitive
696 696 # grafts.
697 697 for r in subset:
698 698 src = _getrevsource(repo, r)
699 699 lineage = None
700 700
701 701 while src is not None:
702 702 if lineage is None:
703 703 lineage = list()
704 704
705 705 lineage.append(r)
706 706
707 707 # The visited lineage is a match if the current source is in the arg
708 708 # set. Since every candidate dest is visited by way of iterating
709 709 # subset, any dests further back in the lineage will be tested by a
710 710 # different iteration over subset. Likewise, if the src was already
711 711 # selected, the current lineage can be selected without going back
712 712 # further.
713 713 if src in args or src in dests:
714 714 dests.update(lineage)
715 715 break
716 716
717 717 r = src
718 718 src = _getrevsource(repo, r)
719 719
720 720 return subset.filter(lambda r: r in dests)
721 721
722 722 def divergent(repo, subset, x):
723 723 """``divergent()``
724 724 Final successors of changesets with an alternative set of final successors.
725 725 """
726 726 # i18n: "divergent" is a keyword
727 727 getargs(x, 0, 0, _("divergent takes no arguments"))
728 728 divergent = obsmod.getrevs(repo, 'divergent')
729 729 return subset.filter(lambda r: r in divergent)
730 730
731 731 def draft(repo, subset, x):
732 732 """``draft()``
733 733 Changeset in draft phase."""
734 734 # i18n: "draft" is a keyword
735 735 getargs(x, 0, 0, _("draft takes no arguments"))
736 736 pc = repo._phasecache
737 737 return subset.filter(lambda r: pc.phase(repo, r) == phases.draft)
738 738
739 739 def extinct(repo, subset, x):
740 740 """``extinct()``
741 741 Obsolete changesets with obsolete descendants only.
742 742 """
743 743 # i18n: "extinct" is a keyword
744 744 getargs(x, 0, 0, _("extinct takes no arguments"))
745 745 extincts = obsmod.getrevs(repo, 'extinct')
746 746 return subset & extincts
747 747
748 748 def extra(repo, subset, x):
749 749 """``extra(label, [value])``
750 750 Changesets with the given label in the extra metadata, with the given
751 751 optional value.
752 752
753 753 If `value` starts with `re:`, the remainder of the value is treated as
754 754 a regular expression. To match a value that actually starts with `re:`,
755 755 use the prefix `literal:`.
756 756 """
757 757
758 758 # i18n: "extra" is a keyword
759 759 l = getargs(x, 1, 2, _('extra takes at least 1 and at most 2 arguments'))
760 760 # i18n: "extra" is a keyword
761 761 label = getstring(l[0], _('first argument to extra must be a string'))
762 762 value = None
763 763
764 764 if len(l) > 1:
765 765 # i18n: "extra" is a keyword
766 766 value = getstring(l[1], _('second argument to extra must be a string'))
767 767 kind, value, matcher = _stringmatcher(value)
768 768
769 769 def _matchvalue(r):
770 770 extra = repo[r].extra()
771 771 return label in extra and (value is None or matcher(extra[label]))
772 772
773 773 return subset.filter(lambda r: _matchvalue(r))
774 774
775 775 def filelog(repo, subset, x):
776 776 """``filelog(pattern)``
777 777 Changesets connected to the specified filelog.
778 778
779 779 For performance reasons, ``filelog()`` does not show every changeset
780 780 that affects the requested file(s). See :hg:`help log` for details. For
781 781 a slower, more accurate result, use ``file()``.
782 782
783 783 The pattern without explicit kind like ``glob:`` is expected to be
784 784 relative to the current directory and match against a file exactly
785 785 for efficiency.
786 786 """
787 787
788 788 # i18n: "filelog" is a keyword
789 789 pat = getstring(x, _("filelog requires a pattern"))
790 790 s = set()
791 791
792 792 if not matchmod.patkind(pat):
793 793 f = pathutil.canonpath(repo.root, repo.getcwd(), pat)
794 794 fl = repo.file(f)
795 795 for fr in fl:
796 796 s.add(fl.linkrev(fr))
797 797 else:
798 798 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=repo[None])
799 799 for f in repo[None]:
800 800 if m(f):
801 801 fl = repo.file(f)
802 802 for fr in fl:
803 803 s.add(fl.linkrev(fr))
804 804
805 805 return subset.filter(lambda r: r in s)
806 806
807 807 def first(repo, subset, x):
808 808 """``first(set, [n])``
809 809 An alias for limit().
810 810 """
811 811 return limit(repo, subset, x)
812 812
813 813 def _follow(repo, subset, x, name, followfirst=False):
814 814 l = getargs(x, 0, 1, _("%s takes no arguments or a filename") % name)
815 815 c = repo['.']
816 816 if l:
817 817 x = getstring(l[0], _("%s expected a filename") % name)
818 818 if x in c:
819 819 cx = c[x]
820 820 s = set(ctx.rev() for ctx in cx.ancestors(followfirst=followfirst))
821 821 # include the revision responsible for the most recent version
822 822 s.add(cx.linkrev())
823 823 else:
824 824 return baseset([])
825 825 else:
826 826 s = _revancestors(repo, baseset([c.rev()]), followfirst)
827 827
828 828 return subset.filter(lambda r: r in s)
829 829
830 830 def follow(repo, subset, x):
831 831 """``follow([file])``
832 832 An alias for ``::.`` (ancestors of the working copy's first parent).
833 833 If a filename is specified, the history of the given file is followed,
834 834 including copies.
835 835 """
836 836 return _follow(repo, subset, x, 'follow')
837 837
838 838 def _followfirst(repo, subset, x):
839 839 # ``followfirst([file])``
840 840 # Like ``follow([file])`` but follows only the first parent of
841 841 # every revision or file revision.
842 842 return _follow(repo, subset, x, '_followfirst', followfirst=True)
843 843
844 844 def getall(repo, subset, x):
845 845 """``all()``
846 846 All changesets, the same as ``0:tip``.
847 847 """
848 848 # i18n: "all" is a keyword
849 849 getargs(x, 0, 0, _("all takes no arguments"))
850 850 return subset
851 851
852 852 def grep(repo, subset, x):
853 853 """``grep(regex)``
854 854 Like ``keyword(string)`` but accepts a regex. Use ``grep(r'...')``
855 855 to ensure special escape characters are handled correctly. Unlike
856 856 ``keyword(string)``, the match is case-sensitive.
857 857 """
858 858 try:
859 859 # i18n: "grep" is a keyword
860 860 gr = re.compile(getstring(x, _("grep requires a string")))
861 861 except re.error, e:
862 862 raise error.ParseError(_('invalid match pattern: %s') % e)
863 863
864 864 def matches(x):
865 865 c = repo[x]
866 866 for e in c.files() + [c.user(), c.description()]:
867 867 if gr.search(e):
868 868 return True
869 869 return False
870 870
871 871 return subset.filter(matches)
872 872
873 873 def _matchfiles(repo, subset, x):
874 874 # _matchfiles takes a revset list of prefixed arguments:
875 875 #
876 876 # [p:foo, i:bar, x:baz]
877 877 #
878 878 # builds a match object from them and filters subset. Allowed
879 879 # prefixes are 'p:' for regular patterns, 'i:' for include
880 880 # patterns and 'x:' for exclude patterns. Use 'r:' prefix to pass
881 881 # a revision identifier, or the empty string to reference the
882 882 # working directory, from which the match object is
883 883 # initialized. Use 'd:' to set the default matching mode, default
884 884 # to 'glob'. At most one 'r:' and 'd:' argument can be passed.
885 885
886 886 # i18n: "_matchfiles" is a keyword
887 887 l = getargs(x, 1, -1, _("_matchfiles requires at least one argument"))
888 888 pats, inc, exc = [], [], []
889 889 hasset = False
890 890 rev, default = None, None
891 891 for arg in l:
892 892 # i18n: "_matchfiles" is a keyword
893 893 s = getstring(arg, _("_matchfiles requires string arguments"))
894 894 prefix, value = s[:2], s[2:]
895 895 if prefix == 'p:':
896 896 pats.append(value)
897 897 elif prefix == 'i:':
898 898 inc.append(value)
899 899 elif prefix == 'x:':
900 900 exc.append(value)
901 901 elif prefix == 'r:':
902 902 if rev is not None:
903 903 # i18n: "_matchfiles" is a keyword
904 904 raise error.ParseError(_('_matchfiles expected at most one '
905 905 'revision'))
906 906 rev = value
907 907 elif prefix == 'd:':
908 908 if default is not None:
909 909 # i18n: "_matchfiles" is a keyword
910 910 raise error.ParseError(_('_matchfiles expected at most one '
911 911 'default mode'))
912 912 default = value
913 913 else:
914 914 # i18n: "_matchfiles" is a keyword
915 915 raise error.ParseError(_('invalid _matchfiles prefix: %s') % prefix)
916 916 if not hasset and matchmod.patkind(value) == 'set':
917 917 hasset = True
918 918 if not default:
919 919 default = 'glob'
920 920
921 921 def matches(x):
922 922 m = None
923 923 c = repo[x]
924 924 if not m or (hasset and rev is None):
925 925 ctx = c
926 926 if rev is not None:
927 927 ctx = repo[rev or None]
928 928 m = matchmod.match(repo.root, repo.getcwd(), pats, include=inc,
929 929 exclude=exc, ctx=ctx, default=default)
930 930 for f in c.files():
931 931 if m(f):
932 932 return True
933 933 return False
934 934
935 935 return subset.filter(matches)
936 936
937 937 def hasfile(repo, subset, x):
938 938 """``file(pattern)``
939 939 Changesets affecting files matched by pattern.
940 940
941 941 For a faster but less accurate result, consider using ``filelog()``
942 942 instead.
943 943
944 944 This predicate uses ``glob:`` as the default kind of pattern.
945 945 """
946 946 # i18n: "file" is a keyword
947 947 pat = getstring(x, _("file requires a pattern"))
948 948 return _matchfiles(repo, subset, ('string', 'p:' + pat))
949 949
950 950 def head(repo, subset, x):
951 951 """``head()``
952 952 Changeset is a named branch head.
953 953 """
954 954 # i18n: "head" is a keyword
955 955 getargs(x, 0, 0, _("head takes no arguments"))
956 956 hs = set()
957 957 for b, ls in repo.branchmap().iteritems():
958 958 hs.update(repo[h].rev() for h in ls)
959 959 return baseset(hs).filter(subset.__contains__)
960 960
961 961 def heads(repo, subset, x):
962 962 """``heads(set)``
963 963 Members of set with no children in set.
964 964 """
965 965 s = getset(repo, subset, x)
966 966 ps = parents(repo, subset, x)
967 967 return s - ps
968 968
969 969 def hidden(repo, subset, x):
970 970 """``hidden()``
971 971 Hidden changesets.
972 972 """
973 973 # i18n: "hidden" is a keyword
974 974 getargs(x, 0, 0, _("hidden takes no arguments"))
975 975 hiddenrevs = repoview.filterrevs(repo, 'visible')
976 976 return subset & hiddenrevs
977 977
978 978 def keyword(repo, subset, x):
979 979 """``keyword(string)``
980 980 Search commit message, user name, and names of changed files for
981 981 string. The match is case-insensitive.
982 982 """
983 983 # i18n: "keyword" is a keyword
984 984 kw = encoding.lower(getstring(x, _("keyword requires a string")))
985 985
986 986 def matches(r):
987 987 c = repo[r]
988 988 return util.any(kw in encoding.lower(t) for t in c.files() + [c.user(),
989 989 c.description()])
990 990
991 991 return subset.filter(matches)
992 992
993 993 def limit(repo, subset, x):
994 994 """``limit(set, [n])``
995 995 First n members of set, defaulting to 1.
996 996 """
997 997 # i18n: "limit" is a keyword
998 998 l = getargs(x, 1, 2, _("limit requires one or two arguments"))
999 999 try:
1000 1000 lim = 1
1001 1001 if len(l) == 2:
1002 1002 # i18n: "limit" is a keyword
1003 1003 lim = int(getstring(l[1], _("limit requires a number")))
1004 1004 except (TypeError, ValueError):
1005 1005 # i18n: "limit" is a keyword
1006 1006 raise error.ParseError(_("limit expects a number"))
1007 1007 ss = subset.set()
1008 1008 os = getset(repo, spanset(repo), l[0])
1009 1009 bs = baseset([])
1010 1010 it = iter(os)
1011 1011 for x in xrange(lim):
1012 1012 try:
1013 1013 y = it.next()
1014 1014 if y in ss:
1015 1015 bs.append(y)
1016 1016 except (StopIteration):
1017 1017 break
1018 1018 return bs
1019 1019
1020 1020 def last(repo, subset, x):
1021 1021 """``last(set, [n])``
1022 1022 Last n members of set, defaulting to 1.
1023 1023 """
1024 1024 # i18n: "last" is a keyword
1025 1025 l = getargs(x, 1, 2, _("last requires one or two arguments"))
1026 1026 try:
1027 1027 lim = 1
1028 1028 if len(l) == 2:
1029 1029 # i18n: "last" is a keyword
1030 1030 lim = int(getstring(l[1], _("last requires a number")))
1031 1031 except (TypeError, ValueError):
1032 1032 # i18n: "last" is a keyword
1033 1033 raise error.ParseError(_("last expects a number"))
1034 1034 ss = subset.set()
1035 1035 os = getset(repo, spanset(repo), l[0])
1036 1036 os.reverse()
1037 1037 bs = baseset([])
1038 1038 it = iter(os)
1039 1039 for x in xrange(lim):
1040 1040 try:
1041 1041 y = it.next()
1042 1042 if y in ss:
1043 1043 bs.append(y)
1044 1044 except (StopIteration):
1045 1045 break
1046 1046 return bs
1047 1047
1048 1048 def maxrev(repo, subset, x):
1049 1049 """``max(set)``
1050 1050 Changeset with highest revision number in set.
1051 1051 """
1052 1052 os = getset(repo, spanset(repo), x)
1053 1053 if os:
1054 1054 m = os.max()
1055 1055 if m in subset:
1056 1056 return baseset([m])
1057 1057 return baseset([])
1058 1058
1059 1059 def merge(repo, subset, x):
1060 1060 """``merge()``
1061 1061 Changeset is a merge changeset.
1062 1062 """
1063 1063 # i18n: "merge" is a keyword
1064 1064 getargs(x, 0, 0, _("merge takes no arguments"))
1065 1065 cl = repo.changelog
1066 1066 return subset.filter(lambda r: cl.parentrevs(r)[1] != -1)
1067 1067
1068 1068 def branchpoint(repo, subset, x):
1069 1069 """``branchpoint()``
1070 1070 Changesets with more than one child.
1071 1071 """
1072 1072 # i18n: "branchpoint" is a keyword
1073 1073 getargs(x, 0, 0, _("branchpoint takes no arguments"))
1074 1074 cl = repo.changelog
1075 1075 if not subset:
1076 1076 return baseset([])
1077 1077 baserev = min(subset)
1078 1078 parentscount = [0]*(len(repo) - baserev)
1079 1079 for r in cl.revs(start=baserev + 1):
1080 1080 for p in cl.parentrevs(r):
1081 1081 if p >= baserev:
1082 1082 parentscount[p - baserev] += 1
1083 1083 return subset.filter(lambda r: parentscount[r - baserev] > 1)
1084 1084
1085 1085 def minrev(repo, subset, x):
1086 1086 """``min(set)``
1087 1087 Changeset with lowest revision number in set.
1088 1088 """
1089 1089 os = getset(repo, spanset(repo), x)
1090 1090 if os:
1091 1091 m = os.min()
1092 1092 if m in subset:
1093 1093 return baseset([m])
1094 1094 return baseset([])
1095 1095
1096 1096 def _missingancestors(repo, subset, x):
1097 1097 # i18n: "_missingancestors" is a keyword
1098 1098 revs, bases = getargs(x, 2, 2,
1099 1099 _("_missingancestors requires two arguments"))
1100 1100 rs = baseset(repo)
1101 1101 revs = getset(repo, rs, revs)
1102 1102 bases = getset(repo, rs, bases)
1103 1103 missing = set(repo.changelog.findmissingrevs(bases, revs))
1104 1104 return baseset([r for r in subset if r in missing])
1105 1105
1106 1106 def modifies(repo, subset, x):
1107 1107 """``modifies(pattern)``
1108 1108 Changesets modifying files matched by pattern.
1109 1109
1110 1110 The pattern without explicit kind like ``glob:`` is expected to be
1111 1111 relative to the current directory and match against a file or a
1112 1112 directory.
1113 1113 """
1114 1114 # i18n: "modifies" is a keyword
1115 1115 pat = getstring(x, _("modifies requires a pattern"))
1116 1116 return checkstatus(repo, subset, pat, 0)
1117 1117
1118 1118 def node_(repo, subset, x):
1119 1119 """``id(string)``
1120 1120 Revision non-ambiguously specified by the given hex string prefix.
1121 1121 """
1122 1122 # i18n: "id" is a keyword
1123 1123 l = getargs(x, 1, 1, _("id requires one argument"))
1124 1124 # i18n: "id" is a keyword
1125 1125 n = getstring(l[0], _("id requires a string"))
1126 1126 if len(n) == 40:
1127 1127 rn = repo[n].rev()
1128 1128 else:
1129 1129 rn = None
1130 1130 pm = repo.changelog._partialmatch(n)
1131 1131 if pm is not None:
1132 1132 rn = repo.changelog.rev(pm)
1133 1133
1134 1134 return subset.filter(lambda r: r == rn)
1135 1135
1136 1136 def obsolete(repo, subset, x):
1137 1137 """``obsolete()``
1138 1138 Mutable changeset with a newer version."""
1139 1139 # i18n: "obsolete" is a keyword
1140 1140 getargs(x, 0, 0, _("obsolete takes no arguments"))
1141 1141 obsoletes = obsmod.getrevs(repo, 'obsolete')
1142 1142 return subset & obsoletes
1143 1143
1144 1144 def origin(repo, subset, x):
1145 1145 """``origin([set])``
1146 1146 Changesets that were specified as a source for the grafts, transplants or
1147 1147 rebases that created the given revisions. Omitting the optional set is the
1148 1148 same as passing all(). If a changeset created by these operations is itself
1149 1149 specified as a source for one of these operations, only the source changeset
1150 1150 for the first operation is selected.
1151 1151 """
1152 1152 if x is not None:
1153 1153 args = getset(repo, spanset(repo), x).set()
1154 1154 else:
1155 1155 args = getall(repo, spanset(repo), x).set()
1156 1156
1157 1157 def _firstsrc(rev):
1158 1158 src = _getrevsource(repo, rev)
1159 1159 if src is None:
1160 1160 return None
1161 1161
1162 1162 while True:
1163 1163 prev = _getrevsource(repo, src)
1164 1164
1165 1165 if prev is None:
1166 1166 return src
1167 1167 src = prev
1168 1168
1169 1169 o = set([_firstsrc(r) for r in args])
1170 1170 return subset.filter(lambda r: r in o)
1171 1171
1172 1172 def outgoing(repo, subset, x):
1173 1173 """``outgoing([path])``
1174 1174 Changesets not found in the specified destination repository, or the
1175 1175 default push location.
1176 1176 """
1177 1177 import hg # avoid start-up nasties
1178 1178 # i18n: "outgoing" is a keyword
1179 1179 l = getargs(x, 0, 1, _("outgoing takes one or no arguments"))
1180 1180 # i18n: "outgoing" is a keyword
1181 1181 dest = l and getstring(l[0], _("outgoing requires a repository path")) or ''
1182 1182 dest = repo.ui.expandpath(dest or 'default-push', dest or 'default')
1183 1183 dest, branches = hg.parseurl(dest)
1184 1184 revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
1185 1185 if revs:
1186 1186 revs = [repo.lookup(rev) for rev in revs]
1187 1187 other = hg.peer(repo, {}, dest)
1188 1188 repo.ui.pushbuffer()
1189 1189 outgoing = discovery.findcommonoutgoing(repo, other, onlyheads=revs)
1190 1190 repo.ui.popbuffer()
1191 1191 cl = repo.changelog
1192 1192 o = set([cl.rev(r) for r in outgoing.missing])
1193 1193 return subset.filter(lambda r: r in o)
1194 1194
1195 1195 def p1(repo, subset, x):
1196 1196 """``p1([set])``
1197 1197 First parent of changesets in set, or the working directory.
1198 1198 """
1199 1199 if x is None:
1200 1200 p = repo[x].p1().rev()
1201 1201 return subset.filter(lambda r: r == p)
1202 1202
1203 1203 ps = set()
1204 1204 cl = repo.changelog
1205 1205 for r in getset(repo, spanset(repo), x):
1206 1206 ps.add(cl.parentrevs(r)[0])
1207 1207 return subset & ps
1208 1208
1209 1209 def p2(repo, subset, x):
1210 1210 """``p2([set])``
1211 1211 Second parent of changesets in set, or the working directory.
1212 1212 """
1213 1213 if x is None:
1214 1214 ps = repo[x].parents()
1215 1215 try:
1216 1216 p = ps[1].rev()
1217 1217 return subset.filter(lambda r: r == p)
1218 1218 except IndexError:
1219 1219 return baseset([])
1220 1220
1221 1221 ps = set()
1222 1222 cl = repo.changelog
1223 1223 for r in getset(repo, spanset(repo), x):
1224 1224 ps.add(cl.parentrevs(r)[1])
1225 1225 return subset & ps
1226 1226
1227 1227 def parents(repo, subset, x):
1228 1228 """``parents([set])``
1229 1229 The set of all parents for all changesets in set, or the working directory.
1230 1230 """
1231 1231 if x is None:
1232 1232 ps = tuple(p.rev() for p in repo[x].parents())
1233 1233 return subset & ps
1234 1234
1235 1235 ps = set()
1236 1236 cl = repo.changelog
1237 1237 for r in getset(repo, spanset(repo), x):
1238 1238 ps.update(cl.parentrevs(r))
1239 1239 return subset & ps
1240 1240
1241 1241 def parentspec(repo, subset, x, n):
1242 1242 """``set^0``
1243 1243 The set.
1244 1244 ``set^1`` (or ``set^``), ``set^2``
1245 1245 First or second parent, respectively, of all changesets in set.
1246 1246 """
1247 1247 try:
1248 1248 n = int(n[1])
1249 1249 if n not in (0, 1, 2):
1250 1250 raise ValueError
1251 1251 except (TypeError, ValueError):
1252 1252 raise error.ParseError(_("^ expects a number 0, 1, or 2"))
1253 1253 ps = set()
1254 1254 cl = repo.changelog
1255 1255 for r in getset(repo, baseset(cl), x):
1256 1256 if n == 0:
1257 1257 ps.add(r)
1258 1258 elif n == 1:
1259 1259 ps.add(cl.parentrevs(r)[0])
1260 1260 elif n == 2:
1261 1261 parents = cl.parentrevs(r)
1262 1262 if len(parents) > 1:
1263 1263 ps.add(parents[1])
1264 1264 return subset & ps
1265 1265
1266 1266 def present(repo, subset, x):
1267 1267 """``present(set)``
1268 1268 An empty set, if any revision in set isn't found; otherwise,
1269 1269 all revisions in set.
1270 1270
1271 1271 If any of specified revisions is not present in the local repository,
1272 1272 the query is normally aborted. But this predicate allows the query
1273 1273 to continue even in such cases.
1274 1274 """
1275 1275 try:
1276 1276 return getset(repo, subset, x)
1277 1277 except error.RepoLookupError:
1278 1278 return baseset([])
1279 1279
1280 1280 def public(repo, subset, x):
1281 1281 """``public()``
1282 1282 Changeset in public phase."""
1283 1283 # i18n: "public" is a keyword
1284 1284 getargs(x, 0, 0, _("public takes no arguments"))
1285 1285 pc = repo._phasecache
1286 1286 return subset.filter(lambda r: pc.phase(repo, r) == phases.public)
1287 1287
1288 1288 def remote(repo, subset, x):
1289 1289 """``remote([id [,path]])``
1290 1290 Local revision that corresponds to the given identifier in a
1291 1291 remote repository, if present. Here, the '.' identifier is a
1292 1292 synonym for the current local branch.
1293 1293 """
1294 1294
1295 1295 import hg # avoid start-up nasties
1296 1296 # i18n: "remote" is a keyword
1297 1297 l = getargs(x, 0, 2, _("remote takes one, two or no arguments"))
1298 1298
1299 1299 q = '.'
1300 1300 if len(l) > 0:
1301 1301 # i18n: "remote" is a keyword
1302 1302 q = getstring(l[0], _("remote requires a string id"))
1303 1303 if q == '.':
1304 1304 q = repo['.'].branch()
1305 1305
1306 1306 dest = ''
1307 1307 if len(l) > 1:
1308 1308 # i18n: "remote" is a keyword
1309 1309 dest = getstring(l[1], _("remote requires a repository path"))
1310 1310 dest = repo.ui.expandpath(dest or 'default')
1311 1311 dest, branches = hg.parseurl(dest)
1312 1312 revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
1313 1313 if revs:
1314 1314 revs = [repo.lookup(rev) for rev in revs]
1315 1315 other = hg.peer(repo, {}, dest)
1316 1316 n = other.lookup(q)
1317 1317 if n in repo:
1318 1318 r = repo[n].rev()
1319 1319 if r in subset:
1320 1320 return baseset([r])
1321 1321 return baseset([])
1322 1322
1323 1323 def removes(repo, subset, x):
1324 1324 """``removes(pattern)``
1325 1325 Changesets which remove files matching pattern.
1326 1326
1327 1327 The pattern without explicit kind like ``glob:`` is expected to be
1328 1328 relative to the current directory and match against a file or a
1329 1329 directory.
1330 1330 """
1331 1331 # i18n: "removes" is a keyword
1332 1332 pat = getstring(x, _("removes requires a pattern"))
1333 1333 return checkstatus(repo, subset, pat, 2)
1334 1334
1335 1335 def rev(repo, subset, x):
1336 1336 """``rev(number)``
1337 1337 Revision with the given numeric identifier.
1338 1338 """
1339 1339 # i18n: "rev" is a keyword
1340 1340 l = getargs(x, 1, 1, _("rev requires one argument"))
1341 1341 try:
1342 1342 # i18n: "rev" is a keyword
1343 1343 l = int(getstring(l[0], _("rev requires a number")))
1344 1344 except (TypeError, ValueError):
1345 1345 # i18n: "rev" is a keyword
1346 1346 raise error.ParseError(_("rev expects a number"))
1347 1347 return subset.filter(lambda r: r == l)
1348 1348
1349 1349 def matching(repo, subset, x):
1350 1350 """``matching(revision [, field])``
1351 1351 Changesets in which a given set of fields match the set of fields in the
1352 1352 selected revision or set.
1353 1353
1354 1354 To match more than one field pass the list of fields to match separated
1355 1355 by spaces (e.g. ``author description``).
1356 1356
1357 1357 Valid fields are most regular revision fields and some special fields.
1358 1358
1359 1359 Regular revision fields are ``description``, ``author``, ``branch``,
1360 1360 ``date``, ``files``, ``phase``, ``parents``, ``substate``, ``user``
1361 1361 and ``diff``.
1362 1362 Note that ``author`` and ``user`` are synonyms. ``diff`` refers to the
1363 1363 contents of the revision. Two revisions matching their ``diff`` will
1364 1364 also match their ``files``.
1365 1365
1366 1366 Special fields are ``summary`` and ``metadata``:
1367 1367 ``summary`` matches the first line of the description.
1368 1368 ``metadata`` is equivalent to matching ``description user date``
1369 1369 (i.e. it matches the main metadata fields).
1370 1370
1371 1371 ``metadata`` is the default field which is used when no fields are
1372 1372 specified. You can match more than one field at a time.
1373 1373 """
1374 1374 # i18n: "matching" is a keyword
1375 1375 l = getargs(x, 1, 2, _("matching takes 1 or 2 arguments"))
1376 1376
1377 1377 revs = getset(repo, baseset(repo.changelog), l[0])
1378 1378
1379 1379 fieldlist = ['metadata']
1380 1380 if len(l) > 1:
1381 1381 fieldlist = getstring(l[1],
1382 1382 # i18n: "matching" is a keyword
1383 1383 _("matching requires a string "
1384 1384 "as its second argument")).split()
1385 1385
1386 1386 # Make sure that there are no repeated fields,
1387 1387 # expand the 'special' 'metadata' field type
1388 1388 # and check the 'files' whenever we check the 'diff'
1389 1389 fields = []
1390 1390 for field in fieldlist:
1391 1391 if field == 'metadata':
1392 1392 fields += ['user', 'description', 'date']
1393 1393 elif field == 'diff':
1394 1394 # a revision matching the diff must also match the files
1395 1395 # since matching the diff is very costly, make sure to
1396 1396 # also match the files first
1397 1397 fields += ['files', 'diff']
1398 1398 else:
1399 1399 if field == 'author':
1400 1400 field = 'user'
1401 1401 fields.append(field)
1402 1402 fields = set(fields)
1403 1403 if 'summary' in fields and 'description' in fields:
1404 1404 # If a revision matches its description it also matches its summary
1405 1405 fields.discard('summary')
1406 1406
1407 1407 # We may want to match more than one field
1408 1408 # Not all fields take the same amount of time to be matched
1409 1409 # Sort the selected fields in order of increasing matching cost
1410 1410 fieldorder = ['phase', 'parents', 'user', 'date', 'branch', 'summary',
1411 1411 'files', 'description', 'substate', 'diff']
1412 1412 def fieldkeyfunc(f):
1413 1413 try:
1414 1414 return fieldorder.index(f)
1415 1415 except ValueError:
1416 1416 # assume an unknown field is very costly
1417 1417 return len(fieldorder)
1418 1418 fields = list(fields)
1419 1419 fields.sort(key=fieldkeyfunc)
1420 1420
1421 1421 # Each field will be matched with its own "getfield" function
1422 1422 # which will be added to the getfieldfuncs array of functions
1423 1423 getfieldfuncs = []
1424 1424 _funcs = {
1425 1425 'user': lambda r: repo[r].user(),
1426 1426 'branch': lambda r: repo[r].branch(),
1427 1427 'date': lambda r: repo[r].date(),
1428 1428 'description': lambda r: repo[r].description(),
1429 1429 'files': lambda r: repo[r].files(),
1430 1430 'parents': lambda r: repo[r].parents(),
1431 1431 'phase': lambda r: repo[r].phase(),
1432 1432 'substate': lambda r: repo[r].substate,
1433 1433 'summary': lambda r: repo[r].description().splitlines()[0],
1434 1434 'diff': lambda r: list(repo[r].diff(git=True),)
1435 1435 }
1436 1436 for info in fields:
1437 1437 getfield = _funcs.get(info, None)
1438 1438 if getfield is None:
1439 1439 raise error.ParseError(
1440 1440 # i18n: "matching" is a keyword
1441 1441 _("unexpected field name passed to matching: %s") % info)
1442 1442 getfieldfuncs.append(getfield)
1443 1443 # convert the getfield array of functions into a "getinfo" function
1444 1444 # which returns an array of field values (or a single value if there
1445 1445 # is only one field to match)
1446 1446 getinfo = lambda r: [f(r) for f in getfieldfuncs]
1447 1447
1448 1448 def matches(x):
1449 1449 for rev in revs:
1450 1450 target = getinfo(rev)
1451 1451 match = True
1452 1452 for n, f in enumerate(getfieldfuncs):
1453 1453 if target[n] != f(x):
1454 1454 match = False
1455 1455 if match:
1456 1456 return True
1457 1457 return False
1458 1458
1459 1459 return subset.filter(matches)
1460 1460
1461 1461 def reverse(repo, subset, x):
1462 1462 """``reverse(set)``
1463 1463 Reverse order of set.
1464 1464 """
1465 1465 l = getset(repo, subset, x)
1466 1466 l.reverse()
1467 1467 return l
1468 1468
1469 1469 def roots(repo, subset, x):
1470 1470 """``roots(set)``
1471 1471 Changesets in set with no parent changeset in set.
1472 1472 """
1473 1473 s = getset(repo, baseset(repo.changelog), x).set()
1474 1474 subset = baseset([r for r in subset if r in s])
1475 1475 cs = _children(repo, subset, s)
1476 1476 return subset - cs
1477 1477
1478 1478 def secret(repo, subset, x):
1479 1479 """``secret()``
1480 1480 Changeset in secret phase."""
1481 1481 # i18n: "secret" is a keyword
1482 1482 getargs(x, 0, 0, _("secret takes no arguments"))
1483 1483 pc = repo._phasecache
1484 1484 return subset.filter(lambda x: pc.phase(repo, x) == phases.secret)
1485 1485
1486 1486 def sort(repo, subset, x):
1487 1487 """``sort(set[, [-]key...])``
1488 1488 Sort set by keys. The default sort order is ascending, specify a key
1489 1489 as ``-key`` to sort in descending order.
1490 1490
1491 1491 The keys can be:
1492 1492
1493 1493 - ``rev`` for the revision number,
1494 1494 - ``branch`` for the branch name,
1495 1495 - ``desc`` for the commit message (description),
1496 1496 - ``user`` for user name (``author`` can be used as an alias),
1497 1497 - ``date`` for the commit date
1498 1498 """
1499 1499 # i18n: "sort" is a keyword
1500 1500 l = getargs(x, 1, 2, _("sort requires one or two arguments"))
1501 1501 keys = "rev"
1502 1502 if len(l) == 2:
1503 1503 # i18n: "sort" is a keyword
1504 1504 keys = getstring(l[1], _("sort spec must be a string"))
1505 1505
1506 1506 s = l[0]
1507 1507 keys = keys.split()
1508 1508 l = []
1509 1509 def invert(s):
1510 1510 return "".join(chr(255 - ord(c)) for c in s)
1511 1511 revs = getset(repo, subset, s)
1512 1512 if keys == ["rev"]:
1513 1513 revs.sort()
1514 1514 return revs
1515 1515 elif keys == ["-rev"]:
1516 1516 revs.sort(reverse=True)
1517 1517 return revs
1518 1518 for r in revs:
1519 1519 c = repo[r]
1520 1520 e = []
1521 1521 for k in keys:
1522 1522 if k == 'rev':
1523 1523 e.append(r)
1524 1524 elif k == '-rev':
1525 1525 e.append(-r)
1526 1526 elif k == 'branch':
1527 1527 e.append(c.branch())
1528 1528 elif k == '-branch':
1529 1529 e.append(invert(c.branch()))
1530 1530 elif k == 'desc':
1531 1531 e.append(c.description())
1532 1532 elif k == '-desc':
1533 1533 e.append(invert(c.description()))
1534 1534 elif k in 'user author':
1535 1535 e.append(c.user())
1536 1536 elif k in '-user -author':
1537 1537 e.append(invert(c.user()))
1538 1538 elif k == 'date':
1539 1539 e.append(c.date()[0])
1540 1540 elif k == '-date':
1541 1541 e.append(-c.date()[0])
1542 1542 else:
1543 1543 raise error.ParseError(_("unknown sort key %r") % k)
1544 1544 e.append(r)
1545 1545 l.append(e)
1546 1546 l.sort()
1547 1547 return baseset([e[-1] for e in l])
1548 1548
1549 1549 def _stringmatcher(pattern):
1550 1550 """
1551 1551 accepts a string, possibly starting with 're:' or 'literal:' prefix.
1552 1552 returns the matcher name, pattern, and matcher function.
1553 1553 missing or unknown prefixes are treated as literal matches.
1554 1554
1555 1555 helper for tests:
1556 1556 >>> def test(pattern, *tests):
1557 1557 ... kind, pattern, matcher = _stringmatcher(pattern)
1558 1558 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
1559 1559
1560 1560 exact matching (no prefix):
1561 1561 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
1562 1562 ('literal', 'abcdefg', [False, False, True])
1563 1563
1564 1564 regex matching ('re:' prefix)
1565 1565 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
1566 1566 ('re', 'a.+b', [False, False, True])
1567 1567
1568 1568 force exact matches ('literal:' prefix)
1569 1569 >>> test('literal:re:foobar', 'foobar', 're:foobar')
1570 1570 ('literal', 're:foobar', [False, True])
1571 1571
1572 1572 unknown prefixes are ignored and treated as literals
1573 1573 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
1574 1574 ('literal', 'foo:bar', [False, False, True])
1575 1575 """
1576 1576 if pattern.startswith('re:'):
1577 1577 pattern = pattern[3:]
1578 1578 try:
1579 1579 regex = re.compile(pattern)
1580 1580 except re.error, e:
1581 1581 raise error.ParseError(_('invalid regular expression: %s')
1582 1582 % e)
1583 1583 return 're', pattern, regex.search
1584 1584 elif pattern.startswith('literal:'):
1585 1585 pattern = pattern[8:]
1586 1586 return 'literal', pattern, pattern.__eq__
1587 1587
1588 1588 def _substringmatcher(pattern):
1589 1589 kind, pattern, matcher = _stringmatcher(pattern)
1590 1590 if kind == 'literal':
1591 1591 matcher = lambda s: pattern in s
1592 1592 return kind, pattern, matcher
1593 1593
1594 1594 def tag(repo, subset, x):
1595 1595 """``tag([name])``
1596 1596 The specified tag by name, or all tagged revisions if no name is given.
1597 1597
1598 1598 If `name` starts with `re:`, the remainder of the name is treated as
1599 1599 a regular expression. To match a tag that actually starts with `re:`,
1600 1600 use the prefix `literal:`.
1601 1601 """
1602 1602 # i18n: "tag" is a keyword
1603 1603 args = getargs(x, 0, 1, _("tag takes one or no arguments"))
1604 1604 cl = repo.changelog
1605 1605 if args:
1606 1606 pattern = getstring(args[0],
1607 1607 # i18n: "tag" is a keyword
1608 1608 _('the argument to tag must be a string'))
1609 1609 kind, pattern, matcher = _stringmatcher(pattern)
1610 1610 if kind == 'literal':
1611 1611 # avoid resolving all tags
1612 1612 tn = repo._tagscache.tags.get(pattern, None)
1613 1613 if tn is None:
1614 1614 raise util.Abort(_("tag '%s' does not exist") % pattern)
1615 1615 s = set([repo[tn].rev()])
1616 1616 else:
1617 1617 s = set([cl.rev(n) for t, n in repo.tagslist() if matcher(t)])
1618 1618 else:
1619 1619 s = set([cl.rev(n) for t, n in repo.tagslist() if t != 'tip'])
1620 1620 return subset & s
1621 1621
1622 1622 def tagged(repo, subset, x):
1623 1623 return tag(repo, subset, x)
1624 1624
1625 1625 def unstable(repo, subset, x):
1626 1626 """``unstable()``
1627 1627 Non-obsolete changesets with obsolete ancestors.
1628 1628 """
1629 1629 # i18n: "unstable" is a keyword
1630 1630 getargs(x, 0, 0, _("unstable takes no arguments"))
1631 1631 unstables = obsmod.getrevs(repo, 'unstable')
1632 1632 return subset & unstables
1633 1633
1634 1634
1635 1635 def user(repo, subset, x):
1636 1636 """``user(string)``
1637 1637 User name contains string. The match is case-insensitive.
1638 1638
1639 1639 If `string` starts with `re:`, the remainder of the string is treated as
1640 1640 a regular expression. To match a user that actually contains `re:`, use
1641 1641 the prefix `literal:`.
1642 1642 """
1643 1643 return author(repo, subset, x)
1644 1644
1645 1645 # for internal use
1646 1646 def _list(repo, subset, x):
1647 1647 s = getstring(x, "internal error")
1648 1648 if not s:
1649 1649 return baseset([])
1650 1650 ls = [repo[r].rev() for r in s.split('\0')]
1651 1651 s = subset.set()
1652 1652 return baseset([r for r in ls if r in s])
1653 1653
1654 1654 # for internal use
1655 1655 def _intlist(repo, subset, x):
1656 1656 s = getstring(x, "internal error")
1657 1657 if not s:
1658 1658 return baseset([])
1659 1659 ls = [int(r) for r in s.split('\0')]
1660 1660 s = subset.set()
1661 1661 return baseset([r for r in ls if r in s])
1662 1662
1663 1663 # for internal use
1664 1664 def _hexlist(repo, subset, x):
1665 1665 s = getstring(x, "internal error")
1666 1666 if not s:
1667 1667 return baseset([])
1668 1668 cl = repo.changelog
1669 1669 ls = [cl.rev(node.bin(r)) for r in s.split('\0')]
1670 1670 s = subset.set()
1671 1671 return baseset([r for r in ls if r in s])
1672 1672
1673 1673 symbols = {
1674 1674 "adds": adds,
1675 1675 "all": getall,
1676 1676 "ancestor": ancestor,
1677 1677 "ancestors": ancestors,
1678 1678 "_firstancestors": _firstancestors,
1679 1679 "author": author,
1680 1680 "only": only,
1681 1681 "bisect": bisect,
1682 1682 "bisected": bisected,
1683 1683 "bookmark": bookmark,
1684 1684 "branch": branch,
1685 1685 "branchpoint": branchpoint,
1686 1686 "bumped": bumped,
1687 1687 "bundle": bundle,
1688 1688 "children": children,
1689 1689 "closed": closed,
1690 1690 "contains": contains,
1691 1691 "converted": converted,
1692 1692 "date": date,
1693 1693 "desc": desc,
1694 1694 "descendants": descendants,
1695 1695 "_firstdescendants": _firstdescendants,
1696 1696 "destination": destination,
1697 1697 "divergent": divergent,
1698 1698 "draft": draft,
1699 1699 "extinct": extinct,
1700 1700 "extra": extra,
1701 1701 "file": hasfile,
1702 1702 "filelog": filelog,
1703 1703 "first": first,
1704 1704 "follow": follow,
1705 1705 "_followfirst": _followfirst,
1706 1706 "grep": grep,
1707 1707 "head": head,
1708 1708 "heads": heads,
1709 1709 "hidden": hidden,
1710 1710 "id": node_,
1711 1711 "keyword": keyword,
1712 1712 "last": last,
1713 1713 "limit": limit,
1714 1714 "_matchfiles": _matchfiles,
1715 1715 "max": maxrev,
1716 1716 "merge": merge,
1717 1717 "min": minrev,
1718 1718 "_missingancestors": _missingancestors,
1719 1719 "modifies": modifies,
1720 1720 "obsolete": obsolete,
1721 1721 "origin": origin,
1722 1722 "outgoing": outgoing,
1723 1723 "p1": p1,
1724 1724 "p2": p2,
1725 1725 "parents": parents,
1726 1726 "present": present,
1727 1727 "public": public,
1728 1728 "remote": remote,
1729 1729 "removes": removes,
1730 1730 "rev": rev,
1731 1731 "reverse": reverse,
1732 1732 "roots": roots,
1733 1733 "sort": sort,
1734 1734 "secret": secret,
1735 1735 "matching": matching,
1736 1736 "tag": tag,
1737 1737 "tagged": tagged,
1738 1738 "user": user,
1739 1739 "unstable": unstable,
1740 1740 "_list": _list,
1741 1741 "_intlist": _intlist,
1742 1742 "_hexlist": _hexlist,
1743 1743 }
1744 1744
1745 1745 # symbols which can't be used for a DoS attack for any given input
1746 1746 # (e.g. those which accept regexes as plain strings shouldn't be included)
1747 1747 # functions that just return a lot of changesets (like all) don't count here
1748 1748 safesymbols = set([
1749 1749 "adds",
1750 1750 "all",
1751 1751 "ancestor",
1752 1752 "ancestors",
1753 1753 "_firstancestors",
1754 1754 "author",
1755 1755 "bisect",
1756 1756 "bisected",
1757 1757 "bookmark",
1758 1758 "branch",
1759 1759 "branchpoint",
1760 1760 "bumped",
1761 1761 "bundle",
1762 1762 "children",
1763 1763 "closed",
1764 1764 "converted",
1765 1765 "date",
1766 1766 "desc",
1767 1767 "descendants",
1768 1768 "_firstdescendants",
1769 1769 "destination",
1770 1770 "divergent",
1771 1771 "draft",
1772 1772 "extinct",
1773 1773 "extra",
1774 1774 "file",
1775 1775 "filelog",
1776 1776 "first",
1777 1777 "follow",
1778 1778 "_followfirst",
1779 1779 "head",
1780 1780 "heads",
1781 1781 "hidden",
1782 1782 "id",
1783 1783 "keyword",
1784 1784 "last",
1785 1785 "limit",
1786 1786 "_matchfiles",
1787 1787 "max",
1788 1788 "merge",
1789 1789 "min",
1790 1790 "_missingancestors",
1791 1791 "modifies",
1792 1792 "obsolete",
1793 1793 "origin",
1794 1794 "outgoing",
1795 1795 "p1",
1796 1796 "p2",
1797 1797 "parents",
1798 1798 "present",
1799 1799 "public",
1800 1800 "remote",
1801 1801 "removes",
1802 1802 "rev",
1803 1803 "reverse",
1804 1804 "roots",
1805 1805 "sort",
1806 1806 "secret",
1807 1807 "matching",
1808 1808 "tag",
1809 1809 "tagged",
1810 1810 "user",
1811 1811 "unstable",
1812 1812 "_list",
1813 1813 "_intlist",
1814 1814 "_hexlist",
1815 1815 ])
1816 1816
1817 1817 methods = {
1818 1818 "range": rangeset,
1819 1819 "dagrange": dagrange,
1820 1820 "string": stringset,
1821 1821 "symbol": symbolset,
1822 1822 "and": andset,
1823 1823 "or": orset,
1824 1824 "not": notset,
1825 1825 "list": listset,
1826 1826 "func": func,
1827 1827 "ancestor": ancestorspec,
1828 1828 "parent": parentspec,
1829 1829 "parentpost": p1,
1830 1830 }
1831 1831
1832 1832 def optimize(x, small):
1833 1833 if x is None:
1834 1834 return 0, x
1835 1835
1836 1836 smallbonus = 1
1837 1837 if small:
1838 1838 smallbonus = .5
1839 1839
1840 1840 op = x[0]
1841 1841 if op == 'minus':
1842 1842 return optimize(('and', x[1], ('not', x[2])), small)
1843 1843 elif op == 'dagrangepre':
1844 1844 return optimize(('func', ('symbol', 'ancestors'), x[1]), small)
1845 1845 elif op == 'dagrangepost':
1846 1846 return optimize(('func', ('symbol', 'descendants'), x[1]), small)
1847 1847 elif op == 'rangepre':
1848 1848 return optimize(('range', ('string', '0'), x[1]), small)
1849 1849 elif op == 'rangepost':
1850 1850 return optimize(('range', x[1], ('string', 'tip')), small)
1851 1851 elif op == 'negate':
1852 1852 return optimize(('string',
1853 1853 '-' + getstring(x[1], _("can't negate that"))), small)
1854 1854 elif op in 'string symbol negate':
1855 1855 return smallbonus, x # single revisions are small
1856 1856 elif op == 'and':
1857 1857 wa, ta = optimize(x[1], True)
1858 1858 wb, tb = optimize(x[2], True)
1859 1859
1860 1860 # (::x and not ::y)/(not ::y and ::x) have a fast path
1861 1861 def ismissingancestors(revs, bases):
1862 1862 return (
1863 1863 revs[0] == 'func'
1864 1864 and getstring(revs[1], _('not a symbol')) == 'ancestors'
1865 1865 and bases[0] == 'not'
1866 1866 and bases[1][0] == 'func'
1867 1867 and getstring(bases[1][1], _('not a symbol')) == 'ancestors')
1868 1868
1869 1869 w = min(wa, wb)
1870 1870 if ismissingancestors(ta, tb):
1871 1871 return w, ('func', ('symbol', '_missingancestors'),
1872 1872 ('list', ta[2], tb[1][2]))
1873 1873 if ismissingancestors(tb, ta):
1874 1874 return w, ('func', ('symbol', '_missingancestors'),
1875 1875 ('list', tb[2], ta[1][2]))
1876 1876
1877 1877 if wa > wb:
1878 1878 return w, (op, tb, ta)
1879 1879 return w, (op, ta, tb)
1880 1880 elif op == 'or':
1881 1881 wa, ta = optimize(x[1], False)
1882 1882 wb, tb = optimize(x[2], False)
1883 1883 if wb < wa:
1884 1884 wb, wa = wa, wb
1885 1885 return max(wa, wb), (op, ta, tb)
1886 1886 elif op == 'not':
1887 1887 o = optimize(x[1], not small)
1888 1888 return o[0], (op, o[1])
1889 1889 elif op == 'parentpost':
1890 1890 o = optimize(x[1], small)
1891 1891 return o[0], (op, o[1])
1892 1892 elif op == 'group':
1893 1893 return optimize(x[1], small)
1894 1894 elif op in 'dagrange range list parent ancestorspec':
1895 1895 if op == 'parent':
1896 1896 # x^:y means (x^) : y, not x ^ (:y)
1897 1897 post = ('parentpost', x[1])
1898 1898 if x[2][0] == 'dagrangepre':
1899 1899 return optimize(('dagrange', post, x[2][1]), small)
1900 1900 elif x[2][0] == 'rangepre':
1901 1901 return optimize(('range', post, x[2][1]), small)
1902 1902
1903 1903 wa, ta = optimize(x[1], small)
1904 1904 wb, tb = optimize(x[2], small)
1905 1905 return wa + wb, (op, ta, tb)
1906 1906 elif op == 'func':
1907 1907 f = getstring(x[1], _("not a symbol"))
1908 1908 wa, ta = optimize(x[2], small)
1909 1909 if f in ("author branch closed date desc file grep keyword "
1910 1910 "outgoing user"):
1911 1911 w = 10 # slow
1912 1912 elif f in "modifies adds removes":
1913 1913 w = 30 # slower
1914 1914 elif f == "contains":
1915 1915 w = 100 # very slow
1916 1916 elif f == "ancestor":
1917 1917 w = 1 * smallbonus
1918 1918 elif f in "reverse limit first":
1919 1919 w = 0
1920 1920 elif f in "sort":
1921 1921 w = 10 # assume most sorts look at changelog
1922 1922 else:
1923 1923 w = 1
1924 1924 return w + wa, (op, x[1], ta)
1925 1925 return 1, x
1926 1926
1927 1927 _aliasarg = ('func', ('symbol', '_aliasarg'))
1928 1928 def _getaliasarg(tree):
1929 1929 """If tree matches ('func', ('symbol', '_aliasarg'), ('string', X))
1930 1930 return X, None otherwise.
1931 1931 """
1932 1932 if (len(tree) == 3 and tree[:2] == _aliasarg
1933 1933 and tree[2][0] == 'string'):
1934 1934 return tree[2][1]
1935 1935 return None
1936 1936
1937 1937 def _checkaliasarg(tree, known=None):
1938 1938 """Check tree contains no _aliasarg construct or only ones which
1939 1939 value is in known. Used to avoid alias placeholders injection.
1940 1940 """
1941 1941 if isinstance(tree, tuple):
1942 1942 arg = _getaliasarg(tree)
1943 1943 if arg is not None and (not known or arg not in known):
1944 1944 raise error.ParseError(_("not a function: %s") % '_aliasarg')
1945 1945 for t in tree:
1946 1946 _checkaliasarg(t, known)
1947 1947
1948 1948 class revsetalias(object):
1949 1949 funcre = re.compile('^([^(]+)\(([^)]+)\)$')
1950 1950 args = None
1951 1951
1952 1952 def __init__(self, name, value):
1953 1953 '''Aliases like:
1954 1954
1955 1955 h = heads(default)
1956 1956 b($1) = ancestors($1) - ancestors(default)
1957 1957 '''
1958 1958 m = self.funcre.search(name)
1959 1959 if m:
1960 1960 self.name = m.group(1)
1961 1961 self.tree = ('func', ('symbol', m.group(1)))
1962 1962 self.args = [x.strip() for x in m.group(2).split(',')]
1963 1963 for arg in self.args:
1964 1964 # _aliasarg() is an unknown symbol only used separate
1965 1965 # alias argument placeholders from regular strings.
1966 1966 value = value.replace(arg, '_aliasarg(%r)' % (arg,))
1967 1967 else:
1968 1968 self.name = name
1969 1969 self.tree = ('symbol', name)
1970 1970
1971 1971 self.replacement, pos = parse(value)
1972 1972 if pos != len(value):
1973 1973 raise error.ParseError(_('invalid token'), pos)
1974 1974 # Check for placeholder injection
1975 1975 _checkaliasarg(self.replacement, self.args)
1976 1976
1977 1977 def _getalias(aliases, tree):
1978 1978 """If tree looks like an unexpanded alias, return it. Return None
1979 1979 otherwise.
1980 1980 """
1981 1981 if isinstance(tree, tuple) and tree:
1982 1982 if tree[0] == 'symbol' and len(tree) == 2:
1983 1983 name = tree[1]
1984 1984 alias = aliases.get(name)
1985 1985 if alias and alias.args is None and alias.tree == tree:
1986 1986 return alias
1987 1987 if tree[0] == 'func' and len(tree) > 1:
1988 1988 if tree[1][0] == 'symbol' and len(tree[1]) == 2:
1989 1989 name = tree[1][1]
1990 1990 alias = aliases.get(name)
1991 1991 if alias and alias.args is not None and alias.tree == tree[:2]:
1992 1992 return alias
1993 1993 return None
1994 1994
1995 1995 def _expandargs(tree, args):
1996 1996 """Replace _aliasarg instances with the substitution value of the
1997 1997 same name in args, recursively.
1998 1998 """
1999 1999 if not tree or not isinstance(tree, tuple):
2000 2000 return tree
2001 2001 arg = _getaliasarg(tree)
2002 2002 if arg is not None:
2003 2003 return args[arg]
2004 2004 return tuple(_expandargs(t, args) for t in tree)
2005 2005
2006 2006 def _expandaliases(aliases, tree, expanding, cache):
2007 2007 """Expand aliases in tree, recursively.
2008 2008
2009 2009 'aliases' is a dictionary mapping user defined aliases to
2010 2010 revsetalias objects.
2011 2011 """
2012 2012 if not isinstance(tree, tuple):
2013 2013 # Do not expand raw strings
2014 2014 return tree
2015 2015 alias = _getalias(aliases, tree)
2016 2016 if alias is not None:
2017 2017 if alias in expanding:
2018 2018 raise error.ParseError(_('infinite expansion of revset alias "%s" '
2019 2019 'detected') % alias.name)
2020 2020 expanding.append(alias)
2021 2021 if alias.name not in cache:
2022 2022 cache[alias.name] = _expandaliases(aliases, alias.replacement,
2023 2023 expanding, cache)
2024 2024 result = cache[alias.name]
2025 2025 expanding.pop()
2026 2026 if alias.args is not None:
2027 2027 l = getlist(tree[2])
2028 2028 if len(l) != len(alias.args):
2029 2029 raise error.ParseError(
2030 2030 _('invalid number of arguments: %s') % len(l))
2031 2031 l = [_expandaliases(aliases, a, [], cache) for a in l]
2032 2032 result = _expandargs(result, dict(zip(alias.args, l)))
2033 2033 else:
2034 2034 result = tuple(_expandaliases(aliases, t, expanding, cache)
2035 2035 for t in tree)
2036 2036 return result
2037 2037
2038 2038 def findaliases(ui, tree):
2039 2039 _checkaliasarg(tree)
2040 2040 aliases = {}
2041 2041 for k, v in ui.configitems('revsetalias'):
2042 2042 alias = revsetalias(k, v)
2043 2043 aliases[alias.name] = alias
2044 2044 return _expandaliases(aliases, tree, [], {})
2045 2045
2046 2046 def parse(spec, lookup=None):
2047 2047 p = parser.parser(tokenize, elements)
2048 2048 return p.parse(spec, lookup=lookup)
2049 2049
2050 2050 def match(ui, spec, repo=None):
2051 2051 if not spec:
2052 2052 raise error.ParseError(_("empty query"))
2053 2053 lookup = None
2054 2054 if repo:
2055 2055 lookup = repo.__contains__
2056 2056 tree, pos = parse(spec, lookup)
2057 2057 if (pos != len(spec)):
2058 2058 raise error.ParseError(_("invalid token"), pos)
2059 2059 if ui:
2060 2060 tree = findaliases(ui, tree)
2061 2061 weight, tree = optimize(tree, True)
2062 2062 def mfunc(repo, subset):
2063 2063 if util.safehasattr(subset, 'set'):
2064 2064 return getset(repo, subset, tree)
2065 2065 return getset(repo, baseset(subset), tree)
2066 2066 return mfunc
2067 2067
2068 2068 def formatspec(expr, *args):
2069 2069 '''
2070 2070 This is a convenience function for using revsets internally, and
2071 2071 escapes arguments appropriately. Aliases are intentionally ignored
2072 2072 so that intended expression behavior isn't accidentally subverted.
2073 2073
2074 2074 Supported arguments:
2075 2075
2076 2076 %r = revset expression, parenthesized
2077 2077 %d = int(arg), no quoting
2078 2078 %s = string(arg), escaped and single-quoted
2079 2079 %b = arg.branch(), escaped and single-quoted
2080 2080 %n = hex(arg), single-quoted
2081 2081 %% = a literal '%'
2082 2082
2083 2083 Prefixing the type with 'l' specifies a parenthesized list of that type.
2084 2084
2085 2085 >>> formatspec('%r:: and %lr', '10 or 11', ("this()", "that()"))
2086 2086 '(10 or 11):: and ((this()) or (that()))'
2087 2087 >>> formatspec('%d:: and not %d::', 10, 20)
2088 2088 '10:: and not 20::'
2089 2089 >>> formatspec('%ld or %ld', [], [1])
2090 2090 "_list('') or 1"
2091 2091 >>> formatspec('keyword(%s)', 'foo\\xe9')
2092 2092 "keyword('foo\\\\xe9')"
2093 2093 >>> b = lambda: 'default'
2094 2094 >>> b.branch = b
2095 2095 >>> formatspec('branch(%b)', b)
2096 2096 "branch('default')"
2097 2097 >>> formatspec('root(%ls)', ['a', 'b', 'c', 'd'])
2098 2098 "root(_list('a\\x00b\\x00c\\x00d'))"
2099 2099 '''
2100 2100
2101 2101 def quote(s):
2102 2102 return repr(str(s))
2103 2103
2104 2104 def argtype(c, arg):
2105 2105 if c == 'd':
2106 2106 return str(int(arg))
2107 2107 elif c == 's':
2108 2108 return quote(arg)
2109 2109 elif c == 'r':
2110 2110 parse(arg) # make sure syntax errors are confined
2111 2111 return '(%s)' % arg
2112 2112 elif c == 'n':
2113 2113 return quote(node.hex(arg))
2114 2114 elif c == 'b':
2115 2115 return quote(arg.branch())
2116 2116
2117 2117 def listexp(s, t):
2118 2118 l = len(s)
2119 2119 if l == 0:
2120 2120 return "_list('')"
2121 2121 elif l == 1:
2122 2122 return argtype(t, s[0])
2123 2123 elif t == 'd':
2124 2124 return "_intlist('%s')" % "\0".join(str(int(a)) for a in s)
2125 2125 elif t == 's':
2126 2126 return "_list('%s')" % "\0".join(s)
2127 2127 elif t == 'n':
2128 2128 return "_hexlist('%s')" % "\0".join(node.hex(a) for a in s)
2129 2129 elif t == 'b':
2130 2130 return "_list('%s')" % "\0".join(a.branch() for a in s)
2131 2131
2132 2132 m = l // 2
2133 2133 return '(%s or %s)' % (listexp(s[:m], t), listexp(s[m:], t))
2134 2134
2135 2135 ret = ''
2136 2136 pos = 0
2137 2137 arg = 0
2138 2138 while pos < len(expr):
2139 2139 c = expr[pos]
2140 2140 if c == '%':
2141 2141 pos += 1
2142 2142 d = expr[pos]
2143 2143 if d == '%':
2144 2144 ret += d
2145 2145 elif d in 'dsnbr':
2146 2146 ret += argtype(d, args[arg])
2147 2147 arg += 1
2148 2148 elif d == 'l':
2149 2149 # a list of some type
2150 2150 pos += 1
2151 2151 d = expr[pos]
2152 2152 ret += listexp(list(args[arg]), d)
2153 2153 arg += 1
2154 2154 else:
2155 2155 raise util.Abort('unexpected revspec format character %s' % d)
2156 2156 else:
2157 2157 ret += c
2158 2158 pos += 1
2159 2159
2160 2160 return ret
2161 2161
2162 2162 def prettyformat(tree):
2163 2163 def _prettyformat(tree, level, lines):
2164 2164 if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
2165 2165 lines.append((level, str(tree)))
2166 2166 else:
2167 2167 lines.append((level, '(%s' % tree[0]))
2168 2168 for s in tree[1:]:
2169 2169 _prettyformat(s, level + 1, lines)
2170 2170 lines[-1:] = [(lines[-1][0], lines[-1][1] + ')')]
2171 2171
2172 2172 lines = []
2173 2173 _prettyformat(tree, 0, lines)
2174 2174 output = '\n'.join((' '*l + s) for l, s in lines)
2175 2175 return output
2176 2176
2177 2177 def depth(tree):
2178 2178 if isinstance(tree, tuple):
2179 2179 return max(map(depth, tree)) + 1
2180 2180 else:
2181 2181 return 0
2182 2182
2183 2183 def funcsused(tree):
2184 2184 if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
2185 2185 return set()
2186 2186 else:
2187 2187 funcs = set()
2188 2188 for s in tree[1:]:
2189 2189 funcs |= funcsused(s)
2190 2190 if tree[0] == 'func':
2191 2191 funcs.add(tree[1][1])
2192 2192 return funcs
2193 2193
2194 2194 class baseset(list):
2195 2195 """Basic data structure that represents a revset and contains the basic
2196 2196 operation that it should be able to perform.
2197 2197
2198 2198 Every method in this class should be implemented by any smartset class.
2199 2199 """
2200 2200 def __init__(self, data=()):
2201 2201 super(baseset, self).__init__(data)
2202 2202 self._set = None
2203 2203
2204 2204 def ascending(self):
2205 2205 """Sorts the set in ascending order (in place).
2206 2206
2207 2207 This is part of the mandatory API for smartset."""
2208 2208 self.sort()
2209 2209
2210 2210 def descending(self):
2211 2211 """Sorts the set in descending order (in place).
2212 2212
2213 2213 This is part of the mandatory API for smartset."""
2214 2214 self.sort(reverse=True)
2215 2215
2216 2216 def min(self):
2217 2217 return min(self)
2218 2218
2219 2219 def max(self):
2220 2220 return max(self)
2221 2221
2222 2222 def set(self):
2223 2223 """Returns a set or a smartset containing all the elements.
2224 2224
2225 2225 The returned structure should be the fastest option for membership
2226 2226 testing.
2227 2227
2228 2228 This is part of the mandatory API for smartset."""
2229 2229 if not self._set:
2230 2230 self._set = set(self)
2231 2231 return self._set
2232 2232
2233 2233 def __sub__(self, other):
2234 2234 """Returns a new object with the substraction of the two collections.
2235 2235
2236 2236 This is part of the mandatory API for smartset."""
2237 2237 if isinstance(other, baseset):
2238 2238 s = other.set()
2239 2239 else:
2240 2240 s = set(other)
2241 2241 return baseset(self.set() - s)
2242 2242
2243 2243 def __and__(self, other):
2244 2244 """Returns a new object with the intersection of the two collections.
2245 2245
2246 2246 This is part of the mandatory API for smartset."""
2247 2247 if isinstance(other, baseset):
2248 2248 other = other.set()
2249 2249 return baseset([y for y in self if y in other])
2250 2250
2251 2251 def __add__(self, other):
2252 2252 """Returns a new object with the union of the two collections.
2253 2253
2254 2254 This is part of the mandatory API for smartset."""
2255 2255 s = self.set()
2256 2256 l = [r for r in other if r not in s]
2257 2257 return baseset(list(self) + l)
2258 2258
2259 2259 def isascending(self):
2260 2260 """Returns True if the collection is ascending order, False if not.
2261 2261
2262 2262 This is part of the mandatory API for smartset."""
2263 2263 return False
2264 2264
2265 2265 def isdescending(self):
2266 2266 """Returns True if the collection is descending order, False if not.
2267 2267
2268 2268 This is part of the mandatory API for smartset."""
2269 2269 return False
2270 2270
2271 2271 def filter(self, condition):
2272 2272 """Returns this smartset filtered by condition as a new smartset.
2273 2273
2274 2274 `condition` is a callable which takes a revision number and returns a
2275 2275 boolean.
2276 2276
2277 2277 This is part of the mandatory API for smartset."""
2278 2278 return lazyset(self, condition)
2279 2279
2280 2280 class _orderedsetmixin(object):
2281 2281 """Mixin class with utility methods for smartsets
2282 2282
2283 2283 This should be extended by smartsets which have the isascending(),
2284 2284 isdescending() and reverse() methods"""
2285 2285
2286 2286 def _first(self):
2287 2287 """return the first revision in the set"""
2288 2288 for r in self:
2289 2289 return r
2290 2290 return None
2291 2291
2292 2292 def _last(self):
2293 2293 """return the last revision in the set"""
2294 2294 self.reverse()
2295 2295 m = self._first()
2296 2296 self.reverse()
2297 2297 return m
2298 2298
2299 2299 def min(self):
2300 2300 """return the smallest element in the set"""
2301 2301 if self.isascending():
2302 2302 return self._first()
2303 2303 return self._last()
2304 2304
2305 2305 def max(self):
2306 2306 """return the largest element in the set"""
2307 2307 if self.isascending():
2308 2308 return self._last()
2309 2309 return self._first()
2310 2310
2311 2311 class lazyset(object):
2312 2312 """Duck type for baseset class which iterates lazily over the revisions in
2313 2313 the subset and contains a function which tests for membership in the
2314 2314 revset
2315 2315 """
2316 2316 def __init__(self, subset, condition=lambda x: True):
2317 2317 """
2318 2318 condition: a function that decide whether a revision in the subset
2319 2319 belongs to the revset or not.
2320 2320 """
2321 2321 self._subset = subset
2322 2322 self._condition = condition
2323 2323 self._cache = {}
2324 2324
2325 2325 def ascending(self):
2326 2326 self._subset.sort()
2327 2327
2328 2328 def descending(self):
2329 2329 self._subset.sort(reverse=True)
2330 2330
2331 2331 def min(self):
2332 2332 return min(self)
2333 2333
2334 2334 def max(self):
2335 2335 return max(self)
2336 2336
2337 2337 def __contains__(self, x):
2338 2338 c = self._cache
2339 2339 if x not in c:
2340 2340 c[x] = x in self._subset and self._condition(x)
2341 2341 return c[x]
2342 2342
2343 2343 def __iter__(self):
2344 2344 cond = self._condition
2345 2345 for x in self._subset:
2346 2346 if cond(x):
2347 2347 yield x
2348 2348
2349 2349 def __and__(self, x):
2350 2350 return lazyset(self, lambda r: r in x)
2351 2351
2352 2352 def __sub__(self, x):
2353 2353 return lazyset(self, lambda r: r not in x)
2354 2354
2355 2355 def __add__(self, x):
2356 2356 return _addset(self, x)
2357 2357
2358 2358 def __nonzero__(self):
2359 2359 for r in self:
2360 2360 return True
2361 2361 return False
2362 2362
2363 2363 def __len__(self):
2364 2364 # Basic implementation to be changed in future patches.
2365 2365 l = baseset([r for r in self])
2366 2366 return len(l)
2367 2367
2368 2368 def __getitem__(self, x):
2369 2369 # Basic implementation to be changed in future patches.
2370 2370 l = baseset([r for r in self])
2371 2371 return l[x]
2372 2372
2373 2373 def sort(self, reverse=False):
2374 2374 if not util.safehasattr(self._subset, 'sort'):
2375 2375 self._subset = baseset(self._subset)
2376 2376 self._subset.sort(reverse=reverse)
2377 2377
2378 2378 def reverse(self):
2379 2379 self._subset.reverse()
2380 2380
2381 2381 def set(self):
2382 2382 return set([r for r in self])
2383 2383
2384 2384 def isascending(self):
2385 2385 return False
2386 2386
2387 2387 def isdescending(self):
2388 2388 return False
2389 2389
2390 2390 def filter(self, l):
2391 2391 return lazyset(self, l)
2392 2392
2393 2393 class orderedlazyset(_orderedsetmixin, lazyset):
2394 2394 """Subclass of lazyset which subset can be ordered either ascending or
2395 2395 descendingly
2396 2396 """
2397 2397 def __init__(self, subset, condition, ascending=True):
2398 2398 super(orderedlazyset, self).__init__(subset, condition)
2399 2399 self._ascending = ascending
2400 2400
2401 2401 def filter(self, l):
2402 2402 return orderedlazyset(self, l, ascending=self._ascending)
2403 2403
2404 2404 def ascending(self):
2405 2405 if not self._ascending:
2406 2406 self.reverse()
2407 2407
2408 2408 def descending(self):
2409 2409 if self._ascending:
2410 2410 self.reverse()
2411 2411
2412 2412 def __and__(self, x):
2413 2413 return orderedlazyset(self, lambda r: r in x,
2414 2414 ascending=self._ascending)
2415 2415
2416 2416 def __sub__(self, x):
2417 2417 return orderedlazyset(self, lambda r: r not in x,
2418 2418 ascending=self._ascending)
2419 2419
2420 2420 def __add__(self, x):
2421 2421 kwargs = {}
2422 2422 if self.isascending() and x.isascending():
2423 2423 kwargs['ascending'] = True
2424 2424 if self.isdescending() and x.isdescending():
2425 2425 kwargs['ascending'] = False
2426 2426 return _addset(self, x, **kwargs)
2427 2427
2428 2428 def sort(self, reverse=False):
2429 2429 if reverse:
2430 2430 if self._ascending:
2431 2431 self._subset.sort(reverse=reverse)
2432 2432 else:
2433 2433 if not self._ascending:
2434 2434 self._subset.sort(reverse=reverse)
2435 2435 self._ascending = not reverse
2436 2436
2437 2437 def isascending(self):
2438 2438 return self._ascending
2439 2439
2440 2440 def isdescending(self):
2441 2441 return not self._ascending
2442 2442
2443 2443 def reverse(self):
2444 2444 self._subset.reverse()
2445 2445 self._ascending = not self._ascending
2446 2446
2447 2447 class _addset(_orderedsetmixin):
2448 2448 """Represent the addition of two sets
2449 2449
2450 2450 Wrapper structure for lazily adding two structures without losing much
2451 2451 performance on the __contains__ method
2452 2452
2453 2453 If the ascending attribute is set, that means the two structures are
2454 2454 ordered in either an ascending or descending way. Therefore, we can add
2455 2455 them mantaining the order by iterating over both at the same time
2456 2456
2457 2457 This class does not duck-type baseset and it's only supposed to be used
2458 2458 internally
2459 2459 """
2460 2460 def __init__(self, revs1, revs2, ascending=None):
2461 2461 self._r1 = revs1
2462 2462 self._r2 = revs2
2463 2463 self._iter = None
2464 2464 self._ascending = ascending
2465 2465 self._genlist = None
2466 2466
2467 2467 @util.propertycache
2468 2468 def _list(self):
2469 2469 if not self._genlist:
2470 2470 self._genlist = baseset(self._iterator())
2471 2471 return self._genlist
2472 2472
2473 2473 def filter(self, condition):
2474 2474 if self._ascending is not None:
2475 2475 return orderedlazyset(self, condition, ascending=self._ascending)
2476 2476 return lazyset(self, condition)
2477 2477
2478 2478 def ascending(self):
2479 2479 if self._ascending is None:
2480 2480 self.sort()
2481 2481 self._ascending = True
2482 2482 else:
2483 2483 if not self._ascending:
2484 2484 self.reverse()
2485 2485
2486 2486 def descending(self):
2487 2487 if self._ascending is None:
2488 2488 self.sort(reverse=True)
2489 2489 self._ascending = False
2490 2490 else:
2491 2491 if self._ascending:
2492 2492 self.reverse()
2493 2493
2494 2494 def __and__(self, other):
2495 2495 filterfunc = other.__contains__
2496 2496 if self._ascending is not None:
2497 2497 return orderedlazyset(self, filterfunc, ascending=self._ascending)
2498 2498 return lazyset(self, filterfunc)
2499 2499
2500 2500 def __sub__(self, other):
2501 2501 filterfunc = lambda r: r not in other
2502 2502 if self._ascending is not None:
2503 2503 return orderedlazyset(self, filterfunc, ascending=self._ascending)
2504 2504 return lazyset(self, filterfunc)
2505 2505
2506 2506 def __add__(self, other):
2507 2507 """When both collections are ascending or descending, preserve the order
2508 2508 """
2509 2509 kwargs = {}
2510 2510 if self._ascending is not None:
2511 2511 if self.isascending() and other.isascending():
2512 2512 kwargs['ascending'] = True
2513 2513 if self.isdescending() and other.isdescending():
2514 2514 kwargs['ascending'] = False
2515 2515 return _addset(self, other, **kwargs)
2516 2516
2517 2517 def _iterator(self):
2518 2518 """Iterate over both collections without repeating elements
2519 2519
2520 2520 If the ascending attribute is not set, iterate over the first one and
2521 2521 then over the second one checking for membership on the first one so we
2522 2522 dont yield any duplicates.
2523 2523
2524 2524 If the ascending attribute is set, iterate over both collections at the
2525 2525 same time, yielding only one value at a time in the given order.
2526 2526 """
2527 2527 if not self._iter:
2528 2528 def gen():
2529 2529 if self._ascending is None:
2530 2530 for r in self._r1:
2531 2531 yield r
2532 2532 s = self._r1.set()
2533 2533 for r in self._r2:
2534 2534 if r not in s:
2535 2535 yield r
2536 2536 else:
2537 2537 iter1 = iter(self._r1)
2538 2538 iter2 = iter(self._r2)
2539 2539
2540 2540 val1 = None
2541 2541 val2 = None
2542 2542
2543 2543 choice = max
2544 2544 if self._ascending:
2545 2545 choice = min
2546 2546 try:
2547 2547 # Consume both iterators in an ordered way until one is
2548 2548 # empty
2549 2549 while True:
2550 2550 if val1 is None:
2551 2551 val1 = iter1.next()
2552 2552 if val2 is None:
2553 2553 val2 = iter2.next()
2554 2554 next = choice(val1, val2)
2555 2555 yield next
2556 2556 if val1 == next:
2557 2557 val1 = None
2558 2558 if val2 == next:
2559 2559 val2 = None
2560 2560 except StopIteration:
2561 2561 # Flush any remaining values and consume the other one
2562 2562 it = iter2
2563 2563 if val1 is not None:
2564 2564 yield val1
2565 2565 it = iter1
2566 2566 elif val2 is not None:
2567 2567 # might have been equality and both are empty
2568 2568 yield val2
2569 2569 for val in it:
2570 2570 yield val
2571 2571
2572 2572 self._iter = _generatorset(gen())
2573 2573
2574 2574 return self._iter
2575 2575
2576 2576 def __iter__(self):
2577 2577 if self._genlist:
2578 2578 return iter(self._genlist)
2579 2579 return iter(self._iterator())
2580 2580
2581 2581 def __contains__(self, x):
2582 2582 return x in self._r1 or x in self._r2
2583 2583
2584 2584 def set(self):
2585 2585 return self
2586 2586
2587 2587 def sort(self, reverse=False):
2588 2588 """Sort the added set
2589 2589
2590 2590 For this we use the cached list with all the generated values and if we
2591 2591 know they are ascending or descending we can sort them in a smart way.
2592 2592 """
2593 2593 if self._ascending is None:
2594 2594 self._list.sort(reverse=reverse)
2595 2595 self._ascending = not reverse
2596 2596 else:
2597 2597 if bool(self._ascending) == bool(reverse):
2598 2598 self.reverse()
2599 2599
2600 2600 def isascending(self):
2601 2601 return self._ascending is not None and self._ascending
2602 2602
2603 2603 def isdescending(self):
2604 2604 return self._ascending is not None and not self._ascending
2605 2605
2606 2606 def reverse(self):
2607 2607 self._list.reverse()
2608 2608 if self._ascending is not None:
2609 2609 self._ascending = not self._ascending
2610 2610
2611 2611 class _generatorset(object):
2612 2612 """Wrap a generator for lazy iteration
2613 2613
2614 2614 Wrapper structure for generators that provides lazy membership and can
2615 2615 be iterated more than once.
2616 2616 When asked for membership it generates values until either it finds the
2617 2617 requested one or has gone through all the elements in the generator
2618 2618
2619 2619 This class does not duck-type baseset and it's only supposed to be used
2620 2620 internally
2621 2621 """
2622 2622 def __init__(self, gen):
2623 2623 """
2624 2624 gen: a generator producing the values for the generatorset.
2625 2625 """
2626 2626 self._gen = gen
2627 self._iter = iter(gen)
2628 2627 self._cache = {}
2629 2628 self._genlist = baseset([])
2630 self._iterated = False
2631 2629 self._finished = False
2632 2630
2633 2631 def __contains__(self, x):
2634 2632 if x in self._cache:
2635 2633 return self._cache[x]
2636 2634
2637 2635 # Use new values only, as existing values would be cached.
2638 2636 for l in self._consumegen():
2639 2637 if l == x:
2640 2638 return True
2641 2639
2642 self._finished = True
2643 2640 self._cache[x] = False
2644 2641 return False
2645 2642
2646 2643 def __iter__(self):
2647 if self._iterated:
2648 # At least a part of the list should be cached if iteration has
2649 # started over the generatorset.
2650 for l in self._genlist:
2651 yield l
2652
2653 for item in self._consumegen():
2654 yield item
2644 if self._finished:
2645 for x in self._genlist:
2646 yield x
2647 return
2648
2649 i = 0
2650 genlist = self._genlist
2651 consume = self._consumegen()
2652 while True:
2653 if i < len(genlist):
2654 yield genlist[i]
2655 else:
2656 yield consume.next()
2657 i += 1
2655 2658
2656 2659 def _consumegen(self):
2657 self._iterated = True
2658
2659 2660 for item in self._gen:
2660 2661 self._cache[item] = True
2661 2662 self._genlist.append(item)
2662 2663 yield item
2663
2664 2664 self._finished = True
2665 2665
2666 2666 def set(self):
2667 2667 return self
2668 2668
2669 2669 def sort(self, reverse=False):
2670 2670 if not self._finished:
2671 2671 for i in self:
2672 2672 continue
2673 2673 self._genlist.sort(reverse=reverse)
2674 2674
2675 2675 class _ascgeneratorset(_generatorset):
2676 2676 """Wrap a generator of ascending elements for lazy iteration
2677 2677
2678 2678 Same structure as _generatorset but stops iterating after it goes past
2679 2679 the value when asked for membership and the element is not contained
2680 2680
2681 2681 This class does not duck-type baseset and it's only supposed to be used
2682 2682 internally
2683 2683 """
2684 2684 def __contains__(self, x):
2685 2685 if x in self._cache:
2686 2686 return self._cache[x]
2687 2687
2688 2688 # Use new values only, as existing values would be cached.
2689 2689 for l in self._consumegen():
2690 2690 if l == x:
2691 2691 return True
2692 2692 if l > x:
2693 2693 break
2694 2694
2695 2695 self._cache[x] = False
2696 2696 return False
2697 2697
2698 2698 class _descgeneratorset(_generatorset):
2699 2699 """Wrap a generator of descending elements for lazy iteration
2700 2700
2701 2701 Same structure as _generatorset but stops iterating after it goes past
2702 2702 the value when asked for membership and the element is not contained
2703 2703
2704 2704 This class does not duck-type baseset and it's only supposed to be used
2705 2705 internally
2706 2706 """
2707 2707 def __contains__(self, x):
2708 2708 if x in self._cache:
2709 2709 return self._cache[x]
2710 2710
2711 2711 # Use new values only, as existing values would be cached.
2712 2712 for l in self._consumegen():
2713 2713 if l == x:
2714 2714 return True
2715 2715 if l < x:
2716 2716 break
2717 2717
2718 2718 self._cache[x] = False
2719 2719 return False
2720 2720
2721 2721 class spanset(_orderedsetmixin):
2722 2722 """Duck type for baseset class which represents a range of revisions and
2723 2723 can work lazily and without having all the range in memory
2724 2724
2725 2725 Note that spanset(x, y) behave almost like xrange(x, y) except for two
2726 2726 notable points:
2727 2727 - when x < y it will be automatically descending,
2728 2728 - revision filtered with this repoview will be skipped.
2729 2729
2730 2730 """
2731 2731 def __init__(self, repo, start=0, end=None):
2732 2732 """
2733 2733 start: first revision included the set
2734 2734 (default to 0)
2735 2735 end: first revision excluded (last+1)
2736 2736 (default to len(repo)
2737 2737
2738 2738 Spanset will be descending if `end` < `start`.
2739 2739 """
2740 2740 self._start = start
2741 2741 if end is not None:
2742 2742 self._end = end
2743 2743 else:
2744 2744 self._end = len(repo)
2745 2745 self._hiddenrevs = repo.changelog.filteredrevs
2746 2746
2747 2747 def ascending(self):
2748 2748 if self._start > self._end:
2749 2749 self.reverse()
2750 2750
2751 2751 def descending(self):
2752 2752 if self._start < self._end:
2753 2753 self.reverse()
2754 2754
2755 2755 def _contained(self, rev):
2756 2756 return (rev <= self._start and rev > self._end) or (rev >= self._start
2757 2757 and rev < self._end)
2758 2758
2759 2759 def __iter__(self):
2760 2760 if self._start <= self._end:
2761 2761 iterrange = xrange(self._start, self._end)
2762 2762 else:
2763 2763 iterrange = xrange(self._start, self._end, -1)
2764 2764
2765 2765 if self._hiddenrevs:
2766 2766 s = self._hiddenrevs
2767 2767 for r in iterrange:
2768 2768 if r not in s:
2769 2769 yield r
2770 2770 else:
2771 2771 for r in iterrange:
2772 2772 yield r
2773 2773
2774 2774 def __contains__(self, x):
2775 2775 return self._contained(x) and not (self._hiddenrevs and rev in
2776 2776 self._hiddenrevs)
2777 2777
2778 2778 def __nonzero__(self):
2779 2779 for r in self:
2780 2780 return True
2781 2781 return False
2782 2782
2783 2783 def __and__(self, x):
2784 2784 if isinstance(x, baseset):
2785 2785 x = x.set()
2786 2786 if self._start <= self._end:
2787 2787 return orderedlazyset(self, lambda r: r in x)
2788 2788 else:
2789 2789 return orderedlazyset(self, lambda r: r in x, ascending=False)
2790 2790
2791 2791 def __sub__(self, x):
2792 2792 if isinstance(x, baseset):
2793 2793 x = x.set()
2794 2794 if self._start <= self._end:
2795 2795 return orderedlazyset(self, lambda r: r not in x)
2796 2796 else:
2797 2797 return orderedlazyset(self, lambda r: r not in x, ascending=False)
2798 2798
2799 2799 def __add__(self, x):
2800 2800 kwargs = {}
2801 2801 if self.isascending() and x.isascending():
2802 2802 kwargs['ascending'] = True
2803 2803 if self.isdescending() and x.isdescending():
2804 2804 kwargs['ascending'] = False
2805 2805 return _addset(self, x, **kwargs)
2806 2806
2807 2807 def __len__(self):
2808 2808 if not self._hiddenrevs:
2809 2809 return abs(self._end - self._start)
2810 2810 else:
2811 2811 count = 0
2812 2812 for rev in self._hiddenrevs:
2813 2813 if self._contained(rev):
2814 2814 count += 1
2815 2815 return abs(self._end - self._start) - count
2816 2816
2817 2817 def __getitem__(self, x):
2818 2818 # Basic implementation to be changed in future patches.
2819 2819 l = baseset([r for r in self])
2820 2820 return l[x]
2821 2821
2822 2822 def sort(self, reverse=False):
2823 2823 if bool(reverse) != (self._start > self._end):
2824 2824 self.reverse()
2825 2825
2826 2826 def reverse(self):
2827 2827 # Just switch the _start and _end parameters
2828 2828 if self._start <= self._end:
2829 2829 self._start, self._end = self._end - 1, self._start - 1
2830 2830 else:
2831 2831 self._start, self._end = self._end + 1, self._start + 1
2832 2832
2833 2833 def set(self):
2834 2834 return self
2835 2835
2836 2836 def isascending(self):
2837 2837 return self._start < self._end
2838 2838
2839 2839 def isdescending(self):
2840 2840 return self._start > self._end
2841 2841
2842 2842 def filter(self, l):
2843 2843 if self._start <= self._end:
2844 2844 return orderedlazyset(self, l)
2845 2845 else:
2846 2846 return orderedlazyset(self, l, ascending=False)
2847 2847
2848 2848 # tell hggettext to extract docstrings from these functions:
2849 2849 i18nfunctions = symbols.values()
General Comments 0
You need to be logged in to leave comments. Login now