##// END OF EJS Templates
revset: drop pre-lazyset optimization for stringset of subset == entire repo...
Yuya Nishihara -
r23724:aafeaba2 stable
parent child Browse files
Show More
@@ -1,2972 +1,2972 b''
1 1 # revset.py - revision set queries for mercurial
2 2 #
3 3 # Copyright 2010 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 import re
9 9 import parser, util, error, discovery, hbisect, phases
10 10 import node
11 11 import heapq
12 12 import match as matchmod
13 13 import ancestor as ancestormod
14 14 from i18n import _
15 15 import encoding
16 16 import obsolete as obsmod
17 17 import pathutil
18 18 import repoview
19 19
20 20 def _revancestors(repo, revs, followfirst):
21 21 """Like revlog.ancestors(), but supports followfirst."""
22 22 cut = followfirst and 1 or None
23 23 cl = repo.changelog
24 24
25 25 def iterate():
26 26 revqueue, revsnode = None, None
27 27 h = []
28 28
29 29 revs.sort(reverse=True)
30 30 revqueue = util.deque(revs)
31 31 if revqueue:
32 32 revsnode = revqueue.popleft()
33 33 heapq.heappush(h, -revsnode)
34 34
35 35 seen = set([node.nullrev])
36 36 while h:
37 37 current = -heapq.heappop(h)
38 38 if current not in seen:
39 39 if revsnode and current == revsnode:
40 40 if revqueue:
41 41 revsnode = revqueue.popleft()
42 42 heapq.heappush(h, -revsnode)
43 43 seen.add(current)
44 44 yield current
45 45 for parent in cl.parentrevs(current)[:cut]:
46 46 if parent != node.nullrev:
47 47 heapq.heappush(h, -parent)
48 48
49 49 return generatorset(iterate(), iterasc=False)
50 50
51 51 def _revdescendants(repo, revs, followfirst):
52 52 """Like revlog.descendants() but supports followfirst."""
53 53 cut = followfirst and 1 or None
54 54
55 55 def iterate():
56 56 cl = repo.changelog
57 57 first = min(revs)
58 58 nullrev = node.nullrev
59 59 if first == nullrev:
60 60 # Are there nodes with a null first parent and a non-null
61 61 # second one? Maybe. Do we care? Probably not.
62 62 for i in cl:
63 63 yield i
64 64 else:
65 65 seen = set(revs)
66 66 for i in cl.revs(first + 1):
67 67 for x in cl.parentrevs(i)[:cut]:
68 68 if x != nullrev and x in seen:
69 69 seen.add(i)
70 70 yield i
71 71 break
72 72
73 73 return generatorset(iterate(), iterasc=True)
74 74
75 75 def _revsbetween(repo, roots, heads):
76 76 """Return all paths between roots and heads, inclusive of both endpoint
77 77 sets."""
78 78 if not roots:
79 79 return baseset()
80 80 parentrevs = repo.changelog.parentrevs
81 81 visit = list(heads)
82 82 reachable = set()
83 83 seen = {}
84 84 minroot = min(roots)
85 85 roots = set(roots)
86 86 # open-code the post-order traversal due to the tiny size of
87 87 # sys.getrecursionlimit()
88 88 while visit:
89 89 rev = visit.pop()
90 90 if rev in roots:
91 91 reachable.add(rev)
92 92 parents = parentrevs(rev)
93 93 seen[rev] = parents
94 94 for parent in parents:
95 95 if parent >= minroot and parent not in seen:
96 96 visit.append(parent)
97 97 if not reachable:
98 98 return baseset()
99 99 for rev in sorted(seen):
100 100 for parent in seen[rev]:
101 101 if parent in reachable:
102 102 reachable.add(rev)
103 103 return baseset(sorted(reachable))
104 104
105 105 elements = {
106 106 "(": (20, ("group", 1, ")"), ("func", 1, ")")),
107 107 "~": (18, None, ("ancestor", 18)),
108 108 "^": (18, None, ("parent", 18), ("parentpost", 18)),
109 109 "-": (5, ("negate", 19), ("minus", 5)),
110 110 "::": (17, ("dagrangepre", 17), ("dagrange", 17),
111 111 ("dagrangepost", 17)),
112 112 "..": (17, ("dagrangepre", 17), ("dagrange", 17),
113 113 ("dagrangepost", 17)),
114 114 ":": (15, ("rangepre", 15), ("range", 15), ("rangepost", 15)),
115 115 "not": (10, ("not", 10)),
116 116 "!": (10, ("not", 10)),
117 117 "and": (5, None, ("and", 5)),
118 118 "&": (5, None, ("and", 5)),
119 119 "or": (4, None, ("or", 4)),
120 120 "|": (4, None, ("or", 4)),
121 121 "+": (4, None, ("or", 4)),
122 122 ",": (2, None, ("list", 2)),
123 123 ")": (0, None, None),
124 124 "symbol": (0, ("symbol",), None),
125 125 "string": (0, ("string",), None),
126 126 "end": (0, None, None),
127 127 }
128 128
129 129 keywords = set(['and', 'or', 'not'])
130 130
131 131 def tokenize(program, lookup=None):
132 132 '''
133 133 Parse a revset statement into a stream of tokens
134 134
135 135 Check that @ is a valid unquoted token character (issue3686):
136 136 >>> list(tokenize("@::"))
137 137 [('symbol', '@', 0), ('::', None, 1), ('end', None, 3)]
138 138
139 139 '''
140 140
141 141 pos, l = 0, len(program)
142 142 while pos < l:
143 143 c = program[pos]
144 144 if c.isspace(): # skip inter-token whitespace
145 145 pass
146 146 elif c == ':' and program[pos:pos + 2] == '::': # look ahead carefully
147 147 yield ('::', None, pos)
148 148 pos += 1 # skip ahead
149 149 elif c == '.' and program[pos:pos + 2] == '..': # look ahead carefully
150 150 yield ('..', None, pos)
151 151 pos += 1 # skip ahead
152 152 elif c in "():,-|&+!~^": # handle simple operators
153 153 yield (c, None, pos)
154 154 elif (c in '"\'' or c == 'r' and
155 155 program[pos:pos + 2] in ("r'", 'r"')): # handle quoted strings
156 156 if c == 'r':
157 157 pos += 1
158 158 c = program[pos]
159 159 decode = lambda x: x
160 160 else:
161 161 decode = lambda x: x.decode('string-escape')
162 162 pos += 1
163 163 s = pos
164 164 while pos < l: # find closing quote
165 165 d = program[pos]
166 166 if d == '\\': # skip over escaped characters
167 167 pos += 2
168 168 continue
169 169 if d == c:
170 170 yield ('string', decode(program[s:pos]), s)
171 171 break
172 172 pos += 1
173 173 else:
174 174 raise error.ParseError(_("unterminated string"), s)
175 175 # gather up a symbol/keyword
176 176 elif c.isalnum() or c in '._@' or ord(c) > 127:
177 177 s = pos
178 178 pos += 1
179 179 while pos < l: # find end of symbol
180 180 d = program[pos]
181 181 if not (d.isalnum() or d in "-._/@" or ord(d) > 127):
182 182 break
183 183 if d == '.' and program[pos - 1] == '.': # special case for ..
184 184 pos -= 1
185 185 break
186 186 pos += 1
187 187 sym = program[s:pos]
188 188 if sym in keywords: # operator keywords
189 189 yield (sym, None, s)
190 190 elif '-' in sym:
191 191 # some jerk gave us foo-bar-baz, try to check if it's a symbol
192 192 if lookup and lookup(sym):
193 193 # looks like a real symbol
194 194 yield ('symbol', sym, s)
195 195 else:
196 196 # looks like an expression
197 197 parts = sym.split('-')
198 198 for p in parts[:-1]:
199 199 if p: # possible consecutive -
200 200 yield ('symbol', p, s)
201 201 s += len(p)
202 202 yield ('-', None, pos)
203 203 s += 1
204 204 if parts[-1]: # possible trailing -
205 205 yield ('symbol', parts[-1], s)
206 206 else:
207 207 yield ('symbol', sym, s)
208 208 pos -= 1
209 209 else:
210 210 raise error.ParseError(_("syntax error"), pos)
211 211 pos += 1
212 212 yield ('end', None, pos)
213 213
214 214 # helpers
215 215
216 216 def getstring(x, err):
217 217 if x and (x[0] == 'string' or x[0] == 'symbol'):
218 218 return x[1]
219 219 raise error.ParseError(err)
220 220
221 221 def getlist(x):
222 222 if not x:
223 223 return []
224 224 if x[0] == 'list':
225 225 return getlist(x[1]) + [x[2]]
226 226 return [x]
227 227
228 228 def getargs(x, min, max, err):
229 229 l = getlist(x)
230 230 if len(l) < min or (max >= 0 and len(l) > max):
231 231 raise error.ParseError(err)
232 232 return l
233 233
234 234 def getset(repo, subset, x):
235 235 if not x:
236 236 raise error.ParseError(_("missing argument"))
237 237 s = methods[x[0]](repo, subset, *x[1:])
238 238 if util.safehasattr(s, 'isascending'):
239 239 return s
240 240 return baseset(s)
241 241
242 242 def _getrevsource(repo, r):
243 243 extra = repo[r].extra()
244 244 for label in ('source', 'transplant_source', 'rebase_source'):
245 245 if label in extra:
246 246 try:
247 247 return repo[extra[label]].rev()
248 248 except error.RepoLookupError:
249 249 pass
250 250 return None
251 251
252 252 # operator methods
253 253
254 254 def stringset(repo, subset, x):
255 255 x = repo[x].rev()
256 256 if x == -1 and len(subset) == len(repo):
257 257 return baseset([-1])
258 if len(subset) == len(repo) or x in subset:
258 if x in subset:
259 259 return baseset([x])
260 260 return baseset()
261 261
262 262 def symbolset(repo, subset, x):
263 263 if x in symbols:
264 264 raise error.ParseError(_("can't use %s here") % x)
265 265 return stringset(repo, subset, x)
266 266
267 267 def rangeset(repo, subset, x, y):
268 268 cl = baseset(repo.changelog)
269 269 m = getset(repo, cl, x)
270 270 n = getset(repo, cl, y)
271 271
272 272 if not m or not n:
273 273 return baseset()
274 274 m, n = m.first(), n.last()
275 275
276 276 if m < n:
277 277 r = spanset(repo, m, n + 1)
278 278 else:
279 279 r = spanset(repo, m, n - 1)
280 280 return r & subset
281 281
282 282 def dagrange(repo, subset, x, y):
283 283 r = spanset(repo)
284 284 xs = _revsbetween(repo, getset(repo, r, x), getset(repo, r, y))
285 285 return xs & subset
286 286
287 287 def andset(repo, subset, x, y):
288 288 return getset(repo, getset(repo, subset, x), y)
289 289
290 290 def orset(repo, subset, x, y):
291 291 xl = getset(repo, subset, x)
292 292 yl = getset(repo, subset - xl, y)
293 293 return xl + yl
294 294
295 295 def notset(repo, subset, x):
296 296 return subset - getset(repo, subset, x)
297 297
298 298 def listset(repo, subset, a, b):
299 299 raise error.ParseError(_("can't use a list in this context"))
300 300
301 301 def func(repo, subset, a, b):
302 302 if a[0] == 'symbol' and a[1] in symbols:
303 303 return symbols[a[1]](repo, subset, b)
304 304 raise error.ParseError(_("not a function: %s") % a[1])
305 305
306 306 # functions
307 307
308 308 def adds(repo, subset, x):
309 309 """``adds(pattern)``
310 310 Changesets that add a file matching pattern.
311 311
312 312 The pattern without explicit kind like ``glob:`` is expected to be
313 313 relative to the current directory and match against a file or a
314 314 directory.
315 315 """
316 316 # i18n: "adds" is a keyword
317 317 pat = getstring(x, _("adds requires a pattern"))
318 318 return checkstatus(repo, subset, pat, 1)
319 319
320 320 def ancestor(repo, subset, x):
321 321 """``ancestor(*changeset)``
322 322 A greatest common ancestor of the changesets.
323 323
324 324 Accepts 0 or more changesets.
325 325 Will return empty list when passed no args.
326 326 Greatest common ancestor of a single changeset is that changeset.
327 327 """
328 328 # i18n: "ancestor" is a keyword
329 329 l = getlist(x)
330 330 rl = spanset(repo)
331 331 anc = None
332 332
333 333 # (getset(repo, rl, i) for i in l) generates a list of lists
334 334 for revs in (getset(repo, rl, i) for i in l):
335 335 for r in revs:
336 336 if anc is None:
337 337 anc = repo[r]
338 338 else:
339 339 anc = anc.ancestor(repo[r])
340 340
341 341 if anc is not None and anc.rev() in subset:
342 342 return baseset([anc.rev()])
343 343 return baseset()
344 344
345 345 def _ancestors(repo, subset, x, followfirst=False):
346 346 heads = getset(repo, spanset(repo), x)
347 347 if not heads:
348 348 return baseset()
349 349 s = _revancestors(repo, heads, followfirst)
350 350 return subset & s
351 351
352 352 def ancestors(repo, subset, x):
353 353 """``ancestors(set)``
354 354 Changesets that are ancestors of a changeset in set.
355 355 """
356 356 return _ancestors(repo, subset, x)
357 357
358 358 def _firstancestors(repo, subset, x):
359 359 # ``_firstancestors(set)``
360 360 # Like ``ancestors(set)`` but follows only the first parents.
361 361 return _ancestors(repo, subset, x, followfirst=True)
362 362
363 363 def ancestorspec(repo, subset, x, n):
364 364 """``set~n``
365 365 Changesets that are the Nth ancestor (first parents only) of a changeset
366 366 in set.
367 367 """
368 368 try:
369 369 n = int(n[1])
370 370 except (TypeError, ValueError):
371 371 raise error.ParseError(_("~ expects a number"))
372 372 ps = set()
373 373 cl = repo.changelog
374 374 for r in getset(repo, baseset(cl), x):
375 375 for i in range(n):
376 376 r = cl.parentrevs(r)[0]
377 377 ps.add(r)
378 378 return subset & ps
379 379
380 380 def author(repo, subset, x):
381 381 """``author(string)``
382 382 Alias for ``user(string)``.
383 383 """
384 384 # i18n: "author" is a keyword
385 385 n = encoding.lower(getstring(x, _("author requires a string")))
386 386 kind, pattern, matcher = _substringmatcher(n)
387 387 return subset.filter(lambda x: matcher(encoding.lower(repo[x].user())))
388 388
389 389 def only(repo, subset, x):
390 390 """``only(set, [set])``
391 391 Changesets that are ancestors of the first set that are not ancestors
392 392 of any other head in the repo. If a second set is specified, the result
393 393 is ancestors of the first set that are not ancestors of the second set
394 394 (i.e. ::<set1> - ::<set2>).
395 395 """
396 396 cl = repo.changelog
397 397 # i18n: "only" is a keyword
398 398 args = getargs(x, 1, 2, _('only takes one or two arguments'))
399 399 include = getset(repo, spanset(repo), args[0])
400 400 if len(args) == 1:
401 401 if not include:
402 402 return baseset()
403 403
404 404 descendants = set(_revdescendants(repo, include, False))
405 405 exclude = [rev for rev in cl.headrevs()
406 406 if not rev in descendants and not rev in include]
407 407 else:
408 408 exclude = getset(repo, spanset(repo), args[1])
409 409
410 410 results = set(ancestormod.missingancestors(include, exclude, cl.parentrevs))
411 411 return subset & results
412 412
413 413 def bisect(repo, subset, x):
414 414 """``bisect(string)``
415 415 Changesets marked in the specified bisect status:
416 416
417 417 - ``good``, ``bad``, ``skip``: csets explicitly marked as good/bad/skip
418 418 - ``goods``, ``bads`` : csets topologically good/bad
419 419 - ``range`` : csets taking part in the bisection
420 420 - ``pruned`` : csets that are goods, bads or skipped
421 421 - ``untested`` : csets whose fate is yet unknown
422 422 - ``ignored`` : csets ignored due to DAG topology
423 423 - ``current`` : the cset currently being bisected
424 424 """
425 425 # i18n: "bisect" is a keyword
426 426 status = getstring(x, _("bisect requires a string")).lower()
427 427 state = set(hbisect.get(repo, status))
428 428 return subset & state
429 429
430 430 # Backward-compatibility
431 431 # - no help entry so that we do not advertise it any more
432 432 def bisected(repo, subset, x):
433 433 return bisect(repo, subset, x)
434 434
435 435 def bookmark(repo, subset, x):
436 436 """``bookmark([name])``
437 437 The named bookmark or all bookmarks.
438 438
439 439 If `name` starts with `re:`, the remainder of the name is treated as
440 440 a regular expression. To match a bookmark that actually starts with `re:`,
441 441 use the prefix `literal:`.
442 442 """
443 443 # i18n: "bookmark" is a keyword
444 444 args = getargs(x, 0, 1, _('bookmark takes one or no arguments'))
445 445 if args:
446 446 bm = getstring(args[0],
447 447 # i18n: "bookmark" is a keyword
448 448 _('the argument to bookmark must be a string'))
449 449 kind, pattern, matcher = _stringmatcher(bm)
450 450 bms = set()
451 451 if kind == 'literal':
452 452 bmrev = repo._bookmarks.get(pattern, None)
453 453 if not bmrev:
454 454 raise util.Abort(_("bookmark '%s' does not exist") % bm)
455 455 bms.add(repo[bmrev].rev())
456 456 else:
457 457 matchrevs = set()
458 458 for name, bmrev in repo._bookmarks.iteritems():
459 459 if matcher(name):
460 460 matchrevs.add(bmrev)
461 461 if not matchrevs:
462 462 raise util.Abort(_("no bookmarks exist that match '%s'")
463 463 % pattern)
464 464 for bmrev in matchrevs:
465 465 bms.add(repo[bmrev].rev())
466 466 else:
467 467 bms = set([repo[r].rev()
468 468 for r in repo._bookmarks.values()])
469 469 bms -= set([node.nullrev])
470 470 return subset & bms
471 471
472 472 def branch(repo, subset, x):
473 473 """``branch(string or set)``
474 474 All changesets belonging to the given branch or the branches of the given
475 475 changesets.
476 476
477 477 If `string` starts with `re:`, the remainder of the name is treated as
478 478 a regular expression. To match a branch that actually starts with `re:`,
479 479 use the prefix `literal:`.
480 480 """
481 481 try:
482 482 b = getstring(x, '')
483 483 except error.ParseError:
484 484 # not a string, but another revspec, e.g. tip()
485 485 pass
486 486 else:
487 487 kind, pattern, matcher = _stringmatcher(b)
488 488 if kind == 'literal':
489 489 # note: falls through to the revspec case if no branch with
490 490 # this name exists
491 491 if pattern in repo.branchmap():
492 492 return subset.filter(lambda r: matcher(repo[r].branch()))
493 493 else:
494 494 return subset.filter(lambda r: matcher(repo[r].branch()))
495 495
496 496 s = getset(repo, spanset(repo), x)
497 497 b = set()
498 498 for r in s:
499 499 b.add(repo[r].branch())
500 500 c = s.__contains__
501 501 return subset.filter(lambda r: c(r) or repo[r].branch() in b)
502 502
503 503 def bumped(repo, subset, x):
504 504 """``bumped()``
505 505 Mutable changesets marked as successors of public changesets.
506 506
507 507 Only non-public and non-obsolete changesets can be `bumped`.
508 508 """
509 509 # i18n: "bumped" is a keyword
510 510 getargs(x, 0, 0, _("bumped takes no arguments"))
511 511 bumped = obsmod.getrevs(repo, 'bumped')
512 512 return subset & bumped
513 513
514 514 def bundle(repo, subset, x):
515 515 """``bundle()``
516 516 Changesets in the bundle.
517 517
518 518 Bundle must be specified by the -R option."""
519 519
520 520 try:
521 521 bundlerevs = repo.changelog.bundlerevs
522 522 except AttributeError:
523 523 raise util.Abort(_("no bundle provided - specify with -R"))
524 524 return subset & bundlerevs
525 525
526 526 def checkstatus(repo, subset, pat, field):
527 527 hasset = matchmod.patkind(pat) == 'set'
528 528
529 529 mcache = [None]
530 530 def matches(x):
531 531 c = repo[x]
532 532 if not mcache[0] or hasset:
533 533 mcache[0] = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
534 534 m = mcache[0]
535 535 fname = None
536 536 if not m.anypats() and len(m.files()) == 1:
537 537 fname = m.files()[0]
538 538 if fname is not None:
539 539 if fname not in c.files():
540 540 return False
541 541 else:
542 542 for f in c.files():
543 543 if m(f):
544 544 break
545 545 else:
546 546 return False
547 547 files = repo.status(c.p1().node(), c.node())[field]
548 548 if fname is not None:
549 549 if fname in files:
550 550 return True
551 551 else:
552 552 for f in files:
553 553 if m(f):
554 554 return True
555 555
556 556 return subset.filter(matches)
557 557
558 558 def _children(repo, narrow, parentset):
559 559 cs = set()
560 560 if not parentset:
561 561 return baseset(cs)
562 562 pr = repo.changelog.parentrevs
563 563 minrev = min(parentset)
564 564 for r in narrow:
565 565 if r <= minrev:
566 566 continue
567 567 for p in pr(r):
568 568 if p in parentset:
569 569 cs.add(r)
570 570 return baseset(cs)
571 571
572 572 def children(repo, subset, x):
573 573 """``children(set)``
574 574 Child changesets of changesets in set.
575 575 """
576 576 s = getset(repo, baseset(repo), x)
577 577 cs = _children(repo, subset, s)
578 578 return subset & cs
579 579
580 580 def closed(repo, subset, x):
581 581 """``closed()``
582 582 Changeset is closed.
583 583 """
584 584 # i18n: "closed" is a keyword
585 585 getargs(x, 0, 0, _("closed takes no arguments"))
586 586 return subset.filter(lambda r: repo[r].closesbranch())
587 587
588 588 def contains(repo, subset, x):
589 589 """``contains(pattern)``
590 590 The revision's manifest contains a file matching pattern (but might not
591 591 modify it). See :hg:`help patterns` for information about file patterns.
592 592
593 593 The pattern without explicit kind like ``glob:`` is expected to be
594 594 relative to the current directory and match against a file exactly
595 595 for efficiency.
596 596 """
597 597 # i18n: "contains" is a keyword
598 598 pat = getstring(x, _("contains requires a pattern"))
599 599
600 600 def matches(x):
601 601 if not matchmod.patkind(pat):
602 602 pats = pathutil.canonpath(repo.root, repo.getcwd(), pat)
603 603 if pats in repo[x]:
604 604 return True
605 605 else:
606 606 c = repo[x]
607 607 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
608 608 for f in c.manifest():
609 609 if m(f):
610 610 return True
611 611 return False
612 612
613 613 return subset.filter(matches)
614 614
615 615 def converted(repo, subset, x):
616 616 """``converted([id])``
617 617 Changesets converted from the given identifier in the old repository if
618 618 present, or all converted changesets if no identifier is specified.
619 619 """
620 620
621 621 # There is exactly no chance of resolving the revision, so do a simple
622 622 # string compare and hope for the best
623 623
624 624 rev = None
625 625 # i18n: "converted" is a keyword
626 626 l = getargs(x, 0, 1, _('converted takes one or no arguments'))
627 627 if l:
628 628 # i18n: "converted" is a keyword
629 629 rev = getstring(l[0], _('converted requires a revision'))
630 630
631 631 def _matchvalue(r):
632 632 source = repo[r].extra().get('convert_revision', None)
633 633 return source is not None and (rev is None or source.startswith(rev))
634 634
635 635 return subset.filter(lambda r: _matchvalue(r))
636 636
637 637 def date(repo, subset, x):
638 638 """``date(interval)``
639 639 Changesets within the interval, see :hg:`help dates`.
640 640 """
641 641 # i18n: "date" is a keyword
642 642 ds = getstring(x, _("date requires a string"))
643 643 dm = util.matchdate(ds)
644 644 return subset.filter(lambda x: dm(repo[x].date()[0]))
645 645
646 646 def desc(repo, subset, x):
647 647 """``desc(string)``
648 648 Search commit message for string. The match is case-insensitive.
649 649 """
650 650 # i18n: "desc" is a keyword
651 651 ds = encoding.lower(getstring(x, _("desc requires a string")))
652 652
653 653 def matches(x):
654 654 c = repo[x]
655 655 return ds in encoding.lower(c.description())
656 656
657 657 return subset.filter(matches)
658 658
659 659 def _descendants(repo, subset, x, followfirst=False):
660 660 roots = getset(repo, spanset(repo), x)
661 661 if not roots:
662 662 return baseset()
663 663 s = _revdescendants(repo, roots, followfirst)
664 664
665 665 # Both sets need to be ascending in order to lazily return the union
666 666 # in the correct order.
667 667 base = subset & roots
668 668 desc = subset & s
669 669 result = base + desc
670 670 if subset.isascending():
671 671 result.sort()
672 672 elif subset.isdescending():
673 673 result.sort(reverse=True)
674 674 else:
675 675 result = subset & result
676 676 return result
677 677
678 678 def descendants(repo, subset, x):
679 679 """``descendants(set)``
680 680 Changesets which are descendants of changesets in set.
681 681 """
682 682 return _descendants(repo, subset, x)
683 683
684 684 def _firstdescendants(repo, subset, x):
685 685 # ``_firstdescendants(set)``
686 686 # Like ``descendants(set)`` but follows only the first parents.
687 687 return _descendants(repo, subset, x, followfirst=True)
688 688
689 689 def destination(repo, subset, x):
690 690 """``destination([set])``
691 691 Changesets that were created by a graft, transplant or rebase operation,
692 692 with the given revisions specified as the source. Omitting the optional set
693 693 is the same as passing all().
694 694 """
695 695 if x is not None:
696 696 sources = getset(repo, spanset(repo), x)
697 697 else:
698 698 sources = getall(repo, spanset(repo), x)
699 699
700 700 dests = set()
701 701
702 702 # subset contains all of the possible destinations that can be returned, so
703 703 # iterate over them and see if their source(s) were provided in the arg set.
704 704 # Even if the immediate src of r is not in the arg set, src's source (or
705 705 # further back) may be. Scanning back further than the immediate src allows
706 706 # transitive transplants and rebases to yield the same results as transitive
707 707 # grafts.
708 708 for r in subset:
709 709 src = _getrevsource(repo, r)
710 710 lineage = None
711 711
712 712 while src is not None:
713 713 if lineage is None:
714 714 lineage = list()
715 715
716 716 lineage.append(r)
717 717
718 718 # The visited lineage is a match if the current source is in the arg
719 719 # set. Since every candidate dest is visited by way of iterating
720 720 # subset, any dests further back in the lineage will be tested by a
721 721 # different iteration over subset. Likewise, if the src was already
722 722 # selected, the current lineage can be selected without going back
723 723 # further.
724 724 if src in sources or src in dests:
725 725 dests.update(lineage)
726 726 break
727 727
728 728 r = src
729 729 src = _getrevsource(repo, r)
730 730
731 731 return subset.filter(dests.__contains__)
732 732
733 733 def divergent(repo, subset, x):
734 734 """``divergent()``
735 735 Final successors of changesets with an alternative set of final successors.
736 736 """
737 737 # i18n: "divergent" is a keyword
738 738 getargs(x, 0, 0, _("divergent takes no arguments"))
739 739 divergent = obsmod.getrevs(repo, 'divergent')
740 740 return subset & divergent
741 741
742 742 def draft(repo, subset, x):
743 743 """``draft()``
744 744 Changeset in draft phase."""
745 745 # i18n: "draft" is a keyword
746 746 getargs(x, 0, 0, _("draft takes no arguments"))
747 747 phase = repo._phasecache.phase
748 748 target = phases.draft
749 749 condition = lambda r: phase(repo, r) == target
750 750 return subset.filter(condition, cache=False)
751 751
752 752 def extinct(repo, subset, x):
753 753 """``extinct()``
754 754 Obsolete changesets with obsolete descendants only.
755 755 """
756 756 # i18n: "extinct" is a keyword
757 757 getargs(x, 0, 0, _("extinct takes no arguments"))
758 758 extincts = obsmod.getrevs(repo, 'extinct')
759 759 return subset & extincts
760 760
761 761 def extra(repo, subset, x):
762 762 """``extra(label, [value])``
763 763 Changesets with the given label in the extra metadata, with the given
764 764 optional value.
765 765
766 766 If `value` starts with `re:`, the remainder of the value is treated as
767 767 a regular expression. To match a value that actually starts with `re:`,
768 768 use the prefix `literal:`.
769 769 """
770 770
771 771 # i18n: "extra" is a keyword
772 772 l = getargs(x, 1, 2, _('extra takes at least 1 and at most 2 arguments'))
773 773 # i18n: "extra" is a keyword
774 774 label = getstring(l[0], _('first argument to extra must be a string'))
775 775 value = None
776 776
777 777 if len(l) > 1:
778 778 # i18n: "extra" is a keyword
779 779 value = getstring(l[1], _('second argument to extra must be a string'))
780 780 kind, value, matcher = _stringmatcher(value)
781 781
782 782 def _matchvalue(r):
783 783 extra = repo[r].extra()
784 784 return label in extra and (value is None or matcher(extra[label]))
785 785
786 786 return subset.filter(lambda r: _matchvalue(r))
787 787
788 788 def filelog(repo, subset, x):
789 789 """``filelog(pattern)``
790 790 Changesets connected to the specified filelog.
791 791
792 792 For performance reasons, visits only revisions mentioned in the file-level
793 793 filelog, rather than filtering through all changesets (much faster, but
794 794 doesn't include deletes or duplicate changes). For a slower, more accurate
795 795 result, use ``file()``.
796 796
797 797 The pattern without explicit kind like ``glob:`` is expected to be
798 798 relative to the current directory and match against a file exactly
799 799 for efficiency.
800 800 """
801 801
802 802 # i18n: "filelog" is a keyword
803 803 pat = getstring(x, _("filelog requires a pattern"))
804 804 s = set()
805 805
806 806 if not matchmod.patkind(pat):
807 807 f = pathutil.canonpath(repo.root, repo.getcwd(), pat)
808 808 fl = repo.file(f)
809 809 for fr in fl:
810 810 s.add(fl.linkrev(fr))
811 811 else:
812 812 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=repo[None])
813 813 for f in repo[None]:
814 814 if m(f):
815 815 fl = repo.file(f)
816 816 for fr in fl:
817 817 s.add(fl.linkrev(fr))
818 818
819 819 return subset & s
820 820
821 821 def first(repo, subset, x):
822 822 """``first(set, [n])``
823 823 An alias for limit().
824 824 """
825 825 return limit(repo, subset, x)
826 826
827 827 def _follow(repo, subset, x, name, followfirst=False):
828 828 l = getargs(x, 0, 1, _("%s takes no arguments or a filename") % name)
829 829 c = repo['.']
830 830 if l:
831 831 x = getstring(l[0], _("%s expected a filename") % name)
832 832 if x in c:
833 833 cx = c[x]
834 834 s = set(ctx.rev() for ctx in cx.ancestors(followfirst=followfirst))
835 835 # include the revision responsible for the most recent version
836 836 s.add(cx.linkrev())
837 837 else:
838 838 return baseset()
839 839 else:
840 840 s = _revancestors(repo, baseset([c.rev()]), followfirst)
841 841
842 842 return subset & s
843 843
844 844 def follow(repo, subset, x):
845 845 """``follow([file])``
846 846 An alias for ``::.`` (ancestors of the working copy's first parent).
847 847 If a filename is specified, the history of the given file is followed,
848 848 including copies.
849 849 """
850 850 return _follow(repo, subset, x, 'follow')
851 851
852 852 def _followfirst(repo, subset, x):
853 853 # ``followfirst([file])``
854 854 # Like ``follow([file])`` but follows only the first parent of
855 855 # every revision or file revision.
856 856 return _follow(repo, subset, x, '_followfirst', followfirst=True)
857 857
858 858 def getall(repo, subset, x):
859 859 """``all()``
860 860 All changesets, the same as ``0:tip``.
861 861 """
862 862 # i18n: "all" is a keyword
863 863 getargs(x, 0, 0, _("all takes no arguments"))
864 864 return subset
865 865
866 866 def grep(repo, subset, x):
867 867 """``grep(regex)``
868 868 Like ``keyword(string)`` but accepts a regex. Use ``grep(r'...')``
869 869 to ensure special escape characters are handled correctly. Unlike
870 870 ``keyword(string)``, the match is case-sensitive.
871 871 """
872 872 try:
873 873 # i18n: "grep" is a keyword
874 874 gr = re.compile(getstring(x, _("grep requires a string")))
875 875 except re.error, e:
876 876 raise error.ParseError(_('invalid match pattern: %s') % e)
877 877
878 878 def matches(x):
879 879 c = repo[x]
880 880 for e in c.files() + [c.user(), c.description()]:
881 881 if gr.search(e):
882 882 return True
883 883 return False
884 884
885 885 return subset.filter(matches)
886 886
887 887 def _matchfiles(repo, subset, x):
888 888 # _matchfiles takes a revset list of prefixed arguments:
889 889 #
890 890 # [p:foo, i:bar, x:baz]
891 891 #
892 892 # builds a match object from them and filters subset. Allowed
893 893 # prefixes are 'p:' for regular patterns, 'i:' for include
894 894 # patterns and 'x:' for exclude patterns. Use 'r:' prefix to pass
895 895 # a revision identifier, or the empty string to reference the
896 896 # working directory, from which the match object is
897 897 # initialized. Use 'd:' to set the default matching mode, default
898 898 # to 'glob'. At most one 'r:' and 'd:' argument can be passed.
899 899
900 900 # i18n: "_matchfiles" is a keyword
901 901 l = getargs(x, 1, -1, _("_matchfiles requires at least one argument"))
902 902 pats, inc, exc = [], [], []
903 903 rev, default = None, None
904 904 for arg in l:
905 905 # i18n: "_matchfiles" is a keyword
906 906 s = getstring(arg, _("_matchfiles requires string arguments"))
907 907 prefix, value = s[:2], s[2:]
908 908 if prefix == 'p:':
909 909 pats.append(value)
910 910 elif prefix == 'i:':
911 911 inc.append(value)
912 912 elif prefix == 'x:':
913 913 exc.append(value)
914 914 elif prefix == 'r:':
915 915 if rev is not None:
916 916 # i18n: "_matchfiles" is a keyword
917 917 raise error.ParseError(_('_matchfiles expected at most one '
918 918 'revision'))
919 919 rev = value
920 920 elif prefix == 'd:':
921 921 if default is not None:
922 922 # i18n: "_matchfiles" is a keyword
923 923 raise error.ParseError(_('_matchfiles expected at most one '
924 924 'default mode'))
925 925 default = value
926 926 else:
927 927 # i18n: "_matchfiles" is a keyword
928 928 raise error.ParseError(_('invalid _matchfiles prefix: %s') % prefix)
929 929 if not default:
930 930 default = 'glob'
931 931
932 932 m = matchmod.match(repo.root, repo.getcwd(), pats, include=inc,
933 933 exclude=exc, ctx=repo[rev], default=default)
934 934
935 935 def matches(x):
936 936 for f in repo[x].files():
937 937 if m(f):
938 938 return True
939 939 return False
940 940
941 941 return subset.filter(matches)
942 942
943 943 def hasfile(repo, subset, x):
944 944 """``file(pattern)``
945 945 Changesets affecting files matched by pattern.
946 946
947 947 For a faster but less accurate result, consider using ``filelog()``
948 948 instead.
949 949
950 950 This predicate uses ``glob:`` as the default kind of pattern.
951 951 """
952 952 # i18n: "file" is a keyword
953 953 pat = getstring(x, _("file requires a pattern"))
954 954 return _matchfiles(repo, subset, ('string', 'p:' + pat))
955 955
956 956 def head(repo, subset, x):
957 957 """``head()``
958 958 Changeset is a named branch head.
959 959 """
960 960 # i18n: "head" is a keyword
961 961 getargs(x, 0, 0, _("head takes no arguments"))
962 962 hs = set()
963 963 for b, ls in repo.branchmap().iteritems():
964 964 hs.update(repo[h].rev() for h in ls)
965 965 return baseset(hs).filter(subset.__contains__)
966 966
967 967 def heads(repo, subset, x):
968 968 """``heads(set)``
969 969 Members of set with no children in set.
970 970 """
971 971 s = getset(repo, subset, x)
972 972 ps = parents(repo, subset, x)
973 973 return s - ps
974 974
975 975 def hidden(repo, subset, x):
976 976 """``hidden()``
977 977 Hidden changesets.
978 978 """
979 979 # i18n: "hidden" is a keyword
980 980 getargs(x, 0, 0, _("hidden takes no arguments"))
981 981 hiddenrevs = repoview.filterrevs(repo, 'visible')
982 982 return subset & hiddenrevs
983 983
984 984 def keyword(repo, subset, x):
985 985 """``keyword(string)``
986 986 Search commit message, user name, and names of changed files for
987 987 string. The match is case-insensitive.
988 988 """
989 989 # i18n: "keyword" is a keyword
990 990 kw = encoding.lower(getstring(x, _("keyword requires a string")))
991 991
992 992 def matches(r):
993 993 c = repo[r]
994 994 return util.any(kw in encoding.lower(t) for t in c.files() + [c.user(),
995 995 c.description()])
996 996
997 997 return subset.filter(matches)
998 998
999 999 def limit(repo, subset, x):
1000 1000 """``limit(set, [n])``
1001 1001 First n members of set, defaulting to 1.
1002 1002 """
1003 1003 # i18n: "limit" is a keyword
1004 1004 l = getargs(x, 1, 2, _("limit requires one or two arguments"))
1005 1005 try:
1006 1006 lim = 1
1007 1007 if len(l) == 2:
1008 1008 # i18n: "limit" is a keyword
1009 1009 lim = int(getstring(l[1], _("limit requires a number")))
1010 1010 except (TypeError, ValueError):
1011 1011 # i18n: "limit" is a keyword
1012 1012 raise error.ParseError(_("limit expects a number"))
1013 1013 ss = subset
1014 1014 os = getset(repo, spanset(repo), l[0])
1015 1015 result = []
1016 1016 it = iter(os)
1017 1017 for x in xrange(lim):
1018 1018 try:
1019 1019 y = it.next()
1020 1020 if y in ss:
1021 1021 result.append(y)
1022 1022 except (StopIteration):
1023 1023 break
1024 1024 return baseset(result)
1025 1025
1026 1026 def last(repo, subset, x):
1027 1027 """``last(set, [n])``
1028 1028 Last n members of set, defaulting to 1.
1029 1029 """
1030 1030 # i18n: "last" is a keyword
1031 1031 l = getargs(x, 1, 2, _("last requires one or two arguments"))
1032 1032 try:
1033 1033 lim = 1
1034 1034 if len(l) == 2:
1035 1035 # i18n: "last" is a keyword
1036 1036 lim = int(getstring(l[1], _("last requires a number")))
1037 1037 except (TypeError, ValueError):
1038 1038 # i18n: "last" is a keyword
1039 1039 raise error.ParseError(_("last expects a number"))
1040 1040 ss = subset
1041 1041 os = getset(repo, spanset(repo), l[0])
1042 1042 os.reverse()
1043 1043 result = []
1044 1044 it = iter(os)
1045 1045 for x in xrange(lim):
1046 1046 try:
1047 1047 y = it.next()
1048 1048 if y in ss:
1049 1049 result.append(y)
1050 1050 except (StopIteration):
1051 1051 break
1052 1052 return baseset(result)
1053 1053
1054 1054 def maxrev(repo, subset, x):
1055 1055 """``max(set)``
1056 1056 Changeset with highest revision number in set.
1057 1057 """
1058 1058 os = getset(repo, spanset(repo), x)
1059 1059 if os:
1060 1060 m = os.max()
1061 1061 if m in subset:
1062 1062 return baseset([m])
1063 1063 return baseset()
1064 1064
1065 1065 def merge(repo, subset, x):
1066 1066 """``merge()``
1067 1067 Changeset is a merge changeset.
1068 1068 """
1069 1069 # i18n: "merge" is a keyword
1070 1070 getargs(x, 0, 0, _("merge takes no arguments"))
1071 1071 cl = repo.changelog
1072 1072 return subset.filter(lambda r: cl.parentrevs(r)[1] != -1)
1073 1073
1074 1074 def branchpoint(repo, subset, x):
1075 1075 """``branchpoint()``
1076 1076 Changesets with more than one child.
1077 1077 """
1078 1078 # i18n: "branchpoint" is a keyword
1079 1079 getargs(x, 0, 0, _("branchpoint takes no arguments"))
1080 1080 cl = repo.changelog
1081 1081 if not subset:
1082 1082 return baseset()
1083 1083 baserev = min(subset)
1084 1084 parentscount = [0]*(len(repo) - baserev)
1085 1085 for r in cl.revs(start=baserev + 1):
1086 1086 for p in cl.parentrevs(r):
1087 1087 if p >= baserev:
1088 1088 parentscount[p - baserev] += 1
1089 1089 return subset.filter(lambda r: parentscount[r - baserev] > 1)
1090 1090
1091 1091 def minrev(repo, subset, x):
1092 1092 """``min(set)``
1093 1093 Changeset with lowest revision number in set.
1094 1094 """
1095 1095 os = getset(repo, spanset(repo), x)
1096 1096 if os:
1097 1097 m = os.min()
1098 1098 if m in subset:
1099 1099 return baseset([m])
1100 1100 return baseset()
1101 1101
1102 1102 def modifies(repo, subset, x):
1103 1103 """``modifies(pattern)``
1104 1104 Changesets modifying files matched by pattern.
1105 1105
1106 1106 The pattern without explicit kind like ``glob:`` is expected to be
1107 1107 relative to the current directory and match against a file or a
1108 1108 directory.
1109 1109 """
1110 1110 # i18n: "modifies" is a keyword
1111 1111 pat = getstring(x, _("modifies requires a pattern"))
1112 1112 return checkstatus(repo, subset, pat, 0)
1113 1113
1114 1114 def node_(repo, subset, x):
1115 1115 """``id(string)``
1116 1116 Revision non-ambiguously specified by the given hex string prefix.
1117 1117 """
1118 1118 # i18n: "id" is a keyword
1119 1119 l = getargs(x, 1, 1, _("id requires one argument"))
1120 1120 # i18n: "id" is a keyword
1121 1121 n = getstring(l[0], _("id requires a string"))
1122 1122 if len(n) == 40:
1123 1123 rn = repo[n].rev()
1124 1124 else:
1125 1125 rn = None
1126 1126 pm = repo.changelog._partialmatch(n)
1127 1127 if pm is not None:
1128 1128 rn = repo.changelog.rev(pm)
1129 1129
1130 1130 if rn is None:
1131 1131 return baseset()
1132 1132 result = baseset([rn])
1133 1133 return result & subset
1134 1134
1135 1135 def obsolete(repo, subset, x):
1136 1136 """``obsolete()``
1137 1137 Mutable changeset with a newer version."""
1138 1138 # i18n: "obsolete" is a keyword
1139 1139 getargs(x, 0, 0, _("obsolete takes no arguments"))
1140 1140 obsoletes = obsmod.getrevs(repo, 'obsolete')
1141 1141 return subset & obsoletes
1142 1142
1143 1143 def origin(repo, subset, x):
1144 1144 """``origin([set])``
1145 1145 Changesets that were specified as a source for the grafts, transplants or
1146 1146 rebases that created the given revisions. Omitting the optional set is the
1147 1147 same as passing all(). If a changeset created by these operations is itself
1148 1148 specified as a source for one of these operations, only the source changeset
1149 1149 for the first operation is selected.
1150 1150 """
1151 1151 if x is not None:
1152 1152 dests = getset(repo, spanset(repo), x)
1153 1153 else:
1154 1154 dests = getall(repo, spanset(repo), x)
1155 1155
1156 1156 def _firstsrc(rev):
1157 1157 src = _getrevsource(repo, rev)
1158 1158 if src is None:
1159 1159 return None
1160 1160
1161 1161 while True:
1162 1162 prev = _getrevsource(repo, src)
1163 1163
1164 1164 if prev is None:
1165 1165 return src
1166 1166 src = prev
1167 1167
1168 1168 o = set([_firstsrc(r) for r in dests])
1169 1169 o -= set([None])
1170 1170 return subset & o
1171 1171
1172 1172 def outgoing(repo, subset, x):
1173 1173 """``outgoing([path])``
1174 1174 Changesets not found in the specified destination repository, or the
1175 1175 default push location.
1176 1176 """
1177 1177 import hg # avoid start-up nasties
1178 1178 # i18n: "outgoing" is a keyword
1179 1179 l = getargs(x, 0, 1, _("outgoing takes one or no arguments"))
1180 1180 # i18n: "outgoing" is a keyword
1181 1181 dest = l and getstring(l[0], _("outgoing requires a repository path")) or ''
1182 1182 dest = repo.ui.expandpath(dest or 'default-push', dest or 'default')
1183 1183 dest, branches = hg.parseurl(dest)
1184 1184 revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
1185 1185 if revs:
1186 1186 revs = [repo.lookup(rev) for rev in revs]
1187 1187 other = hg.peer(repo, {}, dest)
1188 1188 repo.ui.pushbuffer()
1189 1189 outgoing = discovery.findcommonoutgoing(repo, other, onlyheads=revs)
1190 1190 repo.ui.popbuffer()
1191 1191 cl = repo.changelog
1192 1192 o = set([cl.rev(r) for r in outgoing.missing])
1193 1193 return subset & o
1194 1194
1195 1195 def p1(repo, subset, x):
1196 1196 """``p1([set])``
1197 1197 First parent of changesets in set, or the working directory.
1198 1198 """
1199 1199 if x is None:
1200 1200 p = repo[x].p1().rev()
1201 1201 if p >= 0:
1202 1202 return subset & baseset([p])
1203 1203 return baseset()
1204 1204
1205 1205 ps = set()
1206 1206 cl = repo.changelog
1207 1207 for r in getset(repo, spanset(repo), x):
1208 1208 ps.add(cl.parentrevs(r)[0])
1209 1209 ps -= set([node.nullrev])
1210 1210 return subset & ps
1211 1211
1212 1212 def p2(repo, subset, x):
1213 1213 """``p2([set])``
1214 1214 Second parent of changesets in set, or the working directory.
1215 1215 """
1216 1216 if x is None:
1217 1217 ps = repo[x].parents()
1218 1218 try:
1219 1219 p = ps[1].rev()
1220 1220 if p >= 0:
1221 1221 return subset & baseset([p])
1222 1222 return baseset()
1223 1223 except IndexError:
1224 1224 return baseset()
1225 1225
1226 1226 ps = set()
1227 1227 cl = repo.changelog
1228 1228 for r in getset(repo, spanset(repo), x):
1229 1229 ps.add(cl.parentrevs(r)[1])
1230 1230 ps -= set([node.nullrev])
1231 1231 return subset & ps
1232 1232
1233 1233 def parents(repo, subset, x):
1234 1234 """``parents([set])``
1235 1235 The set of all parents for all changesets in set, or the working directory.
1236 1236 """
1237 1237 if x is None:
1238 1238 ps = set(p.rev() for p in repo[x].parents())
1239 1239 else:
1240 1240 ps = set()
1241 1241 cl = repo.changelog
1242 1242 for r in getset(repo, spanset(repo), x):
1243 1243 ps.update(cl.parentrevs(r))
1244 1244 ps -= set([node.nullrev])
1245 1245 return subset & ps
1246 1246
1247 1247 def parentspec(repo, subset, x, n):
1248 1248 """``set^0``
1249 1249 The set.
1250 1250 ``set^1`` (or ``set^``), ``set^2``
1251 1251 First or second parent, respectively, of all changesets in set.
1252 1252 """
1253 1253 try:
1254 1254 n = int(n[1])
1255 1255 if n not in (0, 1, 2):
1256 1256 raise ValueError
1257 1257 except (TypeError, ValueError):
1258 1258 raise error.ParseError(_("^ expects a number 0, 1, or 2"))
1259 1259 ps = set()
1260 1260 cl = repo.changelog
1261 1261 for r in getset(repo, baseset(cl), x):
1262 1262 if n == 0:
1263 1263 ps.add(r)
1264 1264 elif n == 1:
1265 1265 ps.add(cl.parentrevs(r)[0])
1266 1266 elif n == 2:
1267 1267 parents = cl.parentrevs(r)
1268 1268 if len(parents) > 1:
1269 1269 ps.add(parents[1])
1270 1270 return subset & ps
1271 1271
1272 1272 def present(repo, subset, x):
1273 1273 """``present(set)``
1274 1274 An empty set, if any revision in set isn't found; otherwise,
1275 1275 all revisions in set.
1276 1276
1277 1277 If any of specified revisions is not present in the local repository,
1278 1278 the query is normally aborted. But this predicate allows the query
1279 1279 to continue even in such cases.
1280 1280 """
1281 1281 try:
1282 1282 return getset(repo, subset, x)
1283 1283 except error.RepoLookupError:
1284 1284 return baseset()
1285 1285
1286 1286 def public(repo, subset, x):
1287 1287 """``public()``
1288 1288 Changeset in public phase."""
1289 1289 # i18n: "public" is a keyword
1290 1290 getargs(x, 0, 0, _("public takes no arguments"))
1291 1291 phase = repo._phasecache.phase
1292 1292 target = phases.public
1293 1293 condition = lambda r: phase(repo, r) == target
1294 1294 return subset.filter(condition, cache=False)
1295 1295
1296 1296 def remote(repo, subset, x):
1297 1297 """``remote([id [,path]])``
1298 1298 Local revision that corresponds to the given identifier in a
1299 1299 remote repository, if present. Here, the '.' identifier is a
1300 1300 synonym for the current local branch.
1301 1301 """
1302 1302
1303 1303 import hg # avoid start-up nasties
1304 1304 # i18n: "remote" is a keyword
1305 1305 l = getargs(x, 0, 2, _("remote takes one, two or no arguments"))
1306 1306
1307 1307 q = '.'
1308 1308 if len(l) > 0:
1309 1309 # i18n: "remote" is a keyword
1310 1310 q = getstring(l[0], _("remote requires a string id"))
1311 1311 if q == '.':
1312 1312 q = repo['.'].branch()
1313 1313
1314 1314 dest = ''
1315 1315 if len(l) > 1:
1316 1316 # i18n: "remote" is a keyword
1317 1317 dest = getstring(l[1], _("remote requires a repository path"))
1318 1318 dest = repo.ui.expandpath(dest or 'default')
1319 1319 dest, branches = hg.parseurl(dest)
1320 1320 revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
1321 1321 if revs:
1322 1322 revs = [repo.lookup(rev) for rev in revs]
1323 1323 other = hg.peer(repo, {}, dest)
1324 1324 n = other.lookup(q)
1325 1325 if n in repo:
1326 1326 r = repo[n].rev()
1327 1327 if r in subset:
1328 1328 return baseset([r])
1329 1329 return baseset()
1330 1330
1331 1331 def removes(repo, subset, x):
1332 1332 """``removes(pattern)``
1333 1333 Changesets which remove files matching pattern.
1334 1334
1335 1335 The pattern without explicit kind like ``glob:`` is expected to be
1336 1336 relative to the current directory and match against a file or a
1337 1337 directory.
1338 1338 """
1339 1339 # i18n: "removes" is a keyword
1340 1340 pat = getstring(x, _("removes requires a pattern"))
1341 1341 return checkstatus(repo, subset, pat, 2)
1342 1342
1343 1343 def rev(repo, subset, x):
1344 1344 """``rev(number)``
1345 1345 Revision with the given numeric identifier.
1346 1346 """
1347 1347 # i18n: "rev" is a keyword
1348 1348 l = getargs(x, 1, 1, _("rev requires one argument"))
1349 1349 try:
1350 1350 # i18n: "rev" is a keyword
1351 1351 l = int(getstring(l[0], _("rev requires a number")))
1352 1352 except (TypeError, ValueError):
1353 1353 # i18n: "rev" is a keyword
1354 1354 raise error.ParseError(_("rev expects a number"))
1355 1355 if l not in fullreposet(repo):
1356 1356 return baseset()
1357 1357 return subset & baseset([l])
1358 1358
1359 1359 def matching(repo, subset, x):
1360 1360 """``matching(revision [, field])``
1361 1361 Changesets in which a given set of fields match the set of fields in the
1362 1362 selected revision or set.
1363 1363
1364 1364 To match more than one field pass the list of fields to match separated
1365 1365 by spaces (e.g. ``author description``).
1366 1366
1367 1367 Valid fields are most regular revision fields and some special fields.
1368 1368
1369 1369 Regular revision fields are ``description``, ``author``, ``branch``,
1370 1370 ``date``, ``files``, ``phase``, ``parents``, ``substate``, ``user``
1371 1371 and ``diff``.
1372 1372 Note that ``author`` and ``user`` are synonyms. ``diff`` refers to the
1373 1373 contents of the revision. Two revisions matching their ``diff`` will
1374 1374 also match their ``files``.
1375 1375
1376 1376 Special fields are ``summary`` and ``metadata``:
1377 1377 ``summary`` matches the first line of the description.
1378 1378 ``metadata`` is equivalent to matching ``description user date``
1379 1379 (i.e. it matches the main metadata fields).
1380 1380
1381 1381 ``metadata`` is the default field which is used when no fields are
1382 1382 specified. You can match more than one field at a time.
1383 1383 """
1384 1384 # i18n: "matching" is a keyword
1385 1385 l = getargs(x, 1, 2, _("matching takes 1 or 2 arguments"))
1386 1386
1387 1387 revs = getset(repo, baseset(repo.changelog), l[0])
1388 1388
1389 1389 fieldlist = ['metadata']
1390 1390 if len(l) > 1:
1391 1391 fieldlist = getstring(l[1],
1392 1392 # i18n: "matching" is a keyword
1393 1393 _("matching requires a string "
1394 1394 "as its second argument")).split()
1395 1395
1396 1396 # Make sure that there are no repeated fields,
1397 1397 # expand the 'special' 'metadata' field type
1398 1398 # and check the 'files' whenever we check the 'diff'
1399 1399 fields = []
1400 1400 for field in fieldlist:
1401 1401 if field == 'metadata':
1402 1402 fields += ['user', 'description', 'date']
1403 1403 elif field == 'diff':
1404 1404 # a revision matching the diff must also match the files
1405 1405 # since matching the diff is very costly, make sure to
1406 1406 # also match the files first
1407 1407 fields += ['files', 'diff']
1408 1408 else:
1409 1409 if field == 'author':
1410 1410 field = 'user'
1411 1411 fields.append(field)
1412 1412 fields = set(fields)
1413 1413 if 'summary' in fields and 'description' in fields:
1414 1414 # If a revision matches its description it also matches its summary
1415 1415 fields.discard('summary')
1416 1416
1417 1417 # We may want to match more than one field
1418 1418 # Not all fields take the same amount of time to be matched
1419 1419 # Sort the selected fields in order of increasing matching cost
1420 1420 fieldorder = ['phase', 'parents', 'user', 'date', 'branch', 'summary',
1421 1421 'files', 'description', 'substate', 'diff']
1422 1422 def fieldkeyfunc(f):
1423 1423 try:
1424 1424 return fieldorder.index(f)
1425 1425 except ValueError:
1426 1426 # assume an unknown field is very costly
1427 1427 return len(fieldorder)
1428 1428 fields = list(fields)
1429 1429 fields.sort(key=fieldkeyfunc)
1430 1430
1431 1431 # Each field will be matched with its own "getfield" function
1432 1432 # which will be added to the getfieldfuncs array of functions
1433 1433 getfieldfuncs = []
1434 1434 _funcs = {
1435 1435 'user': lambda r: repo[r].user(),
1436 1436 'branch': lambda r: repo[r].branch(),
1437 1437 'date': lambda r: repo[r].date(),
1438 1438 'description': lambda r: repo[r].description(),
1439 1439 'files': lambda r: repo[r].files(),
1440 1440 'parents': lambda r: repo[r].parents(),
1441 1441 'phase': lambda r: repo[r].phase(),
1442 1442 'substate': lambda r: repo[r].substate,
1443 1443 'summary': lambda r: repo[r].description().splitlines()[0],
1444 1444 'diff': lambda r: list(repo[r].diff(git=True),)
1445 1445 }
1446 1446 for info in fields:
1447 1447 getfield = _funcs.get(info, None)
1448 1448 if getfield is None:
1449 1449 raise error.ParseError(
1450 1450 # i18n: "matching" is a keyword
1451 1451 _("unexpected field name passed to matching: %s") % info)
1452 1452 getfieldfuncs.append(getfield)
1453 1453 # convert the getfield array of functions into a "getinfo" function
1454 1454 # which returns an array of field values (or a single value if there
1455 1455 # is only one field to match)
1456 1456 getinfo = lambda r: [f(r) for f in getfieldfuncs]
1457 1457
1458 1458 def matches(x):
1459 1459 for rev in revs:
1460 1460 target = getinfo(rev)
1461 1461 match = True
1462 1462 for n, f in enumerate(getfieldfuncs):
1463 1463 if target[n] != f(x):
1464 1464 match = False
1465 1465 if match:
1466 1466 return True
1467 1467 return False
1468 1468
1469 1469 return subset.filter(matches)
1470 1470
1471 1471 def reverse(repo, subset, x):
1472 1472 """``reverse(set)``
1473 1473 Reverse order of set.
1474 1474 """
1475 1475 l = getset(repo, subset, x)
1476 1476 l.reverse()
1477 1477 return l
1478 1478
1479 1479 def roots(repo, subset, x):
1480 1480 """``roots(set)``
1481 1481 Changesets in set with no parent changeset in set.
1482 1482 """
1483 1483 s = getset(repo, spanset(repo), x)
1484 1484 subset = baseset([r for r in s if r in subset])
1485 1485 cs = _children(repo, subset, s)
1486 1486 return subset - cs
1487 1487
1488 1488 def secret(repo, subset, x):
1489 1489 """``secret()``
1490 1490 Changeset in secret phase."""
1491 1491 # i18n: "secret" is a keyword
1492 1492 getargs(x, 0, 0, _("secret takes no arguments"))
1493 1493 phase = repo._phasecache.phase
1494 1494 target = phases.secret
1495 1495 condition = lambda r: phase(repo, r) == target
1496 1496 return subset.filter(condition, cache=False)
1497 1497
1498 1498 def sort(repo, subset, x):
1499 1499 """``sort(set[, [-]key...])``
1500 1500 Sort set by keys. The default sort order is ascending, specify a key
1501 1501 as ``-key`` to sort in descending order.
1502 1502
1503 1503 The keys can be:
1504 1504
1505 1505 - ``rev`` for the revision number,
1506 1506 - ``branch`` for the branch name,
1507 1507 - ``desc`` for the commit message (description),
1508 1508 - ``user`` for user name (``author`` can be used as an alias),
1509 1509 - ``date`` for the commit date
1510 1510 """
1511 1511 # i18n: "sort" is a keyword
1512 1512 l = getargs(x, 1, 2, _("sort requires one or two arguments"))
1513 1513 keys = "rev"
1514 1514 if len(l) == 2:
1515 1515 # i18n: "sort" is a keyword
1516 1516 keys = getstring(l[1], _("sort spec must be a string"))
1517 1517
1518 1518 s = l[0]
1519 1519 keys = keys.split()
1520 1520 l = []
1521 1521 def invert(s):
1522 1522 return "".join(chr(255 - ord(c)) for c in s)
1523 1523 revs = getset(repo, subset, s)
1524 1524 if keys == ["rev"]:
1525 1525 revs.sort()
1526 1526 return revs
1527 1527 elif keys == ["-rev"]:
1528 1528 revs.sort(reverse=True)
1529 1529 return revs
1530 1530 for r in revs:
1531 1531 c = repo[r]
1532 1532 e = []
1533 1533 for k in keys:
1534 1534 if k == 'rev':
1535 1535 e.append(r)
1536 1536 elif k == '-rev':
1537 1537 e.append(-r)
1538 1538 elif k == 'branch':
1539 1539 e.append(c.branch())
1540 1540 elif k == '-branch':
1541 1541 e.append(invert(c.branch()))
1542 1542 elif k == 'desc':
1543 1543 e.append(c.description())
1544 1544 elif k == '-desc':
1545 1545 e.append(invert(c.description()))
1546 1546 elif k in 'user author':
1547 1547 e.append(c.user())
1548 1548 elif k in '-user -author':
1549 1549 e.append(invert(c.user()))
1550 1550 elif k == 'date':
1551 1551 e.append(c.date()[0])
1552 1552 elif k == '-date':
1553 1553 e.append(-c.date()[0])
1554 1554 else:
1555 1555 raise error.ParseError(_("unknown sort key %r") % k)
1556 1556 e.append(r)
1557 1557 l.append(e)
1558 1558 l.sort()
1559 1559 return baseset([e[-1] for e in l])
1560 1560
1561 1561 def _stringmatcher(pattern):
1562 1562 """
1563 1563 accepts a string, possibly starting with 're:' or 'literal:' prefix.
1564 1564 returns the matcher name, pattern, and matcher function.
1565 1565 missing or unknown prefixes are treated as literal matches.
1566 1566
1567 1567 helper for tests:
1568 1568 >>> def test(pattern, *tests):
1569 1569 ... kind, pattern, matcher = _stringmatcher(pattern)
1570 1570 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
1571 1571
1572 1572 exact matching (no prefix):
1573 1573 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
1574 1574 ('literal', 'abcdefg', [False, False, True])
1575 1575
1576 1576 regex matching ('re:' prefix)
1577 1577 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
1578 1578 ('re', 'a.+b', [False, False, True])
1579 1579
1580 1580 force exact matches ('literal:' prefix)
1581 1581 >>> test('literal:re:foobar', 'foobar', 're:foobar')
1582 1582 ('literal', 're:foobar', [False, True])
1583 1583
1584 1584 unknown prefixes are ignored and treated as literals
1585 1585 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
1586 1586 ('literal', 'foo:bar', [False, False, True])
1587 1587 """
1588 1588 if pattern.startswith('re:'):
1589 1589 pattern = pattern[3:]
1590 1590 try:
1591 1591 regex = re.compile(pattern)
1592 1592 except re.error, e:
1593 1593 raise error.ParseError(_('invalid regular expression: %s')
1594 1594 % e)
1595 1595 return 're', pattern, regex.search
1596 1596 elif pattern.startswith('literal:'):
1597 1597 pattern = pattern[8:]
1598 1598 return 'literal', pattern, pattern.__eq__
1599 1599
1600 1600 def _substringmatcher(pattern):
1601 1601 kind, pattern, matcher = _stringmatcher(pattern)
1602 1602 if kind == 'literal':
1603 1603 matcher = lambda s: pattern in s
1604 1604 return kind, pattern, matcher
1605 1605
1606 1606 def tag(repo, subset, x):
1607 1607 """``tag([name])``
1608 1608 The specified tag by name, or all tagged revisions if no name is given.
1609 1609
1610 1610 If `name` starts with `re:`, the remainder of the name is treated as
1611 1611 a regular expression. To match a tag that actually starts with `re:`,
1612 1612 use the prefix `literal:`.
1613 1613 """
1614 1614 # i18n: "tag" is a keyword
1615 1615 args = getargs(x, 0, 1, _("tag takes one or no arguments"))
1616 1616 cl = repo.changelog
1617 1617 if args:
1618 1618 pattern = getstring(args[0],
1619 1619 # i18n: "tag" is a keyword
1620 1620 _('the argument to tag must be a string'))
1621 1621 kind, pattern, matcher = _stringmatcher(pattern)
1622 1622 if kind == 'literal':
1623 1623 # avoid resolving all tags
1624 1624 tn = repo._tagscache.tags.get(pattern, None)
1625 1625 if tn is None:
1626 1626 raise util.Abort(_("tag '%s' does not exist") % pattern)
1627 1627 s = set([repo[tn].rev()])
1628 1628 else:
1629 1629 s = set([cl.rev(n) for t, n in repo.tagslist() if matcher(t)])
1630 1630 else:
1631 1631 s = set([cl.rev(n) for t, n in repo.tagslist() if t != 'tip'])
1632 1632 return subset & s
1633 1633
1634 1634 def tagged(repo, subset, x):
1635 1635 return tag(repo, subset, x)
1636 1636
1637 1637 def unstable(repo, subset, x):
1638 1638 """``unstable()``
1639 1639 Non-obsolete changesets with obsolete ancestors.
1640 1640 """
1641 1641 # i18n: "unstable" is a keyword
1642 1642 getargs(x, 0, 0, _("unstable takes no arguments"))
1643 1643 unstables = obsmod.getrevs(repo, 'unstable')
1644 1644 return subset & unstables
1645 1645
1646 1646
1647 1647 def user(repo, subset, x):
1648 1648 """``user(string)``
1649 1649 User name contains string. The match is case-insensitive.
1650 1650
1651 1651 If `string` starts with `re:`, the remainder of the string is treated as
1652 1652 a regular expression. To match a user that actually contains `re:`, use
1653 1653 the prefix `literal:`.
1654 1654 """
1655 1655 return author(repo, subset, x)
1656 1656
1657 1657 # for internal use
1658 1658 def _list(repo, subset, x):
1659 1659 s = getstring(x, "internal error")
1660 1660 if not s:
1661 1661 return baseset()
1662 1662 ls = [repo[r].rev() for r in s.split('\0')]
1663 1663 s = subset
1664 1664 return baseset([r for r in ls if r in s])
1665 1665
1666 1666 # for internal use
1667 1667 def _intlist(repo, subset, x):
1668 1668 s = getstring(x, "internal error")
1669 1669 if not s:
1670 1670 return baseset()
1671 1671 ls = [int(r) for r in s.split('\0')]
1672 1672 s = subset
1673 1673 return baseset([r for r in ls if r in s])
1674 1674
1675 1675 # for internal use
1676 1676 def _hexlist(repo, subset, x):
1677 1677 s = getstring(x, "internal error")
1678 1678 if not s:
1679 1679 return baseset()
1680 1680 cl = repo.changelog
1681 1681 ls = [cl.rev(node.bin(r)) for r in s.split('\0')]
1682 1682 s = subset
1683 1683 return baseset([r for r in ls if r in s])
1684 1684
1685 1685 symbols = {
1686 1686 "adds": adds,
1687 1687 "all": getall,
1688 1688 "ancestor": ancestor,
1689 1689 "ancestors": ancestors,
1690 1690 "_firstancestors": _firstancestors,
1691 1691 "author": author,
1692 1692 "only": only,
1693 1693 "bisect": bisect,
1694 1694 "bisected": bisected,
1695 1695 "bookmark": bookmark,
1696 1696 "branch": branch,
1697 1697 "branchpoint": branchpoint,
1698 1698 "bumped": bumped,
1699 1699 "bundle": bundle,
1700 1700 "children": children,
1701 1701 "closed": closed,
1702 1702 "contains": contains,
1703 1703 "converted": converted,
1704 1704 "date": date,
1705 1705 "desc": desc,
1706 1706 "descendants": descendants,
1707 1707 "_firstdescendants": _firstdescendants,
1708 1708 "destination": destination,
1709 1709 "divergent": divergent,
1710 1710 "draft": draft,
1711 1711 "extinct": extinct,
1712 1712 "extra": extra,
1713 1713 "file": hasfile,
1714 1714 "filelog": filelog,
1715 1715 "first": first,
1716 1716 "follow": follow,
1717 1717 "_followfirst": _followfirst,
1718 1718 "grep": grep,
1719 1719 "head": head,
1720 1720 "heads": heads,
1721 1721 "hidden": hidden,
1722 1722 "id": node_,
1723 1723 "keyword": keyword,
1724 1724 "last": last,
1725 1725 "limit": limit,
1726 1726 "_matchfiles": _matchfiles,
1727 1727 "max": maxrev,
1728 1728 "merge": merge,
1729 1729 "min": minrev,
1730 1730 "modifies": modifies,
1731 1731 "obsolete": obsolete,
1732 1732 "origin": origin,
1733 1733 "outgoing": outgoing,
1734 1734 "p1": p1,
1735 1735 "p2": p2,
1736 1736 "parents": parents,
1737 1737 "present": present,
1738 1738 "public": public,
1739 1739 "remote": remote,
1740 1740 "removes": removes,
1741 1741 "rev": rev,
1742 1742 "reverse": reverse,
1743 1743 "roots": roots,
1744 1744 "sort": sort,
1745 1745 "secret": secret,
1746 1746 "matching": matching,
1747 1747 "tag": tag,
1748 1748 "tagged": tagged,
1749 1749 "user": user,
1750 1750 "unstable": unstable,
1751 1751 "_list": _list,
1752 1752 "_intlist": _intlist,
1753 1753 "_hexlist": _hexlist,
1754 1754 }
1755 1755
1756 1756 # symbols which can't be used for a DoS attack for any given input
1757 1757 # (e.g. those which accept regexes as plain strings shouldn't be included)
1758 1758 # functions that just return a lot of changesets (like all) don't count here
1759 1759 safesymbols = set([
1760 1760 "adds",
1761 1761 "all",
1762 1762 "ancestor",
1763 1763 "ancestors",
1764 1764 "_firstancestors",
1765 1765 "author",
1766 1766 "bisect",
1767 1767 "bisected",
1768 1768 "bookmark",
1769 1769 "branch",
1770 1770 "branchpoint",
1771 1771 "bumped",
1772 1772 "bundle",
1773 1773 "children",
1774 1774 "closed",
1775 1775 "converted",
1776 1776 "date",
1777 1777 "desc",
1778 1778 "descendants",
1779 1779 "_firstdescendants",
1780 1780 "destination",
1781 1781 "divergent",
1782 1782 "draft",
1783 1783 "extinct",
1784 1784 "extra",
1785 1785 "file",
1786 1786 "filelog",
1787 1787 "first",
1788 1788 "follow",
1789 1789 "_followfirst",
1790 1790 "head",
1791 1791 "heads",
1792 1792 "hidden",
1793 1793 "id",
1794 1794 "keyword",
1795 1795 "last",
1796 1796 "limit",
1797 1797 "_matchfiles",
1798 1798 "max",
1799 1799 "merge",
1800 1800 "min",
1801 1801 "modifies",
1802 1802 "obsolete",
1803 1803 "origin",
1804 1804 "outgoing",
1805 1805 "p1",
1806 1806 "p2",
1807 1807 "parents",
1808 1808 "present",
1809 1809 "public",
1810 1810 "remote",
1811 1811 "removes",
1812 1812 "rev",
1813 1813 "reverse",
1814 1814 "roots",
1815 1815 "sort",
1816 1816 "secret",
1817 1817 "matching",
1818 1818 "tag",
1819 1819 "tagged",
1820 1820 "user",
1821 1821 "unstable",
1822 1822 "_list",
1823 1823 "_intlist",
1824 1824 "_hexlist",
1825 1825 ])
1826 1826
1827 1827 methods = {
1828 1828 "range": rangeset,
1829 1829 "dagrange": dagrange,
1830 1830 "string": stringset,
1831 1831 "symbol": symbolset,
1832 1832 "and": andset,
1833 1833 "or": orset,
1834 1834 "not": notset,
1835 1835 "list": listset,
1836 1836 "func": func,
1837 1837 "ancestor": ancestorspec,
1838 1838 "parent": parentspec,
1839 1839 "parentpost": p1,
1840 1840 }
1841 1841
1842 1842 def optimize(x, small):
1843 1843 if x is None:
1844 1844 return 0, x
1845 1845
1846 1846 smallbonus = 1
1847 1847 if small:
1848 1848 smallbonus = .5
1849 1849
1850 1850 op = x[0]
1851 1851 if op == 'minus':
1852 1852 return optimize(('and', x[1], ('not', x[2])), small)
1853 1853 elif op == 'dagrangepre':
1854 1854 return optimize(('func', ('symbol', 'ancestors'), x[1]), small)
1855 1855 elif op == 'dagrangepost':
1856 1856 return optimize(('func', ('symbol', 'descendants'), x[1]), small)
1857 1857 elif op == 'rangepre':
1858 1858 return optimize(('range', ('string', '0'), x[1]), small)
1859 1859 elif op == 'rangepost':
1860 1860 return optimize(('range', x[1], ('string', 'tip')), small)
1861 1861 elif op == 'negate':
1862 1862 return optimize(('string',
1863 1863 '-' + getstring(x[1], _("can't negate that"))), small)
1864 1864 elif op in 'string symbol negate':
1865 1865 return smallbonus, x # single revisions are small
1866 1866 elif op == 'and':
1867 1867 wa, ta = optimize(x[1], True)
1868 1868 wb, tb = optimize(x[2], True)
1869 1869
1870 1870 # (::x and not ::y)/(not ::y and ::x) have a fast path
1871 1871 def isonly(revs, bases):
1872 1872 return (
1873 1873 revs[0] == 'func'
1874 1874 and getstring(revs[1], _('not a symbol')) == 'ancestors'
1875 1875 and bases[0] == 'not'
1876 1876 and bases[1][0] == 'func'
1877 1877 and getstring(bases[1][1], _('not a symbol')) == 'ancestors')
1878 1878
1879 1879 w = min(wa, wb)
1880 1880 if isonly(ta, tb):
1881 1881 return w, ('func', ('symbol', 'only'), ('list', ta[2], tb[1][2]))
1882 1882 if isonly(tb, ta):
1883 1883 return w, ('func', ('symbol', 'only'), ('list', tb[2], ta[1][2]))
1884 1884
1885 1885 if wa > wb:
1886 1886 return w, (op, tb, ta)
1887 1887 return w, (op, ta, tb)
1888 1888 elif op == 'or':
1889 1889 wa, ta = optimize(x[1], False)
1890 1890 wb, tb = optimize(x[2], False)
1891 1891 if wb < wa:
1892 1892 wb, wa = wa, wb
1893 1893 return max(wa, wb), (op, ta, tb)
1894 1894 elif op == 'not':
1895 1895 o = optimize(x[1], not small)
1896 1896 return o[0], (op, o[1])
1897 1897 elif op == 'parentpost':
1898 1898 o = optimize(x[1], small)
1899 1899 return o[0], (op, o[1])
1900 1900 elif op == 'group':
1901 1901 return optimize(x[1], small)
1902 1902 elif op in 'dagrange range list parent ancestorspec':
1903 1903 if op == 'parent':
1904 1904 # x^:y means (x^) : y, not x ^ (:y)
1905 1905 post = ('parentpost', x[1])
1906 1906 if x[2][0] == 'dagrangepre':
1907 1907 return optimize(('dagrange', post, x[2][1]), small)
1908 1908 elif x[2][0] == 'rangepre':
1909 1909 return optimize(('range', post, x[2][1]), small)
1910 1910
1911 1911 wa, ta = optimize(x[1], small)
1912 1912 wb, tb = optimize(x[2], small)
1913 1913 return wa + wb, (op, ta, tb)
1914 1914 elif op == 'func':
1915 1915 f = getstring(x[1], _("not a symbol"))
1916 1916 wa, ta = optimize(x[2], small)
1917 1917 if f in ("author branch closed date desc file grep keyword "
1918 1918 "outgoing user"):
1919 1919 w = 10 # slow
1920 1920 elif f in "modifies adds removes":
1921 1921 w = 30 # slower
1922 1922 elif f == "contains":
1923 1923 w = 100 # very slow
1924 1924 elif f == "ancestor":
1925 1925 w = 1 * smallbonus
1926 1926 elif f in "reverse limit first _intlist":
1927 1927 w = 0
1928 1928 elif f in "sort":
1929 1929 w = 10 # assume most sorts look at changelog
1930 1930 else:
1931 1931 w = 1
1932 1932 return w + wa, (op, x[1], ta)
1933 1933 return 1, x
1934 1934
1935 1935 _aliasarg = ('func', ('symbol', '_aliasarg'))
1936 1936 def _getaliasarg(tree):
1937 1937 """If tree matches ('func', ('symbol', '_aliasarg'), ('string', X))
1938 1938 return X, None otherwise.
1939 1939 """
1940 1940 if (len(tree) == 3 and tree[:2] == _aliasarg
1941 1941 and tree[2][0] == 'string'):
1942 1942 return tree[2][1]
1943 1943 return None
1944 1944
1945 1945 def _checkaliasarg(tree, known=None):
1946 1946 """Check tree contains no _aliasarg construct or only ones which
1947 1947 value is in known. Used to avoid alias placeholders injection.
1948 1948 """
1949 1949 if isinstance(tree, tuple):
1950 1950 arg = _getaliasarg(tree)
1951 1951 if arg is not None and (not known or arg not in known):
1952 1952 raise error.ParseError(_("not a function: %s") % '_aliasarg')
1953 1953 for t in tree:
1954 1954 _checkaliasarg(t, known)
1955 1955
1956 1956 class revsetalias(object):
1957 1957 funcre = re.compile('^([^(]+)\(([^)]+)\)$')
1958 1958 args = None
1959 1959
1960 1960 def __init__(self, name, value):
1961 1961 '''Aliases like:
1962 1962
1963 1963 h = heads(default)
1964 1964 b($1) = ancestors($1) - ancestors(default)
1965 1965 '''
1966 1966 m = self.funcre.search(name)
1967 1967 if m:
1968 1968 self.name = m.group(1)
1969 1969 self.tree = ('func', ('symbol', m.group(1)))
1970 1970 self.args = [x.strip() for x in m.group(2).split(',')]
1971 1971 for arg in self.args:
1972 1972 # _aliasarg() is an unknown symbol only used separate
1973 1973 # alias argument placeholders from regular strings.
1974 1974 value = value.replace(arg, '_aliasarg(%r)' % (arg,))
1975 1975 else:
1976 1976 self.name = name
1977 1977 self.tree = ('symbol', name)
1978 1978
1979 1979 self.replacement, pos = parse(value)
1980 1980 if pos != len(value):
1981 1981 raise error.ParseError(_('invalid token'), pos)
1982 1982 # Check for placeholder injection
1983 1983 _checkaliasarg(self.replacement, self.args)
1984 1984
1985 1985 def _getalias(aliases, tree):
1986 1986 """If tree looks like an unexpanded alias, return it. Return None
1987 1987 otherwise.
1988 1988 """
1989 1989 if isinstance(tree, tuple) and tree:
1990 1990 if tree[0] == 'symbol' and len(tree) == 2:
1991 1991 name = tree[1]
1992 1992 alias = aliases.get(name)
1993 1993 if alias and alias.args is None and alias.tree == tree:
1994 1994 return alias
1995 1995 if tree[0] == 'func' and len(tree) > 1:
1996 1996 if tree[1][0] == 'symbol' and len(tree[1]) == 2:
1997 1997 name = tree[1][1]
1998 1998 alias = aliases.get(name)
1999 1999 if alias and alias.args is not None and alias.tree == tree[:2]:
2000 2000 return alias
2001 2001 return None
2002 2002
2003 2003 def _expandargs(tree, args):
2004 2004 """Replace _aliasarg instances with the substitution value of the
2005 2005 same name in args, recursively.
2006 2006 """
2007 2007 if not tree or not isinstance(tree, tuple):
2008 2008 return tree
2009 2009 arg = _getaliasarg(tree)
2010 2010 if arg is not None:
2011 2011 return args[arg]
2012 2012 return tuple(_expandargs(t, args) for t in tree)
2013 2013
2014 2014 def _expandaliases(aliases, tree, expanding, cache):
2015 2015 """Expand aliases in tree, recursively.
2016 2016
2017 2017 'aliases' is a dictionary mapping user defined aliases to
2018 2018 revsetalias objects.
2019 2019 """
2020 2020 if not isinstance(tree, tuple):
2021 2021 # Do not expand raw strings
2022 2022 return tree
2023 2023 alias = _getalias(aliases, tree)
2024 2024 if alias is not None:
2025 2025 if alias in expanding:
2026 2026 raise error.ParseError(_('infinite expansion of revset alias "%s" '
2027 2027 'detected') % alias.name)
2028 2028 expanding.append(alias)
2029 2029 if alias.name not in cache:
2030 2030 cache[alias.name] = _expandaliases(aliases, alias.replacement,
2031 2031 expanding, cache)
2032 2032 result = cache[alias.name]
2033 2033 expanding.pop()
2034 2034 if alias.args is not None:
2035 2035 l = getlist(tree[2])
2036 2036 if len(l) != len(alias.args):
2037 2037 raise error.ParseError(
2038 2038 _('invalid number of arguments: %s') % len(l))
2039 2039 l = [_expandaliases(aliases, a, [], cache) for a in l]
2040 2040 result = _expandargs(result, dict(zip(alias.args, l)))
2041 2041 else:
2042 2042 result = tuple(_expandaliases(aliases, t, expanding, cache)
2043 2043 for t in tree)
2044 2044 return result
2045 2045
2046 2046 def findaliases(ui, tree):
2047 2047 _checkaliasarg(tree)
2048 2048 aliases = {}
2049 2049 for k, v in ui.configitems('revsetalias'):
2050 2050 alias = revsetalias(k, v)
2051 2051 aliases[alias.name] = alias
2052 2052 return _expandaliases(aliases, tree, [], {})
2053 2053
2054 2054 def parse(spec, lookup=None):
2055 2055 p = parser.parser(tokenize, elements)
2056 2056 return p.parse(spec, lookup=lookup)
2057 2057
2058 2058 def match(ui, spec, repo=None):
2059 2059 if not spec:
2060 2060 raise error.ParseError(_("empty query"))
2061 2061 lookup = None
2062 2062 if repo:
2063 2063 lookup = repo.__contains__
2064 2064 tree, pos = parse(spec, lookup)
2065 2065 if (pos != len(spec)):
2066 2066 raise error.ParseError(_("invalid token"), pos)
2067 2067 if ui:
2068 2068 tree = findaliases(ui, tree)
2069 2069 weight, tree = optimize(tree, True)
2070 2070 def mfunc(repo, subset):
2071 2071 if util.safehasattr(subset, 'isascending'):
2072 2072 result = getset(repo, subset, tree)
2073 2073 else:
2074 2074 result = getset(repo, baseset(subset), tree)
2075 2075 return result
2076 2076 return mfunc
2077 2077
2078 2078 def formatspec(expr, *args):
2079 2079 '''
2080 2080 This is a convenience function for using revsets internally, and
2081 2081 escapes arguments appropriately. Aliases are intentionally ignored
2082 2082 so that intended expression behavior isn't accidentally subverted.
2083 2083
2084 2084 Supported arguments:
2085 2085
2086 2086 %r = revset expression, parenthesized
2087 2087 %d = int(arg), no quoting
2088 2088 %s = string(arg), escaped and single-quoted
2089 2089 %b = arg.branch(), escaped and single-quoted
2090 2090 %n = hex(arg), single-quoted
2091 2091 %% = a literal '%'
2092 2092
2093 2093 Prefixing the type with 'l' specifies a parenthesized list of that type.
2094 2094
2095 2095 >>> formatspec('%r:: and %lr', '10 or 11', ("this()", "that()"))
2096 2096 '(10 or 11):: and ((this()) or (that()))'
2097 2097 >>> formatspec('%d:: and not %d::', 10, 20)
2098 2098 '10:: and not 20::'
2099 2099 >>> formatspec('%ld or %ld', [], [1])
2100 2100 "_list('') or 1"
2101 2101 >>> formatspec('keyword(%s)', 'foo\\xe9')
2102 2102 "keyword('foo\\\\xe9')"
2103 2103 >>> b = lambda: 'default'
2104 2104 >>> b.branch = b
2105 2105 >>> formatspec('branch(%b)', b)
2106 2106 "branch('default')"
2107 2107 >>> formatspec('root(%ls)', ['a', 'b', 'c', 'd'])
2108 2108 "root(_list('a\\x00b\\x00c\\x00d'))"
2109 2109 '''
2110 2110
2111 2111 def quote(s):
2112 2112 return repr(str(s))
2113 2113
2114 2114 def argtype(c, arg):
2115 2115 if c == 'd':
2116 2116 return str(int(arg))
2117 2117 elif c == 's':
2118 2118 return quote(arg)
2119 2119 elif c == 'r':
2120 2120 parse(arg) # make sure syntax errors are confined
2121 2121 return '(%s)' % arg
2122 2122 elif c == 'n':
2123 2123 return quote(node.hex(arg))
2124 2124 elif c == 'b':
2125 2125 return quote(arg.branch())
2126 2126
2127 2127 def listexp(s, t):
2128 2128 l = len(s)
2129 2129 if l == 0:
2130 2130 return "_list('')"
2131 2131 elif l == 1:
2132 2132 return argtype(t, s[0])
2133 2133 elif t == 'd':
2134 2134 return "_intlist('%s')" % "\0".join(str(int(a)) for a in s)
2135 2135 elif t == 's':
2136 2136 return "_list('%s')" % "\0".join(s)
2137 2137 elif t == 'n':
2138 2138 return "_hexlist('%s')" % "\0".join(node.hex(a) for a in s)
2139 2139 elif t == 'b':
2140 2140 return "_list('%s')" % "\0".join(a.branch() for a in s)
2141 2141
2142 2142 m = l // 2
2143 2143 return '(%s or %s)' % (listexp(s[:m], t), listexp(s[m:], t))
2144 2144
2145 2145 ret = ''
2146 2146 pos = 0
2147 2147 arg = 0
2148 2148 while pos < len(expr):
2149 2149 c = expr[pos]
2150 2150 if c == '%':
2151 2151 pos += 1
2152 2152 d = expr[pos]
2153 2153 if d == '%':
2154 2154 ret += d
2155 2155 elif d in 'dsnbr':
2156 2156 ret += argtype(d, args[arg])
2157 2157 arg += 1
2158 2158 elif d == 'l':
2159 2159 # a list of some type
2160 2160 pos += 1
2161 2161 d = expr[pos]
2162 2162 ret += listexp(list(args[arg]), d)
2163 2163 arg += 1
2164 2164 else:
2165 2165 raise util.Abort('unexpected revspec format character %s' % d)
2166 2166 else:
2167 2167 ret += c
2168 2168 pos += 1
2169 2169
2170 2170 return ret
2171 2171
2172 2172 def prettyformat(tree):
2173 2173 def _prettyformat(tree, level, lines):
2174 2174 if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
2175 2175 lines.append((level, str(tree)))
2176 2176 else:
2177 2177 lines.append((level, '(%s' % tree[0]))
2178 2178 for s in tree[1:]:
2179 2179 _prettyformat(s, level + 1, lines)
2180 2180 lines[-1:] = [(lines[-1][0], lines[-1][1] + ')')]
2181 2181
2182 2182 lines = []
2183 2183 _prettyformat(tree, 0, lines)
2184 2184 output = '\n'.join((' '*l + s) for l, s in lines)
2185 2185 return output
2186 2186
2187 2187 def depth(tree):
2188 2188 if isinstance(tree, tuple):
2189 2189 return max(map(depth, tree)) + 1
2190 2190 else:
2191 2191 return 0
2192 2192
2193 2193 def funcsused(tree):
2194 2194 if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
2195 2195 return set()
2196 2196 else:
2197 2197 funcs = set()
2198 2198 for s in tree[1:]:
2199 2199 funcs |= funcsused(s)
2200 2200 if tree[0] == 'func':
2201 2201 funcs.add(tree[1][1])
2202 2202 return funcs
2203 2203
2204 2204 class abstractsmartset(object):
2205 2205
2206 2206 def __nonzero__(self):
2207 2207 """True if the smartset is not empty"""
2208 2208 raise NotImplementedError()
2209 2209
2210 2210 def __contains__(self, rev):
2211 2211 """provide fast membership testing"""
2212 2212 raise NotImplementedError()
2213 2213
2214 2214 def __iter__(self):
2215 2215 """iterate the set in the order it is supposed to be iterated"""
2216 2216 raise NotImplementedError()
2217 2217
2218 2218 # Attributes containing a function to perform a fast iteration in a given
2219 2219 # direction. A smartset can have none, one, or both defined.
2220 2220 #
2221 2221 # Default value is None instead of a function returning None to avoid
2222 2222 # initializing an iterator just for testing if a fast method exists.
2223 2223 fastasc = None
2224 2224 fastdesc = None
2225 2225
2226 2226 def isascending(self):
2227 2227 """True if the set will iterate in ascending order"""
2228 2228 raise NotImplementedError()
2229 2229
2230 2230 def isdescending(self):
2231 2231 """True if the set will iterate in descending order"""
2232 2232 raise NotImplementedError()
2233 2233
2234 2234 def min(self):
2235 2235 """return the minimum element in the set"""
2236 2236 if self.fastasc is not None:
2237 2237 for r in self.fastasc():
2238 2238 return r
2239 2239 raise ValueError('arg is an empty sequence')
2240 2240 return min(self)
2241 2241
2242 2242 def max(self):
2243 2243 """return the maximum element in the set"""
2244 2244 if self.fastdesc is not None:
2245 2245 for r in self.fastdesc():
2246 2246 return r
2247 2247 raise ValueError('arg is an empty sequence')
2248 2248 return max(self)
2249 2249
2250 2250 def first(self):
2251 2251 """return the first element in the set (user iteration perspective)
2252 2252
2253 2253 Return None if the set is empty"""
2254 2254 raise NotImplementedError()
2255 2255
2256 2256 def last(self):
2257 2257 """return the last element in the set (user iteration perspective)
2258 2258
2259 2259 Return None if the set is empty"""
2260 2260 raise NotImplementedError()
2261 2261
2262 2262 def __len__(self):
2263 2263 """return the length of the smartsets
2264 2264
2265 2265 This can be expensive on smartset that could be lazy otherwise."""
2266 2266 raise NotImplementedError()
2267 2267
2268 2268 def reverse(self):
2269 2269 """reverse the expected iteration order"""
2270 2270 raise NotImplementedError()
2271 2271
2272 2272 def sort(self, reverse=True):
2273 2273 """get the set to iterate in an ascending or descending order"""
2274 2274 raise NotImplementedError()
2275 2275
2276 2276 def __and__(self, other):
2277 2277 """Returns a new object with the intersection of the two collections.
2278 2278
2279 2279 This is part of the mandatory API for smartset."""
2280 2280 return self.filter(other.__contains__, cache=False)
2281 2281
2282 2282 def __add__(self, other):
2283 2283 """Returns a new object with the union of the two collections.
2284 2284
2285 2285 This is part of the mandatory API for smartset."""
2286 2286 return addset(self, other)
2287 2287
2288 2288 def __sub__(self, other):
2289 2289 """Returns a new object with the substraction of the two collections.
2290 2290
2291 2291 This is part of the mandatory API for smartset."""
2292 2292 c = other.__contains__
2293 2293 return self.filter(lambda r: not c(r), cache=False)
2294 2294
2295 2295 def filter(self, condition, cache=True):
2296 2296 """Returns this smartset filtered by condition as a new smartset.
2297 2297
2298 2298 `condition` is a callable which takes a revision number and returns a
2299 2299 boolean.
2300 2300
2301 2301 This is part of the mandatory API for smartset."""
2302 2302 # builtin cannot be cached. but do not needs to
2303 2303 if cache and util.safehasattr(condition, 'func_code'):
2304 2304 condition = util.cachefunc(condition)
2305 2305 return filteredset(self, condition)
2306 2306
2307 2307 class baseset(abstractsmartset):
2308 2308 """Basic data structure that represents a revset and contains the basic
2309 2309 operation that it should be able to perform.
2310 2310
2311 2311 Every method in this class should be implemented by any smartset class.
2312 2312 """
2313 2313 def __init__(self, data=()):
2314 2314 if not isinstance(data, list):
2315 2315 data = list(data)
2316 2316 self._list = data
2317 2317 self._ascending = None
2318 2318
2319 2319 @util.propertycache
2320 2320 def _set(self):
2321 2321 return set(self._list)
2322 2322
2323 2323 @util.propertycache
2324 2324 def _asclist(self):
2325 2325 asclist = self._list[:]
2326 2326 asclist.sort()
2327 2327 return asclist
2328 2328
2329 2329 def __iter__(self):
2330 2330 if self._ascending is None:
2331 2331 return iter(self._list)
2332 2332 elif self._ascending:
2333 2333 return iter(self._asclist)
2334 2334 else:
2335 2335 return reversed(self._asclist)
2336 2336
2337 2337 def fastasc(self):
2338 2338 return iter(self._asclist)
2339 2339
2340 2340 def fastdesc(self):
2341 2341 return reversed(self._asclist)
2342 2342
2343 2343 @util.propertycache
2344 2344 def __contains__(self):
2345 2345 return self._set.__contains__
2346 2346
2347 2347 def __nonzero__(self):
2348 2348 return bool(self._list)
2349 2349
2350 2350 def sort(self, reverse=False):
2351 2351 self._ascending = not bool(reverse)
2352 2352
2353 2353 def reverse(self):
2354 2354 if self._ascending is None:
2355 2355 self._list.reverse()
2356 2356 else:
2357 2357 self._ascending = not self._ascending
2358 2358
2359 2359 def __len__(self):
2360 2360 return len(self._list)
2361 2361
2362 2362 def isascending(self):
2363 2363 """Returns True if the collection is ascending order, False if not.
2364 2364
2365 2365 This is part of the mandatory API for smartset."""
2366 2366 if len(self) <= 1:
2367 2367 return True
2368 2368 return self._ascending is not None and self._ascending
2369 2369
2370 2370 def isdescending(self):
2371 2371 """Returns True if the collection is descending order, False if not.
2372 2372
2373 2373 This is part of the mandatory API for smartset."""
2374 2374 if len(self) <= 1:
2375 2375 return True
2376 2376 return self._ascending is not None and not self._ascending
2377 2377
2378 2378 def first(self):
2379 2379 if self:
2380 2380 if self._ascending is None:
2381 2381 return self._list[0]
2382 2382 elif self._ascending:
2383 2383 return self._asclist[0]
2384 2384 else:
2385 2385 return self._asclist[-1]
2386 2386 return None
2387 2387
2388 2388 def last(self):
2389 2389 if self:
2390 2390 if self._ascending is None:
2391 2391 return self._list[-1]
2392 2392 elif self._ascending:
2393 2393 return self._asclist[-1]
2394 2394 else:
2395 2395 return self._asclist[0]
2396 2396 return None
2397 2397
2398 2398 class filteredset(abstractsmartset):
2399 2399 """Duck type for baseset class which iterates lazily over the revisions in
2400 2400 the subset and contains a function which tests for membership in the
2401 2401 revset
2402 2402 """
2403 2403 def __init__(self, subset, condition=lambda x: True):
2404 2404 """
2405 2405 condition: a function that decide whether a revision in the subset
2406 2406 belongs to the revset or not.
2407 2407 """
2408 2408 self._subset = subset
2409 2409 self._condition = condition
2410 2410 self._cache = {}
2411 2411
2412 2412 def __contains__(self, x):
2413 2413 c = self._cache
2414 2414 if x not in c:
2415 2415 v = c[x] = x in self._subset and self._condition(x)
2416 2416 return v
2417 2417 return c[x]
2418 2418
2419 2419 def __iter__(self):
2420 2420 return self._iterfilter(self._subset)
2421 2421
2422 2422 def _iterfilter(self, it):
2423 2423 cond = self._condition
2424 2424 for x in it:
2425 2425 if cond(x):
2426 2426 yield x
2427 2427
2428 2428 @property
2429 2429 def fastasc(self):
2430 2430 it = self._subset.fastasc
2431 2431 if it is None:
2432 2432 return None
2433 2433 return lambda: self._iterfilter(it())
2434 2434
2435 2435 @property
2436 2436 def fastdesc(self):
2437 2437 it = self._subset.fastdesc
2438 2438 if it is None:
2439 2439 return None
2440 2440 return lambda: self._iterfilter(it())
2441 2441
2442 2442 def __nonzero__(self):
2443 2443 for r in self:
2444 2444 return True
2445 2445 return False
2446 2446
2447 2447 def __len__(self):
2448 2448 # Basic implementation to be changed in future patches.
2449 2449 l = baseset([r for r in self])
2450 2450 return len(l)
2451 2451
2452 2452 def sort(self, reverse=False):
2453 2453 self._subset.sort(reverse=reverse)
2454 2454
2455 2455 def reverse(self):
2456 2456 self._subset.reverse()
2457 2457
2458 2458 def isascending(self):
2459 2459 return self._subset.isascending()
2460 2460
2461 2461 def isdescending(self):
2462 2462 return self._subset.isdescending()
2463 2463
2464 2464 def first(self):
2465 2465 for x in self:
2466 2466 return x
2467 2467 return None
2468 2468
2469 2469 def last(self):
2470 2470 it = None
2471 2471 if self._subset.isascending:
2472 2472 it = self.fastdesc
2473 2473 elif self._subset.isdescending:
2474 2474 it = self.fastdesc
2475 2475 if it is None:
2476 2476 # slowly consume everything. This needs improvement
2477 2477 it = lambda: reversed(list(self))
2478 2478 for x in it():
2479 2479 return x
2480 2480 return None
2481 2481
2482 2482 class addset(abstractsmartset):
2483 2483 """Represent the addition of two sets
2484 2484
2485 2485 Wrapper structure for lazily adding two structures without losing much
2486 2486 performance on the __contains__ method
2487 2487
2488 2488 If the ascending attribute is set, that means the two structures are
2489 2489 ordered in either an ascending or descending way. Therefore, we can add
2490 2490 them maintaining the order by iterating over both at the same time
2491 2491 """
2492 2492 def __init__(self, revs1, revs2, ascending=None):
2493 2493 self._r1 = revs1
2494 2494 self._r2 = revs2
2495 2495 self._iter = None
2496 2496 self._ascending = ascending
2497 2497 self._genlist = None
2498 2498 self._asclist = None
2499 2499
2500 2500 def __len__(self):
2501 2501 return len(self._list)
2502 2502
2503 2503 def __nonzero__(self):
2504 2504 return bool(self._r1) or bool(self._r2)
2505 2505
2506 2506 @util.propertycache
2507 2507 def _list(self):
2508 2508 if not self._genlist:
2509 2509 self._genlist = baseset(self._iterator())
2510 2510 return self._genlist
2511 2511
2512 2512 def _iterator(self):
2513 2513 """Iterate over both collections without repeating elements
2514 2514
2515 2515 If the ascending attribute is not set, iterate over the first one and
2516 2516 then over the second one checking for membership on the first one so we
2517 2517 dont yield any duplicates.
2518 2518
2519 2519 If the ascending attribute is set, iterate over both collections at the
2520 2520 same time, yielding only one value at a time in the given order.
2521 2521 """
2522 2522 if self._ascending is None:
2523 2523 def gen():
2524 2524 for r in self._r1:
2525 2525 yield r
2526 2526 inr1 = self._r1.__contains__
2527 2527 for r in self._r2:
2528 2528 if not inr1(r):
2529 2529 yield r
2530 2530 gen = gen()
2531 2531 else:
2532 2532 iter1 = iter(self._r1)
2533 2533 iter2 = iter(self._r2)
2534 2534 gen = self._iterordered(self._ascending, iter1, iter2)
2535 2535 return gen
2536 2536
2537 2537 def __iter__(self):
2538 2538 if self._ascending is None:
2539 2539 if self._genlist:
2540 2540 return iter(self._genlist)
2541 2541 return iter(self._iterator())
2542 2542 self._trysetasclist()
2543 2543 if self._ascending:
2544 2544 it = self.fastasc
2545 2545 else:
2546 2546 it = self.fastdesc
2547 2547 if it is None:
2548 2548 # consume the gen and try again
2549 2549 self._list
2550 2550 return iter(self)
2551 2551 return it()
2552 2552
2553 2553 def _trysetasclist(self):
2554 2554 """populate the _asclist attribut if possible and necessary"""
2555 2555 if self._genlist is not None and self._asclist is None:
2556 2556 self._asclist = sorted(self._genlist)
2557 2557
2558 2558 @property
2559 2559 def fastasc(self):
2560 2560 self._trysetasclist()
2561 2561 if self._asclist is not None:
2562 2562 return self._asclist.__iter__
2563 2563 iter1 = self._r1.fastasc
2564 2564 iter2 = self._r2.fastasc
2565 2565 if None in (iter1, iter2):
2566 2566 return None
2567 2567 return lambda: self._iterordered(True, iter1(), iter2())
2568 2568
2569 2569 @property
2570 2570 def fastdesc(self):
2571 2571 self._trysetasclist()
2572 2572 if self._asclist is not None:
2573 2573 return self._asclist.__reversed__
2574 2574 iter1 = self._r1.fastdesc
2575 2575 iter2 = self._r2.fastdesc
2576 2576 if None in (iter1, iter2):
2577 2577 return None
2578 2578 return lambda: self._iterordered(False, iter1(), iter2())
2579 2579
2580 2580 def _iterordered(self, ascending, iter1, iter2):
2581 2581 """produce an ordered iteration from two iterators with the same order
2582 2582
2583 2583 The ascending is used to indicated the iteration direction.
2584 2584 """
2585 2585 choice = max
2586 2586 if ascending:
2587 2587 choice = min
2588 2588
2589 2589 val1 = None
2590 2590 val2 = None
2591 2591
2592 2592 choice = max
2593 2593 if ascending:
2594 2594 choice = min
2595 2595 try:
2596 2596 # Consume both iterators in an ordered way until one is
2597 2597 # empty
2598 2598 while True:
2599 2599 if val1 is None:
2600 2600 val1 = iter1.next()
2601 2601 if val2 is None:
2602 2602 val2 = iter2.next()
2603 2603 next = choice(val1, val2)
2604 2604 yield next
2605 2605 if val1 == next:
2606 2606 val1 = None
2607 2607 if val2 == next:
2608 2608 val2 = None
2609 2609 except StopIteration:
2610 2610 # Flush any remaining values and consume the other one
2611 2611 it = iter2
2612 2612 if val1 is not None:
2613 2613 yield val1
2614 2614 it = iter1
2615 2615 elif val2 is not None:
2616 2616 # might have been equality and both are empty
2617 2617 yield val2
2618 2618 for val in it:
2619 2619 yield val
2620 2620
2621 2621 def __contains__(self, x):
2622 2622 return x in self._r1 or x in self._r2
2623 2623
2624 2624 def sort(self, reverse=False):
2625 2625 """Sort the added set
2626 2626
2627 2627 For this we use the cached list with all the generated values and if we
2628 2628 know they are ascending or descending we can sort them in a smart way.
2629 2629 """
2630 2630 self._ascending = not reverse
2631 2631
2632 2632 def isascending(self):
2633 2633 return self._ascending is not None and self._ascending
2634 2634
2635 2635 def isdescending(self):
2636 2636 return self._ascending is not None and not self._ascending
2637 2637
2638 2638 def reverse(self):
2639 2639 if self._ascending is None:
2640 2640 self._list.reverse()
2641 2641 else:
2642 2642 self._ascending = not self._ascending
2643 2643
2644 2644 def first(self):
2645 2645 for x in self:
2646 2646 return x
2647 2647 return None
2648 2648
2649 2649 def last(self):
2650 2650 self.reverse()
2651 2651 val = self.first()
2652 2652 self.reverse()
2653 2653 return val
2654 2654
2655 2655 class generatorset(abstractsmartset):
2656 2656 """Wrap a generator for lazy iteration
2657 2657
2658 2658 Wrapper structure for generators that provides lazy membership and can
2659 2659 be iterated more than once.
2660 2660 When asked for membership it generates values until either it finds the
2661 2661 requested one or has gone through all the elements in the generator
2662 2662 """
2663 2663 def __init__(self, gen, iterasc=None):
2664 2664 """
2665 2665 gen: a generator producing the values for the generatorset.
2666 2666 """
2667 2667 self._gen = gen
2668 2668 self._asclist = None
2669 2669 self._cache = {}
2670 2670 self._genlist = []
2671 2671 self._finished = False
2672 2672 self._ascending = True
2673 2673 if iterasc is not None:
2674 2674 if iterasc:
2675 2675 self.fastasc = self._iterator
2676 2676 self.__contains__ = self._asccontains
2677 2677 else:
2678 2678 self.fastdesc = self._iterator
2679 2679 self.__contains__ = self._desccontains
2680 2680
2681 2681 def __nonzero__(self):
2682 2682 for r in self:
2683 2683 return True
2684 2684 return False
2685 2685
2686 2686 def __contains__(self, x):
2687 2687 if x in self._cache:
2688 2688 return self._cache[x]
2689 2689
2690 2690 # Use new values only, as existing values would be cached.
2691 2691 for l in self._consumegen():
2692 2692 if l == x:
2693 2693 return True
2694 2694
2695 2695 self._cache[x] = False
2696 2696 return False
2697 2697
2698 2698 def _asccontains(self, x):
2699 2699 """version of contains optimised for ascending generator"""
2700 2700 if x in self._cache:
2701 2701 return self._cache[x]
2702 2702
2703 2703 # Use new values only, as existing values would be cached.
2704 2704 for l in self._consumegen():
2705 2705 if l == x:
2706 2706 return True
2707 2707 if l > x:
2708 2708 break
2709 2709
2710 2710 self._cache[x] = False
2711 2711 return False
2712 2712
2713 2713 def _desccontains(self, x):
2714 2714 """version of contains optimised for descending generator"""
2715 2715 if x in self._cache:
2716 2716 return self._cache[x]
2717 2717
2718 2718 # Use new values only, as existing values would be cached.
2719 2719 for l in self._consumegen():
2720 2720 if l == x:
2721 2721 return True
2722 2722 if l < x:
2723 2723 break
2724 2724
2725 2725 self._cache[x] = False
2726 2726 return False
2727 2727
2728 2728 def __iter__(self):
2729 2729 if self._ascending:
2730 2730 it = self.fastasc
2731 2731 else:
2732 2732 it = self.fastdesc
2733 2733 if it is not None:
2734 2734 return it()
2735 2735 # we need to consume the iterator
2736 2736 for x in self._consumegen():
2737 2737 pass
2738 2738 # recall the same code
2739 2739 return iter(self)
2740 2740
2741 2741 def _iterator(self):
2742 2742 if self._finished:
2743 2743 return iter(self._genlist)
2744 2744
2745 2745 # We have to use this complex iteration strategy to allow multiple
2746 2746 # iterations at the same time. We need to be able to catch revision
2747 2747 # removed from `consumegen` and added to genlist in another instance.
2748 2748 #
2749 2749 # Getting rid of it would provide an about 15% speed up on this
2750 2750 # iteration.
2751 2751 genlist = self._genlist
2752 2752 nextrev = self._consumegen().next
2753 2753 _len = len # cache global lookup
2754 2754 def gen():
2755 2755 i = 0
2756 2756 while True:
2757 2757 if i < _len(genlist):
2758 2758 yield genlist[i]
2759 2759 else:
2760 2760 yield nextrev()
2761 2761 i += 1
2762 2762 return gen()
2763 2763
2764 2764 def _consumegen(self):
2765 2765 cache = self._cache
2766 2766 genlist = self._genlist.append
2767 2767 for item in self._gen:
2768 2768 cache[item] = True
2769 2769 genlist(item)
2770 2770 yield item
2771 2771 if not self._finished:
2772 2772 self._finished = True
2773 2773 asc = self._genlist[:]
2774 2774 asc.sort()
2775 2775 self._asclist = asc
2776 2776 self.fastasc = asc.__iter__
2777 2777 self.fastdesc = asc.__reversed__
2778 2778
2779 2779 def __len__(self):
2780 2780 for x in self._consumegen():
2781 2781 pass
2782 2782 return len(self._genlist)
2783 2783
2784 2784 def sort(self, reverse=False):
2785 2785 self._ascending = not reverse
2786 2786
2787 2787 def reverse(self):
2788 2788 self._ascending = not self._ascending
2789 2789
2790 2790 def isascending(self):
2791 2791 return self._ascending
2792 2792
2793 2793 def isdescending(self):
2794 2794 return not self._ascending
2795 2795
2796 2796 def first(self):
2797 2797 if self._ascending:
2798 2798 it = self.fastasc
2799 2799 else:
2800 2800 it = self.fastdesc
2801 2801 if it is None:
2802 2802 # we need to consume all and try again
2803 2803 for x in self._consumegen():
2804 2804 pass
2805 2805 return self.first()
2806 2806 if self:
2807 2807 return it().next()
2808 2808 return None
2809 2809
2810 2810 def last(self):
2811 2811 if self._ascending:
2812 2812 it = self.fastdesc
2813 2813 else:
2814 2814 it = self.fastasc
2815 2815 if it is None:
2816 2816 # we need to consume all and try again
2817 2817 for x in self._consumegen():
2818 2818 pass
2819 2819 return self.first()
2820 2820 if self:
2821 2821 return it().next()
2822 2822 return None
2823 2823
2824 2824 def spanset(repo, start=None, end=None):
2825 2825 """factory function to dispatch between fullreposet and actual spanset
2826 2826
2827 2827 Feel free to update all spanset call sites and kill this function at some
2828 2828 point.
2829 2829 """
2830 2830 if start is None and end is None:
2831 2831 return fullreposet(repo)
2832 2832 return _spanset(repo, start, end)
2833 2833
2834 2834
2835 2835 class _spanset(abstractsmartset):
2836 2836 """Duck type for baseset class which represents a range of revisions and
2837 2837 can work lazily and without having all the range in memory
2838 2838
2839 2839 Note that spanset(x, y) behave almost like xrange(x, y) except for two
2840 2840 notable points:
2841 2841 - when x < y it will be automatically descending,
2842 2842 - revision filtered with this repoview will be skipped.
2843 2843
2844 2844 """
2845 2845 def __init__(self, repo, start=0, end=None):
2846 2846 """
2847 2847 start: first revision included the set
2848 2848 (default to 0)
2849 2849 end: first revision excluded (last+1)
2850 2850 (default to len(repo)
2851 2851
2852 2852 Spanset will be descending if `end` < `start`.
2853 2853 """
2854 2854 if end is None:
2855 2855 end = len(repo)
2856 2856 self._ascending = start <= end
2857 2857 if not self._ascending:
2858 2858 start, end = end + 1, start +1
2859 2859 self._start = start
2860 2860 self._end = end
2861 2861 self._hiddenrevs = repo.changelog.filteredrevs
2862 2862
2863 2863 def sort(self, reverse=False):
2864 2864 self._ascending = not reverse
2865 2865
2866 2866 def reverse(self):
2867 2867 self._ascending = not self._ascending
2868 2868
2869 2869 def _iterfilter(self, iterrange):
2870 2870 s = self._hiddenrevs
2871 2871 for r in iterrange:
2872 2872 if r not in s:
2873 2873 yield r
2874 2874
2875 2875 def __iter__(self):
2876 2876 if self._ascending:
2877 2877 return self.fastasc()
2878 2878 else:
2879 2879 return self.fastdesc()
2880 2880
2881 2881 def fastasc(self):
2882 2882 iterrange = xrange(self._start, self._end)
2883 2883 if self._hiddenrevs:
2884 2884 return self._iterfilter(iterrange)
2885 2885 return iter(iterrange)
2886 2886
2887 2887 def fastdesc(self):
2888 2888 iterrange = xrange(self._end - 1, self._start - 1, -1)
2889 2889 if self._hiddenrevs:
2890 2890 return self._iterfilter(iterrange)
2891 2891 return iter(iterrange)
2892 2892
2893 2893 def __contains__(self, rev):
2894 2894 hidden = self._hiddenrevs
2895 2895 return ((self._start <= rev < self._end)
2896 2896 and not (hidden and rev in hidden))
2897 2897
2898 2898 def __nonzero__(self):
2899 2899 for r in self:
2900 2900 return True
2901 2901 return False
2902 2902
2903 2903 def __len__(self):
2904 2904 if not self._hiddenrevs:
2905 2905 return abs(self._end - self._start)
2906 2906 else:
2907 2907 count = 0
2908 2908 start = self._start
2909 2909 end = self._end
2910 2910 for rev in self._hiddenrevs:
2911 2911 if (end < rev <= start) or (start <= rev < end):
2912 2912 count += 1
2913 2913 return abs(self._end - self._start) - count
2914 2914
2915 2915 def isascending(self):
2916 2916 return self._start <= self._end
2917 2917
2918 2918 def isdescending(self):
2919 2919 return self._start >= self._end
2920 2920
2921 2921 def first(self):
2922 2922 if self._ascending:
2923 2923 it = self.fastasc
2924 2924 else:
2925 2925 it = self.fastdesc
2926 2926 for x in it():
2927 2927 return x
2928 2928 return None
2929 2929
2930 2930 def last(self):
2931 2931 if self._ascending:
2932 2932 it = self.fastdesc
2933 2933 else:
2934 2934 it = self.fastasc
2935 2935 for x in it():
2936 2936 return x
2937 2937 return None
2938 2938
2939 2939 class fullreposet(_spanset):
2940 2940 """a set containing all revisions in the repo
2941 2941
2942 2942 This class exists to host special optimisation.
2943 2943 """
2944 2944
2945 2945 def __init__(self, repo):
2946 2946 super(fullreposet, self).__init__(repo)
2947 2947
2948 2948 def __and__(self, other):
2949 2949 """fullrepo & other -> other
2950 2950
2951 2951 As self contains the whole repo, all of the other set should also be in
2952 2952 self. Therefor `self & other = other`.
2953 2953
2954 2954 This boldly assumes the other contains valid revs only.
2955 2955 """
2956 2956 # other not a smartset, make is so
2957 2957 if not util.safehasattr(other, 'isascending'):
2958 2958 # filter out hidden revision
2959 2959 # (this boldly assumes all smartset are pure)
2960 2960 #
2961 2961 # `other` was used with "&", let's assume this is a set like
2962 2962 # object.
2963 2963 other = baseset(other - self._hiddenrevs)
2964 2964
2965 2965 if self.isascending():
2966 2966 other.sort()
2967 2967 else:
2968 2968 other.sort(reverse)
2969 2969 return other
2970 2970
2971 2971 # tell hggettext to extract docstrings from these functions:
2972 2972 i18nfunctions = symbols.values()
General Comments 0
You need to be logged in to leave comments. Login now