##// END OF EJS Templates
fullreposet: use `sort` to enforce the order...
Pierre-Yves David -
r22833:6f4109aa default
parent child Browse files
Show More
@@ -1,3034 +1,3030
1 1 # revset.py - revision set queries for mercurial
2 2 #
3 3 # Copyright 2010 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 import re
9 9 import parser, util, error, discovery, hbisect, phases
10 10 import node
11 11 import heapq
12 12 import match as matchmod
13 13 import ancestor as ancestormod
14 14 from i18n import _
15 15 import encoding
16 16 import obsolete as obsmod
17 17 import pathutil
18 18 import repoview
19 19
20 20 def _revancestors(repo, revs, followfirst):
21 21 """Like revlog.ancestors(), but supports followfirst."""
22 22 cut = followfirst and 1 or None
23 23 cl = repo.changelog
24 24
25 25 def iterate():
26 26 revqueue, revsnode = None, None
27 27 h = []
28 28
29 29 revs.sort(reverse=True)
30 30 revqueue = util.deque(revs)
31 31 if revqueue:
32 32 revsnode = revqueue.popleft()
33 33 heapq.heappush(h, -revsnode)
34 34
35 35 seen = set([node.nullrev])
36 36 while h:
37 37 current = -heapq.heappop(h)
38 38 if current not in seen:
39 39 if revsnode and current == revsnode:
40 40 if revqueue:
41 41 revsnode = revqueue.popleft()
42 42 heapq.heappush(h, -revsnode)
43 43 seen.add(current)
44 44 yield current
45 45 for parent in cl.parentrevs(current)[:cut]:
46 46 if parent != node.nullrev:
47 47 heapq.heappush(h, -parent)
48 48
49 49 return generatorset(iterate(), iterasc=False)
50 50
51 51 def _revdescendants(repo, revs, followfirst):
52 52 """Like revlog.descendants() but supports followfirst."""
53 53 cut = followfirst and 1 or None
54 54
55 55 def iterate():
56 56 cl = repo.changelog
57 57 first = min(revs)
58 58 nullrev = node.nullrev
59 59 if first == nullrev:
60 60 # Are there nodes with a null first parent and a non-null
61 61 # second one? Maybe. Do we care? Probably not.
62 62 for i in cl:
63 63 yield i
64 64 else:
65 65 seen = set(revs)
66 66 for i in cl.revs(first + 1):
67 67 for x in cl.parentrevs(i)[:cut]:
68 68 if x != nullrev and x in seen:
69 69 seen.add(i)
70 70 yield i
71 71 break
72 72
73 73 return generatorset(iterate(), iterasc=True)
74 74
75 75 def _revsbetween(repo, roots, heads):
76 76 """Return all paths between roots and heads, inclusive of both endpoint
77 77 sets."""
78 78 if not roots:
79 79 return baseset()
80 80 parentrevs = repo.changelog.parentrevs
81 81 visit = list(heads)
82 82 reachable = set()
83 83 seen = {}
84 84 minroot = min(roots)
85 85 roots = set(roots)
86 86 # open-code the post-order traversal due to the tiny size of
87 87 # sys.getrecursionlimit()
88 88 while visit:
89 89 rev = visit.pop()
90 90 if rev in roots:
91 91 reachable.add(rev)
92 92 parents = parentrevs(rev)
93 93 seen[rev] = parents
94 94 for parent in parents:
95 95 if parent >= minroot and parent not in seen:
96 96 visit.append(parent)
97 97 if not reachable:
98 98 return baseset()
99 99 for rev in sorted(seen):
100 100 for parent in seen[rev]:
101 101 if parent in reachable:
102 102 reachable.add(rev)
103 103 return baseset(sorted(reachable))
104 104
105 105 elements = {
106 106 "(": (20, ("group", 1, ")"), ("func", 1, ")")),
107 107 "~": (18, None, ("ancestor", 18)),
108 108 "^": (18, None, ("parent", 18), ("parentpost", 18)),
109 109 "-": (5, ("negate", 19), ("minus", 5)),
110 110 "::": (17, ("dagrangepre", 17), ("dagrange", 17),
111 111 ("dagrangepost", 17)),
112 112 "..": (17, ("dagrangepre", 17), ("dagrange", 17),
113 113 ("dagrangepost", 17)),
114 114 ":": (15, ("rangepre", 15), ("range", 15), ("rangepost", 15)),
115 115 "not": (10, ("not", 10)),
116 116 "!": (10, ("not", 10)),
117 117 "and": (5, None, ("and", 5)),
118 118 "&": (5, None, ("and", 5)),
119 119 "or": (4, None, ("or", 4)),
120 120 "|": (4, None, ("or", 4)),
121 121 "+": (4, None, ("or", 4)),
122 122 ",": (2, None, ("list", 2)),
123 123 ")": (0, None, None),
124 124 "symbol": (0, ("symbol",), None),
125 125 "string": (0, ("string",), None),
126 126 "end": (0, None, None),
127 127 }
128 128
129 129 keywords = set(['and', 'or', 'not'])
130 130
131 131 def tokenize(program, lookup=None):
132 132 '''
133 133 Parse a revset statement into a stream of tokens
134 134
135 135 Check that @ is a valid unquoted token character (issue3686):
136 136 >>> list(tokenize("@::"))
137 137 [('symbol', '@', 0), ('::', None, 1), ('end', None, 3)]
138 138
139 139 '''
140 140
141 141 pos, l = 0, len(program)
142 142 while pos < l:
143 143 c = program[pos]
144 144 if c.isspace(): # skip inter-token whitespace
145 145 pass
146 146 elif c == ':' and program[pos:pos + 2] == '::': # look ahead carefully
147 147 yield ('::', None, pos)
148 148 pos += 1 # skip ahead
149 149 elif c == '.' and program[pos:pos + 2] == '..': # look ahead carefully
150 150 yield ('..', None, pos)
151 151 pos += 1 # skip ahead
152 152 elif c in "():,-|&+!~^": # handle simple operators
153 153 yield (c, None, pos)
154 154 elif (c in '"\'' or c == 'r' and
155 155 program[pos:pos + 2] in ("r'", 'r"')): # handle quoted strings
156 156 if c == 'r':
157 157 pos += 1
158 158 c = program[pos]
159 159 decode = lambda x: x
160 160 else:
161 161 decode = lambda x: x.decode('string-escape')
162 162 pos += 1
163 163 s = pos
164 164 while pos < l: # find closing quote
165 165 d = program[pos]
166 166 if d == '\\': # skip over escaped characters
167 167 pos += 2
168 168 continue
169 169 if d == c:
170 170 yield ('string', decode(program[s:pos]), s)
171 171 break
172 172 pos += 1
173 173 else:
174 174 raise error.ParseError(_("unterminated string"), s)
175 175 # gather up a symbol/keyword
176 176 elif c.isalnum() or c in '._@' or ord(c) > 127:
177 177 s = pos
178 178 pos += 1
179 179 while pos < l: # find end of symbol
180 180 d = program[pos]
181 181 if not (d.isalnum() or d in "-._/@" or ord(d) > 127):
182 182 break
183 183 if d == '.' and program[pos - 1] == '.': # special case for ..
184 184 pos -= 1
185 185 break
186 186 pos += 1
187 187 sym = program[s:pos]
188 188 if sym in keywords: # operator keywords
189 189 yield (sym, None, s)
190 190 elif '-' in sym:
191 191 # some jerk gave us foo-bar-baz, try to check if it's a symbol
192 192 if lookup and lookup(sym):
193 193 # looks like a real symbol
194 194 yield ('symbol', sym, s)
195 195 else:
196 196 # looks like an expression
197 197 parts = sym.split('-')
198 198 for p in parts[:-1]:
199 199 if p: # possible consecutive -
200 200 yield ('symbol', p, s)
201 201 s += len(p)
202 202 yield ('-', None, pos)
203 203 s += 1
204 204 if parts[-1]: # possible trailing -
205 205 yield ('symbol', parts[-1], s)
206 206 else:
207 207 yield ('symbol', sym, s)
208 208 pos -= 1
209 209 else:
210 210 raise error.ParseError(_("syntax error"), pos)
211 211 pos += 1
212 212 yield ('end', None, pos)
213 213
214 214 # helpers
215 215
216 216 def getstring(x, err):
217 217 if x and (x[0] == 'string' or x[0] == 'symbol'):
218 218 return x[1]
219 219 raise error.ParseError(err)
220 220
221 221 def getlist(x):
222 222 if not x:
223 223 return []
224 224 if x[0] == 'list':
225 225 return getlist(x[1]) + [x[2]]
226 226 return [x]
227 227
228 228 def getargs(x, min, max, err):
229 229 l = getlist(x)
230 230 if len(l) < min or (max >= 0 and len(l) > max):
231 231 raise error.ParseError(err)
232 232 return l
233 233
234 234 def getset(repo, subset, x):
235 235 if not x:
236 236 raise error.ParseError(_("missing argument"))
237 237 s = methods[x[0]](repo, subset, *x[1:])
238 238 if util.safehasattr(s, 'set'):
239 239 return s
240 240 return baseset(s)
241 241
242 242 def _getrevsource(repo, r):
243 243 extra = repo[r].extra()
244 244 for label in ('source', 'transplant_source', 'rebase_source'):
245 245 if label in extra:
246 246 try:
247 247 return repo[extra[label]].rev()
248 248 except error.RepoLookupError:
249 249 pass
250 250 return None
251 251
252 252 # operator methods
253 253
254 254 def stringset(repo, subset, x):
255 255 x = repo[x].rev()
256 256 if x == -1 and len(subset) == len(repo):
257 257 return baseset([-1])
258 258 if len(subset) == len(repo) or x in subset:
259 259 return baseset([x])
260 260 return baseset()
261 261
262 262 def symbolset(repo, subset, x):
263 263 if x in symbols:
264 264 raise error.ParseError(_("can't use %s here") % x)
265 265 return stringset(repo, subset, x)
266 266
267 267 def rangeset(repo, subset, x, y):
268 268 cl = baseset(repo.changelog)
269 269 m = getset(repo, cl, x)
270 270 n = getset(repo, cl, y)
271 271
272 272 if not m or not n:
273 273 return baseset()
274 274 m, n = m.first(), n.last()
275 275
276 276 if m < n:
277 277 r = spanset(repo, m, n + 1)
278 278 else:
279 279 r = spanset(repo, m, n - 1)
280 280 return r & subset
281 281
282 282 def dagrange(repo, subset, x, y):
283 283 r = spanset(repo)
284 284 xs = _revsbetween(repo, getset(repo, r, x), getset(repo, r, y))
285 285 s = subset.set()
286 286 return xs.filter(s.__contains__)
287 287
288 288 def andset(repo, subset, x, y):
289 289 return getset(repo, getset(repo, subset, x), y)
290 290
291 291 def orset(repo, subset, x, y):
292 292 xl = getset(repo, subset, x)
293 293 yl = getset(repo, subset - xl, y)
294 294 return xl + yl
295 295
296 296 def notset(repo, subset, x):
297 297 return subset - getset(repo, subset, x)
298 298
299 299 def listset(repo, subset, a, b):
300 300 raise error.ParseError(_("can't use a list in this context"))
301 301
302 302 def func(repo, subset, a, b):
303 303 if a[0] == 'symbol' and a[1] in symbols:
304 304 return symbols[a[1]](repo, subset, b)
305 305 raise error.ParseError(_("not a function: %s") % a[1])
306 306
307 307 # functions
308 308
309 309 def adds(repo, subset, x):
310 310 """``adds(pattern)``
311 311 Changesets that add a file matching pattern.
312 312
313 313 The pattern without explicit kind like ``glob:`` is expected to be
314 314 relative to the current directory and match against a file or a
315 315 directory.
316 316 """
317 317 # i18n: "adds" is a keyword
318 318 pat = getstring(x, _("adds requires a pattern"))
319 319 return checkstatus(repo, subset, pat, 1)
320 320
321 321 def ancestor(repo, subset, x):
322 322 """``ancestor(*changeset)``
323 323 A greatest common ancestor of the changesets.
324 324
325 325 Accepts 0 or more changesets.
326 326 Will return empty list when passed no args.
327 327 Greatest common ancestor of a single changeset is that changeset.
328 328 """
329 329 # i18n: "ancestor" is a keyword
330 330 l = getlist(x)
331 331 rl = spanset(repo)
332 332 anc = None
333 333
334 334 # (getset(repo, rl, i) for i in l) generates a list of lists
335 335 for revs in (getset(repo, rl, i) for i in l):
336 336 for r in revs:
337 337 if anc is None:
338 338 anc = repo[r]
339 339 else:
340 340 anc = anc.ancestor(repo[r])
341 341
342 342 if anc is not None and anc.rev() in subset:
343 343 return baseset([anc.rev()])
344 344 return baseset()
345 345
346 346 def _ancestors(repo, subset, x, followfirst=False):
347 347 args = getset(repo, spanset(repo), x)
348 348 if not args:
349 349 return baseset()
350 350 s = _revancestors(repo, args, followfirst)
351 351 return subset.filter(s.__contains__)
352 352
353 353 def ancestors(repo, subset, x):
354 354 """``ancestors(set)``
355 355 Changesets that are ancestors of a changeset in set.
356 356 """
357 357 return _ancestors(repo, subset, x)
358 358
359 359 def _firstancestors(repo, subset, x):
360 360 # ``_firstancestors(set)``
361 361 # Like ``ancestors(set)`` but follows only the first parents.
362 362 return _ancestors(repo, subset, x, followfirst=True)
363 363
364 364 def ancestorspec(repo, subset, x, n):
365 365 """``set~n``
366 366 Changesets that are the Nth ancestor (first parents only) of a changeset
367 367 in set.
368 368 """
369 369 try:
370 370 n = int(n[1])
371 371 except (TypeError, ValueError):
372 372 raise error.ParseError(_("~ expects a number"))
373 373 ps = set()
374 374 cl = repo.changelog
375 375 for r in getset(repo, baseset(cl), x):
376 376 for i in range(n):
377 377 r = cl.parentrevs(r)[0]
378 378 ps.add(r)
379 379 return subset & ps
380 380
381 381 def author(repo, subset, x):
382 382 """``author(string)``
383 383 Alias for ``user(string)``.
384 384 """
385 385 # i18n: "author" is a keyword
386 386 n = encoding.lower(getstring(x, _("author requires a string")))
387 387 kind, pattern, matcher = _substringmatcher(n)
388 388 return subset.filter(lambda x: matcher(encoding.lower(repo[x].user())))
389 389
390 390 def only(repo, subset, x):
391 391 """``only(set, [set])``
392 392 Changesets that are ancestors of the first set that are not ancestors
393 393 of any other head in the repo. If a second set is specified, the result
394 394 is ancestors of the first set that are not ancestors of the second set
395 395 (i.e. ::<set1> - ::<set2>).
396 396 """
397 397 cl = repo.changelog
398 398 # i18n: "only" is a keyword
399 399 args = getargs(x, 1, 2, _('only takes one or two arguments'))
400 400 include = getset(repo, spanset(repo), args[0]).set()
401 401 if len(args) == 1:
402 402 if len(include) == 0:
403 403 return baseset()
404 404
405 405 descendants = set(_revdescendants(repo, include, False))
406 406 exclude = [rev for rev in cl.headrevs()
407 407 if not rev in descendants and not rev in include]
408 408 else:
409 409 exclude = getset(repo, spanset(repo), args[1])
410 410
411 411 results = set(ancestormod.missingancestors(include, exclude, cl.parentrevs))
412 412 return filteredset(subset, results.__contains__)
413 413
414 414 def bisect(repo, subset, x):
415 415 """``bisect(string)``
416 416 Changesets marked in the specified bisect status:
417 417
418 418 - ``good``, ``bad``, ``skip``: csets explicitly marked as good/bad/skip
419 419 - ``goods``, ``bads`` : csets topologically good/bad
420 420 - ``range`` : csets taking part in the bisection
421 421 - ``pruned`` : csets that are goods, bads or skipped
422 422 - ``untested`` : csets whose fate is yet unknown
423 423 - ``ignored`` : csets ignored due to DAG topology
424 424 - ``current`` : the cset currently being bisected
425 425 """
426 426 # i18n: "bisect" is a keyword
427 427 status = getstring(x, _("bisect requires a string")).lower()
428 428 state = set(hbisect.get(repo, status))
429 429 return subset & state
430 430
431 431 # Backward-compatibility
432 432 # - no help entry so that we do not advertise it any more
433 433 def bisected(repo, subset, x):
434 434 return bisect(repo, subset, x)
435 435
436 436 def bookmark(repo, subset, x):
437 437 """``bookmark([name])``
438 438 The named bookmark or all bookmarks.
439 439
440 440 If `name` starts with `re:`, the remainder of the name is treated as
441 441 a regular expression. To match a bookmark that actually starts with `re:`,
442 442 use the prefix `literal:`.
443 443 """
444 444 # i18n: "bookmark" is a keyword
445 445 args = getargs(x, 0, 1, _('bookmark takes one or no arguments'))
446 446 if args:
447 447 bm = getstring(args[0],
448 448 # i18n: "bookmark" is a keyword
449 449 _('the argument to bookmark must be a string'))
450 450 kind, pattern, matcher = _stringmatcher(bm)
451 451 bms = set()
452 452 if kind == 'literal':
453 453 bmrev = repo._bookmarks.get(pattern, None)
454 454 if not bmrev:
455 455 raise util.Abort(_("bookmark '%s' does not exist") % bm)
456 456 bms.add(repo[bmrev].rev())
457 457 else:
458 458 matchrevs = set()
459 459 for name, bmrev in repo._bookmarks.iteritems():
460 460 if matcher(name):
461 461 matchrevs.add(bmrev)
462 462 if not matchrevs:
463 463 raise util.Abort(_("no bookmarks exist that match '%s'")
464 464 % pattern)
465 465 for bmrev in matchrevs:
466 466 bms.add(repo[bmrev].rev())
467 467 else:
468 468 bms = set([repo[r].rev()
469 469 for r in repo._bookmarks.values()])
470 470 bms -= set([node.nullrev])
471 471 return subset & bms
472 472
473 473 def branch(repo, subset, x):
474 474 """``branch(string or set)``
475 475 All changesets belonging to the given branch or the branches of the given
476 476 changesets.
477 477
478 478 If `string` starts with `re:`, the remainder of the name is treated as
479 479 a regular expression. To match a branch that actually starts with `re:`,
480 480 use the prefix `literal:`.
481 481 """
482 482 try:
483 483 b = getstring(x, '')
484 484 except error.ParseError:
485 485 # not a string, but another revspec, e.g. tip()
486 486 pass
487 487 else:
488 488 kind, pattern, matcher = _stringmatcher(b)
489 489 if kind == 'literal':
490 490 # note: falls through to the revspec case if no branch with
491 491 # this name exists
492 492 if pattern in repo.branchmap():
493 493 return subset.filter(lambda r: matcher(repo[r].branch()))
494 494 else:
495 495 return subset.filter(lambda r: matcher(repo[r].branch()))
496 496
497 497 s = getset(repo, spanset(repo), x)
498 498 b = set()
499 499 for r in s:
500 500 b.add(repo[r].branch())
501 501 s = s.set()
502 502 return subset.filter(lambda r: r in s or repo[r].branch() in b)
503 503
504 504 def bumped(repo, subset, x):
505 505 """``bumped()``
506 506 Mutable changesets marked as successors of public changesets.
507 507
508 508 Only non-public and non-obsolete changesets can be `bumped`.
509 509 """
510 510 # i18n: "bumped" is a keyword
511 511 getargs(x, 0, 0, _("bumped takes no arguments"))
512 512 bumped = obsmod.getrevs(repo, 'bumped')
513 513 return subset & bumped
514 514
515 515 def bundle(repo, subset, x):
516 516 """``bundle()``
517 517 Changesets in the bundle.
518 518
519 519 Bundle must be specified by the -R option."""
520 520
521 521 try:
522 522 bundlerevs = repo.changelog.bundlerevs
523 523 except AttributeError:
524 524 raise util.Abort(_("no bundle provided - specify with -R"))
525 525 return subset & bundlerevs
526 526
527 527 def checkstatus(repo, subset, pat, field):
528 528 hasset = matchmod.patkind(pat) == 'set'
529 529
530 530 def matches(x):
531 531 m = None
532 532 fname = None
533 533 c = repo[x]
534 534 if not m or hasset:
535 535 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
536 536 if not m.anypats() and len(m.files()) == 1:
537 537 fname = m.files()[0]
538 538 if fname is not None:
539 539 if fname not in c.files():
540 540 return False
541 541 else:
542 542 for f in c.files():
543 543 if m(f):
544 544 break
545 545 else:
546 546 return False
547 547 files = repo.status(c.p1().node(), c.node())[field]
548 548 if fname is not None:
549 549 if fname in files:
550 550 return True
551 551 else:
552 552 for f in files:
553 553 if m(f):
554 554 return True
555 555
556 556 return subset.filter(matches)
557 557
558 558 def _children(repo, narrow, parentset):
559 559 cs = set()
560 560 if not parentset:
561 561 return baseset(cs)
562 562 pr = repo.changelog.parentrevs
563 563 minrev = min(parentset)
564 564 for r in narrow:
565 565 if r <= minrev:
566 566 continue
567 567 for p in pr(r):
568 568 if p in parentset:
569 569 cs.add(r)
570 570 return baseset(cs)
571 571
572 572 def children(repo, subset, x):
573 573 """``children(set)``
574 574 Child changesets of changesets in set.
575 575 """
576 576 s = getset(repo, baseset(repo), x).set()
577 577 cs = _children(repo, subset, s)
578 578 return subset & cs
579 579
580 580 def closed(repo, subset, x):
581 581 """``closed()``
582 582 Changeset is closed.
583 583 """
584 584 # i18n: "closed" is a keyword
585 585 getargs(x, 0, 0, _("closed takes no arguments"))
586 586 return subset.filter(lambda r: repo[r].closesbranch())
587 587
588 588 def contains(repo, subset, x):
589 589 """``contains(pattern)``
590 590 The revision's manifest contains a file matching pattern (but might not
591 591 modify it). See :hg:`help patterns` for information about file patterns.
592 592
593 593 The pattern without explicit kind like ``glob:`` is expected to be
594 594 relative to the current directory and match against a file exactly
595 595 for efficiency.
596 596 """
597 597 # i18n: "contains" is a keyword
598 598 pat = getstring(x, _("contains requires a pattern"))
599 599
600 600 def matches(x):
601 601 if not matchmod.patkind(pat):
602 602 pats = pathutil.canonpath(repo.root, repo.getcwd(), pat)
603 603 if pats in repo[x]:
604 604 return True
605 605 else:
606 606 c = repo[x]
607 607 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
608 608 for f in c.manifest():
609 609 if m(f):
610 610 return True
611 611 return False
612 612
613 613 return subset.filter(matches)
614 614
615 615 def converted(repo, subset, x):
616 616 """``converted([id])``
617 617 Changesets converted from the given identifier in the old repository if
618 618 present, or all converted changesets if no identifier is specified.
619 619 """
620 620
621 621 # There is exactly no chance of resolving the revision, so do a simple
622 622 # string compare and hope for the best
623 623
624 624 rev = None
625 625 # i18n: "converted" is a keyword
626 626 l = getargs(x, 0, 1, _('converted takes one or no arguments'))
627 627 if l:
628 628 # i18n: "converted" is a keyword
629 629 rev = getstring(l[0], _('converted requires a revision'))
630 630
631 631 def _matchvalue(r):
632 632 source = repo[r].extra().get('convert_revision', None)
633 633 return source is not None and (rev is None or source.startswith(rev))
634 634
635 635 return subset.filter(lambda r: _matchvalue(r))
636 636
637 637 def date(repo, subset, x):
638 638 """``date(interval)``
639 639 Changesets within the interval, see :hg:`help dates`.
640 640 """
641 641 # i18n: "date" is a keyword
642 642 ds = getstring(x, _("date requires a string"))
643 643 dm = util.matchdate(ds)
644 644 return subset.filter(lambda x: dm(repo[x].date()[0]))
645 645
646 646 def desc(repo, subset, x):
647 647 """``desc(string)``
648 648 Search commit message for string. The match is case-insensitive.
649 649 """
650 650 # i18n: "desc" is a keyword
651 651 ds = encoding.lower(getstring(x, _("desc requires a string")))
652 652
653 653 def matches(x):
654 654 c = repo[x]
655 655 return ds in encoding.lower(c.description())
656 656
657 657 return subset.filter(matches)
658 658
659 659 def _descendants(repo, subset, x, followfirst=False):
660 660 args = getset(repo, spanset(repo), x)
661 661 if not args:
662 662 return baseset()
663 663 s = _revdescendants(repo, args, followfirst)
664 664
665 665 # Both sets need to be ascending in order to lazily return the union
666 666 # in the correct order.
667 667 args.sort()
668 668 result = (filteredset(s, subset.__contains__, ascending=True) +
669 669 filteredset(args, subset.__contains__, ascending=True))
670 670
671 671 return result
672 672
673 673 def descendants(repo, subset, x):
674 674 """``descendants(set)``
675 675 Changesets which are descendants of changesets in set.
676 676 """
677 677 return _descendants(repo, subset, x)
678 678
679 679 def _firstdescendants(repo, subset, x):
680 680 # ``_firstdescendants(set)``
681 681 # Like ``descendants(set)`` but follows only the first parents.
682 682 return _descendants(repo, subset, x, followfirst=True)
683 683
684 684 def destination(repo, subset, x):
685 685 """``destination([set])``
686 686 Changesets that were created by a graft, transplant or rebase operation,
687 687 with the given revisions specified as the source. Omitting the optional set
688 688 is the same as passing all().
689 689 """
690 690 if x is not None:
691 691 args = getset(repo, spanset(repo), x).set()
692 692 else:
693 693 args = getall(repo, spanset(repo), x).set()
694 694
695 695 dests = set()
696 696
697 697 # subset contains all of the possible destinations that can be returned, so
698 698 # iterate over them and see if their source(s) were provided in the args.
699 699 # Even if the immediate src of r is not in the args, src's source (or
700 700 # further back) may be. Scanning back further than the immediate src allows
701 701 # transitive transplants and rebases to yield the same results as transitive
702 702 # grafts.
703 703 for r in subset:
704 704 src = _getrevsource(repo, r)
705 705 lineage = None
706 706
707 707 while src is not None:
708 708 if lineage is None:
709 709 lineage = list()
710 710
711 711 lineage.append(r)
712 712
713 713 # The visited lineage is a match if the current source is in the arg
714 714 # set. Since every candidate dest is visited by way of iterating
715 715 # subset, any dests further back in the lineage will be tested by a
716 716 # different iteration over subset. Likewise, if the src was already
717 717 # selected, the current lineage can be selected without going back
718 718 # further.
719 719 if src in args or src in dests:
720 720 dests.update(lineage)
721 721 break
722 722
723 723 r = src
724 724 src = _getrevsource(repo, r)
725 725
726 726 return subset.filter(dests.__contains__)
727 727
728 728 def divergent(repo, subset, x):
729 729 """``divergent()``
730 730 Final successors of changesets with an alternative set of final successors.
731 731 """
732 732 # i18n: "divergent" is a keyword
733 733 getargs(x, 0, 0, _("divergent takes no arguments"))
734 734 divergent = obsmod.getrevs(repo, 'divergent')
735 735 return subset & divergent
736 736
737 737 def draft(repo, subset, x):
738 738 """``draft()``
739 739 Changeset in draft phase."""
740 740 # i18n: "draft" is a keyword
741 741 getargs(x, 0, 0, _("draft takes no arguments"))
742 742 pc = repo._phasecache
743 743 return subset.filter(lambda r: pc.phase(repo, r) == phases.draft)
744 744
745 745 def extinct(repo, subset, x):
746 746 """``extinct()``
747 747 Obsolete changesets with obsolete descendants only.
748 748 """
749 749 # i18n: "extinct" is a keyword
750 750 getargs(x, 0, 0, _("extinct takes no arguments"))
751 751 extincts = obsmod.getrevs(repo, 'extinct')
752 752 return subset & extincts
753 753
754 754 def extra(repo, subset, x):
755 755 """``extra(label, [value])``
756 756 Changesets with the given label in the extra metadata, with the given
757 757 optional value.
758 758
759 759 If `value` starts with `re:`, the remainder of the value is treated as
760 760 a regular expression. To match a value that actually starts with `re:`,
761 761 use the prefix `literal:`.
762 762 """
763 763
764 764 # i18n: "extra" is a keyword
765 765 l = getargs(x, 1, 2, _('extra takes at least 1 and at most 2 arguments'))
766 766 # i18n: "extra" is a keyword
767 767 label = getstring(l[0], _('first argument to extra must be a string'))
768 768 value = None
769 769
770 770 if len(l) > 1:
771 771 # i18n: "extra" is a keyword
772 772 value = getstring(l[1], _('second argument to extra must be a string'))
773 773 kind, value, matcher = _stringmatcher(value)
774 774
775 775 def _matchvalue(r):
776 776 extra = repo[r].extra()
777 777 return label in extra and (value is None or matcher(extra[label]))
778 778
779 779 return subset.filter(lambda r: _matchvalue(r))
780 780
781 781 def filelog(repo, subset, x):
782 782 """``filelog(pattern)``
783 783 Changesets connected to the specified filelog.
784 784
785 785 For performance reasons, visits only revisions mentioned in the file-level
786 786 filelog, rather than filtering through all changesets (much faster, but
787 787 doesn't include deletes or duplicate changes). For a slower, more accurate
788 788 result, use ``file()``.
789 789
790 790 The pattern without explicit kind like ``glob:`` is expected to be
791 791 relative to the current directory and match against a file exactly
792 792 for efficiency.
793 793 """
794 794
795 795 # i18n: "filelog" is a keyword
796 796 pat = getstring(x, _("filelog requires a pattern"))
797 797 s = set()
798 798
799 799 if not matchmod.patkind(pat):
800 800 f = pathutil.canonpath(repo.root, repo.getcwd(), pat)
801 801 fl = repo.file(f)
802 802 for fr in fl:
803 803 s.add(fl.linkrev(fr))
804 804 else:
805 805 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=repo[None])
806 806 for f in repo[None]:
807 807 if m(f):
808 808 fl = repo.file(f)
809 809 for fr in fl:
810 810 s.add(fl.linkrev(fr))
811 811
812 812 return subset & s
813 813
814 814 def first(repo, subset, x):
815 815 """``first(set, [n])``
816 816 An alias for limit().
817 817 """
818 818 return limit(repo, subset, x)
819 819
820 820 def _follow(repo, subset, x, name, followfirst=False):
821 821 l = getargs(x, 0, 1, _("%s takes no arguments or a filename") % name)
822 822 c = repo['.']
823 823 if l:
824 824 x = getstring(l[0], _("%s expected a filename") % name)
825 825 if x in c:
826 826 cx = c[x]
827 827 s = set(ctx.rev() for ctx in cx.ancestors(followfirst=followfirst))
828 828 # include the revision responsible for the most recent version
829 829 s.add(cx.linkrev())
830 830 else:
831 831 return baseset()
832 832 else:
833 833 s = _revancestors(repo, baseset([c.rev()]), followfirst)
834 834
835 835 return subset & s
836 836
837 837 def follow(repo, subset, x):
838 838 """``follow([file])``
839 839 An alias for ``::.`` (ancestors of the working copy's first parent).
840 840 If a filename is specified, the history of the given file is followed,
841 841 including copies.
842 842 """
843 843 return _follow(repo, subset, x, 'follow')
844 844
845 845 def _followfirst(repo, subset, x):
846 846 # ``followfirst([file])``
847 847 # Like ``follow([file])`` but follows only the first parent of
848 848 # every revision or file revision.
849 849 return _follow(repo, subset, x, '_followfirst', followfirst=True)
850 850
851 851 def getall(repo, subset, x):
852 852 """``all()``
853 853 All changesets, the same as ``0:tip``.
854 854 """
855 855 # i18n: "all" is a keyword
856 856 getargs(x, 0, 0, _("all takes no arguments"))
857 857 return subset
858 858
859 859 def grep(repo, subset, x):
860 860 """``grep(regex)``
861 861 Like ``keyword(string)`` but accepts a regex. Use ``grep(r'...')``
862 862 to ensure special escape characters are handled correctly. Unlike
863 863 ``keyword(string)``, the match is case-sensitive.
864 864 """
865 865 try:
866 866 # i18n: "grep" is a keyword
867 867 gr = re.compile(getstring(x, _("grep requires a string")))
868 868 except re.error, e:
869 869 raise error.ParseError(_('invalid match pattern: %s') % e)
870 870
871 871 def matches(x):
872 872 c = repo[x]
873 873 for e in c.files() + [c.user(), c.description()]:
874 874 if gr.search(e):
875 875 return True
876 876 return False
877 877
878 878 return subset.filter(matches)
879 879
880 880 def _matchfiles(repo, subset, x):
881 881 # _matchfiles takes a revset list of prefixed arguments:
882 882 #
883 883 # [p:foo, i:bar, x:baz]
884 884 #
885 885 # builds a match object from them and filters subset. Allowed
886 886 # prefixes are 'p:' for regular patterns, 'i:' for include
887 887 # patterns and 'x:' for exclude patterns. Use 'r:' prefix to pass
888 888 # a revision identifier, or the empty string to reference the
889 889 # working directory, from which the match object is
890 890 # initialized. Use 'd:' to set the default matching mode, default
891 891 # to 'glob'. At most one 'r:' and 'd:' argument can be passed.
892 892
893 893 # i18n: "_matchfiles" is a keyword
894 894 l = getargs(x, 1, -1, _("_matchfiles requires at least one argument"))
895 895 pats, inc, exc = [], [], []
896 896 hasset = False
897 897 rev, default = None, None
898 898 for arg in l:
899 899 # i18n: "_matchfiles" is a keyword
900 900 s = getstring(arg, _("_matchfiles requires string arguments"))
901 901 prefix, value = s[:2], s[2:]
902 902 if prefix == 'p:':
903 903 pats.append(value)
904 904 elif prefix == 'i:':
905 905 inc.append(value)
906 906 elif prefix == 'x:':
907 907 exc.append(value)
908 908 elif prefix == 'r:':
909 909 if rev is not None:
910 910 # i18n: "_matchfiles" is a keyword
911 911 raise error.ParseError(_('_matchfiles expected at most one '
912 912 'revision'))
913 913 rev = value
914 914 elif prefix == 'd:':
915 915 if default is not None:
916 916 # i18n: "_matchfiles" is a keyword
917 917 raise error.ParseError(_('_matchfiles expected at most one '
918 918 'default mode'))
919 919 default = value
920 920 else:
921 921 # i18n: "_matchfiles" is a keyword
922 922 raise error.ParseError(_('invalid _matchfiles prefix: %s') % prefix)
923 923 if not hasset and matchmod.patkind(value) == 'set':
924 924 hasset = True
925 925 if not default:
926 926 default = 'glob'
927 927
928 928 def matches(x):
929 929 m = None
930 930 c = repo[x]
931 931 if not m or (hasset and rev is None):
932 932 ctx = c
933 933 if rev is not None:
934 934 ctx = repo[rev or None]
935 935 m = matchmod.match(repo.root, repo.getcwd(), pats, include=inc,
936 936 exclude=exc, ctx=ctx, default=default)
937 937 for f in c.files():
938 938 if m(f):
939 939 return True
940 940 return False
941 941
942 942 return subset.filter(matches)
943 943
944 944 def hasfile(repo, subset, x):
945 945 """``file(pattern)``
946 946 Changesets affecting files matched by pattern.
947 947
948 948 For a faster but less accurate result, consider using ``filelog()``
949 949 instead.
950 950
951 951 This predicate uses ``glob:`` as the default kind of pattern.
952 952 """
953 953 # i18n: "file" is a keyword
954 954 pat = getstring(x, _("file requires a pattern"))
955 955 return _matchfiles(repo, subset, ('string', 'p:' + pat))
956 956
957 957 def head(repo, subset, x):
958 958 """``head()``
959 959 Changeset is a named branch head.
960 960 """
961 961 # i18n: "head" is a keyword
962 962 getargs(x, 0, 0, _("head takes no arguments"))
963 963 hs = set()
964 964 for b, ls in repo.branchmap().iteritems():
965 965 hs.update(repo[h].rev() for h in ls)
966 966 return baseset(hs).filter(subset.__contains__)
967 967
968 968 def heads(repo, subset, x):
969 969 """``heads(set)``
970 970 Members of set with no children in set.
971 971 """
972 972 s = getset(repo, subset, x)
973 973 ps = parents(repo, subset, x)
974 974 return s - ps
975 975
976 976 def hidden(repo, subset, x):
977 977 """``hidden()``
978 978 Hidden changesets.
979 979 """
980 980 # i18n: "hidden" is a keyword
981 981 getargs(x, 0, 0, _("hidden takes no arguments"))
982 982 hiddenrevs = repoview.filterrevs(repo, 'visible')
983 983 return subset & hiddenrevs
984 984
985 985 def keyword(repo, subset, x):
986 986 """``keyword(string)``
987 987 Search commit message, user name, and names of changed files for
988 988 string. The match is case-insensitive.
989 989 """
990 990 # i18n: "keyword" is a keyword
991 991 kw = encoding.lower(getstring(x, _("keyword requires a string")))
992 992
993 993 def matches(r):
994 994 c = repo[r]
995 995 return util.any(kw in encoding.lower(t) for t in c.files() + [c.user(),
996 996 c.description()])
997 997
998 998 return subset.filter(matches)
999 999
1000 1000 def limit(repo, subset, x):
1001 1001 """``limit(set, [n])``
1002 1002 First n members of set, defaulting to 1.
1003 1003 """
1004 1004 # i18n: "limit" is a keyword
1005 1005 l = getargs(x, 1, 2, _("limit requires one or two arguments"))
1006 1006 try:
1007 1007 lim = 1
1008 1008 if len(l) == 2:
1009 1009 # i18n: "limit" is a keyword
1010 1010 lim = int(getstring(l[1], _("limit requires a number")))
1011 1011 except (TypeError, ValueError):
1012 1012 # i18n: "limit" is a keyword
1013 1013 raise error.ParseError(_("limit expects a number"))
1014 1014 ss = subset.set()
1015 1015 os = getset(repo, spanset(repo), l[0])
1016 1016 result = []
1017 1017 it = iter(os)
1018 1018 for x in xrange(lim):
1019 1019 try:
1020 1020 y = it.next()
1021 1021 if y in ss:
1022 1022 result.append(y)
1023 1023 except (StopIteration):
1024 1024 break
1025 1025 return baseset(result)
1026 1026
1027 1027 def last(repo, subset, x):
1028 1028 """``last(set, [n])``
1029 1029 Last n members of set, defaulting to 1.
1030 1030 """
1031 1031 # i18n: "last" is a keyword
1032 1032 l = getargs(x, 1, 2, _("last requires one or two arguments"))
1033 1033 try:
1034 1034 lim = 1
1035 1035 if len(l) == 2:
1036 1036 # i18n: "last" is a keyword
1037 1037 lim = int(getstring(l[1], _("last requires a number")))
1038 1038 except (TypeError, ValueError):
1039 1039 # i18n: "last" is a keyword
1040 1040 raise error.ParseError(_("last expects a number"))
1041 1041 ss = subset.set()
1042 1042 os = getset(repo, spanset(repo), l[0])
1043 1043 os.reverse()
1044 1044 result = []
1045 1045 it = iter(os)
1046 1046 for x in xrange(lim):
1047 1047 try:
1048 1048 y = it.next()
1049 1049 if y in ss:
1050 1050 result.append(y)
1051 1051 except (StopIteration):
1052 1052 break
1053 1053 return baseset(result)
1054 1054
1055 1055 def maxrev(repo, subset, x):
1056 1056 """``max(set)``
1057 1057 Changeset with highest revision number in set.
1058 1058 """
1059 1059 os = getset(repo, spanset(repo), x)
1060 1060 if os:
1061 1061 m = os.max()
1062 1062 if m in subset:
1063 1063 return baseset([m])
1064 1064 return baseset()
1065 1065
1066 1066 def merge(repo, subset, x):
1067 1067 """``merge()``
1068 1068 Changeset is a merge changeset.
1069 1069 """
1070 1070 # i18n: "merge" is a keyword
1071 1071 getargs(x, 0, 0, _("merge takes no arguments"))
1072 1072 cl = repo.changelog
1073 1073 return subset.filter(lambda r: cl.parentrevs(r)[1] != -1)
1074 1074
1075 1075 def branchpoint(repo, subset, x):
1076 1076 """``branchpoint()``
1077 1077 Changesets with more than one child.
1078 1078 """
1079 1079 # i18n: "branchpoint" is a keyword
1080 1080 getargs(x, 0, 0, _("branchpoint takes no arguments"))
1081 1081 cl = repo.changelog
1082 1082 if not subset:
1083 1083 return baseset()
1084 1084 baserev = min(subset)
1085 1085 parentscount = [0]*(len(repo) - baserev)
1086 1086 for r in cl.revs(start=baserev + 1):
1087 1087 for p in cl.parentrevs(r):
1088 1088 if p >= baserev:
1089 1089 parentscount[p - baserev] += 1
1090 1090 return subset.filter(lambda r: parentscount[r - baserev] > 1)
1091 1091
1092 1092 def minrev(repo, subset, x):
1093 1093 """``min(set)``
1094 1094 Changeset with lowest revision number in set.
1095 1095 """
1096 1096 os = getset(repo, spanset(repo), x)
1097 1097 if os:
1098 1098 m = os.min()
1099 1099 if m in subset:
1100 1100 return baseset([m])
1101 1101 return baseset()
1102 1102
1103 1103 def modifies(repo, subset, x):
1104 1104 """``modifies(pattern)``
1105 1105 Changesets modifying files matched by pattern.
1106 1106
1107 1107 The pattern without explicit kind like ``glob:`` is expected to be
1108 1108 relative to the current directory and match against a file or a
1109 1109 directory.
1110 1110 """
1111 1111 # i18n: "modifies" is a keyword
1112 1112 pat = getstring(x, _("modifies requires a pattern"))
1113 1113 return checkstatus(repo, subset, pat, 0)
1114 1114
1115 1115 def node_(repo, subset, x):
1116 1116 """``id(string)``
1117 1117 Revision non-ambiguously specified by the given hex string prefix.
1118 1118 """
1119 1119 # i18n: "id" is a keyword
1120 1120 l = getargs(x, 1, 1, _("id requires one argument"))
1121 1121 # i18n: "id" is a keyword
1122 1122 n = getstring(l[0], _("id requires a string"))
1123 1123 if len(n) == 40:
1124 1124 rn = repo[n].rev()
1125 1125 else:
1126 1126 rn = None
1127 1127 pm = repo.changelog._partialmatch(n)
1128 1128 if pm is not None:
1129 1129 rn = repo.changelog.rev(pm)
1130 1130
1131 1131 return subset.filter(lambda r: r == rn)
1132 1132
1133 1133 def obsolete(repo, subset, x):
1134 1134 """``obsolete()``
1135 1135 Mutable changeset with a newer version."""
1136 1136 # i18n: "obsolete" is a keyword
1137 1137 getargs(x, 0, 0, _("obsolete takes no arguments"))
1138 1138 obsoletes = obsmod.getrevs(repo, 'obsolete')
1139 1139 return subset & obsoletes
1140 1140
1141 1141 def origin(repo, subset, x):
1142 1142 """``origin([set])``
1143 1143 Changesets that were specified as a source for the grafts, transplants or
1144 1144 rebases that created the given revisions. Omitting the optional set is the
1145 1145 same as passing all(). If a changeset created by these operations is itself
1146 1146 specified as a source for one of these operations, only the source changeset
1147 1147 for the first operation is selected.
1148 1148 """
1149 1149 if x is not None:
1150 1150 args = getset(repo, spanset(repo), x).set()
1151 1151 else:
1152 1152 args = getall(repo, spanset(repo), x).set()
1153 1153
1154 1154 def _firstsrc(rev):
1155 1155 src = _getrevsource(repo, rev)
1156 1156 if src is None:
1157 1157 return None
1158 1158
1159 1159 while True:
1160 1160 prev = _getrevsource(repo, src)
1161 1161
1162 1162 if prev is None:
1163 1163 return src
1164 1164 src = prev
1165 1165
1166 1166 o = set([_firstsrc(r) for r in args])
1167 1167 o -= set([None])
1168 1168 return subset & o
1169 1169
1170 1170 def outgoing(repo, subset, x):
1171 1171 """``outgoing([path])``
1172 1172 Changesets not found in the specified destination repository, or the
1173 1173 default push location.
1174 1174 """
1175 1175 import hg # avoid start-up nasties
1176 1176 # i18n: "outgoing" is a keyword
1177 1177 l = getargs(x, 0, 1, _("outgoing takes one or no arguments"))
1178 1178 # i18n: "outgoing" is a keyword
1179 1179 dest = l and getstring(l[0], _("outgoing requires a repository path")) or ''
1180 1180 dest = repo.ui.expandpath(dest or 'default-push', dest or 'default')
1181 1181 dest, branches = hg.parseurl(dest)
1182 1182 revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
1183 1183 if revs:
1184 1184 revs = [repo.lookup(rev) for rev in revs]
1185 1185 other = hg.peer(repo, {}, dest)
1186 1186 repo.ui.pushbuffer()
1187 1187 outgoing = discovery.findcommonoutgoing(repo, other, onlyheads=revs)
1188 1188 repo.ui.popbuffer()
1189 1189 cl = repo.changelog
1190 1190 o = set([cl.rev(r) for r in outgoing.missing])
1191 1191 return subset & o
1192 1192
1193 1193 def p1(repo, subset, x):
1194 1194 """``p1([set])``
1195 1195 First parent of changesets in set, or the working directory.
1196 1196 """
1197 1197 if x is None:
1198 1198 p = repo[x].p1().rev()
1199 1199 if p >= 0:
1200 1200 return subset & baseset([p])
1201 1201 return baseset()
1202 1202
1203 1203 ps = set()
1204 1204 cl = repo.changelog
1205 1205 for r in getset(repo, spanset(repo), x):
1206 1206 ps.add(cl.parentrevs(r)[0])
1207 1207 ps -= set([node.nullrev])
1208 1208 return subset & ps
1209 1209
1210 1210 def p2(repo, subset, x):
1211 1211 """``p2([set])``
1212 1212 Second parent of changesets in set, or the working directory.
1213 1213 """
1214 1214 if x is None:
1215 1215 ps = repo[x].parents()
1216 1216 try:
1217 1217 p = ps[1].rev()
1218 1218 if p >= 0:
1219 1219 return subset & baseset([p])
1220 1220 return baseset()
1221 1221 except IndexError:
1222 1222 return baseset()
1223 1223
1224 1224 ps = set()
1225 1225 cl = repo.changelog
1226 1226 for r in getset(repo, spanset(repo), x):
1227 1227 ps.add(cl.parentrevs(r)[1])
1228 1228 ps -= set([node.nullrev])
1229 1229 return subset & ps
1230 1230
1231 1231 def parents(repo, subset, x):
1232 1232 """``parents([set])``
1233 1233 The set of all parents for all changesets in set, or the working directory.
1234 1234 """
1235 1235 if x is None:
1236 1236 ps = set(p.rev() for p in repo[x].parents())
1237 1237 else:
1238 1238 ps = set()
1239 1239 cl = repo.changelog
1240 1240 for r in getset(repo, spanset(repo), x):
1241 1241 ps.update(cl.parentrevs(r))
1242 1242 ps -= set([node.nullrev])
1243 1243 return subset & ps
1244 1244
1245 1245 def parentspec(repo, subset, x, n):
1246 1246 """``set^0``
1247 1247 The set.
1248 1248 ``set^1`` (or ``set^``), ``set^2``
1249 1249 First or second parent, respectively, of all changesets in set.
1250 1250 """
1251 1251 try:
1252 1252 n = int(n[1])
1253 1253 if n not in (0, 1, 2):
1254 1254 raise ValueError
1255 1255 except (TypeError, ValueError):
1256 1256 raise error.ParseError(_("^ expects a number 0, 1, or 2"))
1257 1257 ps = set()
1258 1258 cl = repo.changelog
1259 1259 for r in getset(repo, baseset(cl), x):
1260 1260 if n == 0:
1261 1261 ps.add(r)
1262 1262 elif n == 1:
1263 1263 ps.add(cl.parentrevs(r)[0])
1264 1264 elif n == 2:
1265 1265 parents = cl.parentrevs(r)
1266 1266 if len(parents) > 1:
1267 1267 ps.add(parents[1])
1268 1268 return subset & ps
1269 1269
1270 1270 def present(repo, subset, x):
1271 1271 """``present(set)``
1272 1272 An empty set, if any revision in set isn't found; otherwise,
1273 1273 all revisions in set.
1274 1274
1275 1275 If any of specified revisions is not present in the local repository,
1276 1276 the query is normally aborted. But this predicate allows the query
1277 1277 to continue even in such cases.
1278 1278 """
1279 1279 try:
1280 1280 return getset(repo, subset, x)
1281 1281 except error.RepoLookupError:
1282 1282 return baseset()
1283 1283
1284 1284 def public(repo, subset, x):
1285 1285 """``public()``
1286 1286 Changeset in public phase."""
1287 1287 # i18n: "public" is a keyword
1288 1288 getargs(x, 0, 0, _("public takes no arguments"))
1289 1289 pc = repo._phasecache
1290 1290 return subset.filter(lambda r: pc.phase(repo, r) == phases.public)
1291 1291
1292 1292 def remote(repo, subset, x):
1293 1293 """``remote([id [,path]])``
1294 1294 Local revision that corresponds to the given identifier in a
1295 1295 remote repository, if present. Here, the '.' identifier is a
1296 1296 synonym for the current local branch.
1297 1297 """
1298 1298
1299 1299 import hg # avoid start-up nasties
1300 1300 # i18n: "remote" is a keyword
1301 1301 l = getargs(x, 0, 2, _("remote takes one, two or no arguments"))
1302 1302
1303 1303 q = '.'
1304 1304 if len(l) > 0:
1305 1305 # i18n: "remote" is a keyword
1306 1306 q = getstring(l[0], _("remote requires a string id"))
1307 1307 if q == '.':
1308 1308 q = repo['.'].branch()
1309 1309
1310 1310 dest = ''
1311 1311 if len(l) > 1:
1312 1312 # i18n: "remote" is a keyword
1313 1313 dest = getstring(l[1], _("remote requires a repository path"))
1314 1314 dest = repo.ui.expandpath(dest or 'default')
1315 1315 dest, branches = hg.parseurl(dest)
1316 1316 revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
1317 1317 if revs:
1318 1318 revs = [repo.lookup(rev) for rev in revs]
1319 1319 other = hg.peer(repo, {}, dest)
1320 1320 n = other.lookup(q)
1321 1321 if n in repo:
1322 1322 r = repo[n].rev()
1323 1323 if r in subset:
1324 1324 return baseset([r])
1325 1325 return baseset()
1326 1326
1327 1327 def removes(repo, subset, x):
1328 1328 """``removes(pattern)``
1329 1329 Changesets which remove files matching pattern.
1330 1330
1331 1331 The pattern without explicit kind like ``glob:`` is expected to be
1332 1332 relative to the current directory and match against a file or a
1333 1333 directory.
1334 1334 """
1335 1335 # i18n: "removes" is a keyword
1336 1336 pat = getstring(x, _("removes requires a pattern"))
1337 1337 return checkstatus(repo, subset, pat, 2)
1338 1338
1339 1339 def rev(repo, subset, x):
1340 1340 """``rev(number)``
1341 1341 Revision with the given numeric identifier.
1342 1342 """
1343 1343 # i18n: "rev" is a keyword
1344 1344 l = getargs(x, 1, 1, _("rev requires one argument"))
1345 1345 try:
1346 1346 # i18n: "rev" is a keyword
1347 1347 l = int(getstring(l[0], _("rev requires a number")))
1348 1348 except (TypeError, ValueError):
1349 1349 # i18n: "rev" is a keyword
1350 1350 raise error.ParseError(_("rev expects a number"))
1351 1351 return subset & baseset([l])
1352 1352
1353 1353 def matching(repo, subset, x):
1354 1354 """``matching(revision [, field])``
1355 1355 Changesets in which a given set of fields match the set of fields in the
1356 1356 selected revision or set.
1357 1357
1358 1358 To match more than one field pass the list of fields to match separated
1359 1359 by spaces (e.g. ``author description``).
1360 1360
1361 1361 Valid fields are most regular revision fields and some special fields.
1362 1362
1363 1363 Regular revision fields are ``description``, ``author``, ``branch``,
1364 1364 ``date``, ``files``, ``phase``, ``parents``, ``substate``, ``user``
1365 1365 and ``diff``.
1366 1366 Note that ``author`` and ``user`` are synonyms. ``diff`` refers to the
1367 1367 contents of the revision. Two revisions matching their ``diff`` will
1368 1368 also match their ``files``.
1369 1369
1370 1370 Special fields are ``summary`` and ``metadata``:
1371 1371 ``summary`` matches the first line of the description.
1372 1372 ``metadata`` is equivalent to matching ``description user date``
1373 1373 (i.e. it matches the main metadata fields).
1374 1374
1375 1375 ``metadata`` is the default field which is used when no fields are
1376 1376 specified. You can match more than one field at a time.
1377 1377 """
1378 1378 # i18n: "matching" is a keyword
1379 1379 l = getargs(x, 1, 2, _("matching takes 1 or 2 arguments"))
1380 1380
1381 1381 revs = getset(repo, baseset(repo.changelog), l[0])
1382 1382
1383 1383 fieldlist = ['metadata']
1384 1384 if len(l) > 1:
1385 1385 fieldlist = getstring(l[1],
1386 1386 # i18n: "matching" is a keyword
1387 1387 _("matching requires a string "
1388 1388 "as its second argument")).split()
1389 1389
1390 1390 # Make sure that there are no repeated fields,
1391 1391 # expand the 'special' 'metadata' field type
1392 1392 # and check the 'files' whenever we check the 'diff'
1393 1393 fields = []
1394 1394 for field in fieldlist:
1395 1395 if field == 'metadata':
1396 1396 fields += ['user', 'description', 'date']
1397 1397 elif field == 'diff':
1398 1398 # a revision matching the diff must also match the files
1399 1399 # since matching the diff is very costly, make sure to
1400 1400 # also match the files first
1401 1401 fields += ['files', 'diff']
1402 1402 else:
1403 1403 if field == 'author':
1404 1404 field = 'user'
1405 1405 fields.append(field)
1406 1406 fields = set(fields)
1407 1407 if 'summary' in fields and 'description' in fields:
1408 1408 # If a revision matches its description it also matches its summary
1409 1409 fields.discard('summary')
1410 1410
1411 1411 # We may want to match more than one field
1412 1412 # Not all fields take the same amount of time to be matched
1413 1413 # Sort the selected fields in order of increasing matching cost
1414 1414 fieldorder = ['phase', 'parents', 'user', 'date', 'branch', 'summary',
1415 1415 'files', 'description', 'substate', 'diff']
1416 1416 def fieldkeyfunc(f):
1417 1417 try:
1418 1418 return fieldorder.index(f)
1419 1419 except ValueError:
1420 1420 # assume an unknown field is very costly
1421 1421 return len(fieldorder)
1422 1422 fields = list(fields)
1423 1423 fields.sort(key=fieldkeyfunc)
1424 1424
1425 1425 # Each field will be matched with its own "getfield" function
1426 1426 # which will be added to the getfieldfuncs array of functions
1427 1427 getfieldfuncs = []
1428 1428 _funcs = {
1429 1429 'user': lambda r: repo[r].user(),
1430 1430 'branch': lambda r: repo[r].branch(),
1431 1431 'date': lambda r: repo[r].date(),
1432 1432 'description': lambda r: repo[r].description(),
1433 1433 'files': lambda r: repo[r].files(),
1434 1434 'parents': lambda r: repo[r].parents(),
1435 1435 'phase': lambda r: repo[r].phase(),
1436 1436 'substate': lambda r: repo[r].substate,
1437 1437 'summary': lambda r: repo[r].description().splitlines()[0],
1438 1438 'diff': lambda r: list(repo[r].diff(git=True),)
1439 1439 }
1440 1440 for info in fields:
1441 1441 getfield = _funcs.get(info, None)
1442 1442 if getfield is None:
1443 1443 raise error.ParseError(
1444 1444 # i18n: "matching" is a keyword
1445 1445 _("unexpected field name passed to matching: %s") % info)
1446 1446 getfieldfuncs.append(getfield)
1447 1447 # convert the getfield array of functions into a "getinfo" function
1448 1448 # which returns an array of field values (or a single value if there
1449 1449 # is only one field to match)
1450 1450 getinfo = lambda r: [f(r) for f in getfieldfuncs]
1451 1451
1452 1452 def matches(x):
1453 1453 for rev in revs:
1454 1454 target = getinfo(rev)
1455 1455 match = True
1456 1456 for n, f in enumerate(getfieldfuncs):
1457 1457 if target[n] != f(x):
1458 1458 match = False
1459 1459 if match:
1460 1460 return True
1461 1461 return False
1462 1462
1463 1463 return subset.filter(matches)
1464 1464
1465 1465 def reverse(repo, subset, x):
1466 1466 """``reverse(set)``
1467 1467 Reverse order of set.
1468 1468 """
1469 1469 l = getset(repo, subset, x)
1470 1470 l.reverse()
1471 1471 return l
1472 1472
1473 1473 def roots(repo, subset, x):
1474 1474 """``roots(set)``
1475 1475 Changesets in set with no parent changeset in set.
1476 1476 """
1477 1477 s = getset(repo, spanset(repo), x).set()
1478 1478 subset = baseset([r for r in s if r in subset.set()])
1479 1479 cs = _children(repo, subset, s)
1480 1480 return subset - cs
1481 1481
1482 1482 def secret(repo, subset, x):
1483 1483 """``secret()``
1484 1484 Changeset in secret phase."""
1485 1485 # i18n: "secret" is a keyword
1486 1486 getargs(x, 0, 0, _("secret takes no arguments"))
1487 1487 pc = repo._phasecache
1488 1488 return subset.filter(lambda x: pc.phase(repo, x) == phases.secret)
1489 1489
1490 1490 def sort(repo, subset, x):
1491 1491 """``sort(set[, [-]key...])``
1492 1492 Sort set by keys. The default sort order is ascending, specify a key
1493 1493 as ``-key`` to sort in descending order.
1494 1494
1495 1495 The keys can be:
1496 1496
1497 1497 - ``rev`` for the revision number,
1498 1498 - ``branch`` for the branch name,
1499 1499 - ``desc`` for the commit message (description),
1500 1500 - ``user`` for user name (``author`` can be used as an alias),
1501 1501 - ``date`` for the commit date
1502 1502 """
1503 1503 # i18n: "sort" is a keyword
1504 1504 l = getargs(x, 1, 2, _("sort requires one or two arguments"))
1505 1505 keys = "rev"
1506 1506 if len(l) == 2:
1507 1507 # i18n: "sort" is a keyword
1508 1508 keys = getstring(l[1], _("sort spec must be a string"))
1509 1509
1510 1510 s = l[0]
1511 1511 keys = keys.split()
1512 1512 l = []
1513 1513 def invert(s):
1514 1514 return "".join(chr(255 - ord(c)) for c in s)
1515 1515 revs = getset(repo, subset, s)
1516 1516 if keys == ["rev"]:
1517 1517 revs.sort()
1518 1518 return revs
1519 1519 elif keys == ["-rev"]:
1520 1520 revs.sort(reverse=True)
1521 1521 return revs
1522 1522 for r in revs:
1523 1523 c = repo[r]
1524 1524 e = []
1525 1525 for k in keys:
1526 1526 if k == 'rev':
1527 1527 e.append(r)
1528 1528 elif k == '-rev':
1529 1529 e.append(-r)
1530 1530 elif k == 'branch':
1531 1531 e.append(c.branch())
1532 1532 elif k == '-branch':
1533 1533 e.append(invert(c.branch()))
1534 1534 elif k == 'desc':
1535 1535 e.append(c.description())
1536 1536 elif k == '-desc':
1537 1537 e.append(invert(c.description()))
1538 1538 elif k in 'user author':
1539 1539 e.append(c.user())
1540 1540 elif k in '-user -author':
1541 1541 e.append(invert(c.user()))
1542 1542 elif k == 'date':
1543 1543 e.append(c.date()[0])
1544 1544 elif k == '-date':
1545 1545 e.append(-c.date()[0])
1546 1546 else:
1547 1547 raise error.ParseError(_("unknown sort key %r") % k)
1548 1548 e.append(r)
1549 1549 l.append(e)
1550 1550 l.sort()
1551 1551 return baseset([e[-1] for e in l])
1552 1552
1553 1553 def _stringmatcher(pattern):
1554 1554 """
1555 1555 accepts a string, possibly starting with 're:' or 'literal:' prefix.
1556 1556 returns the matcher name, pattern, and matcher function.
1557 1557 missing or unknown prefixes are treated as literal matches.
1558 1558
1559 1559 helper for tests:
1560 1560 >>> def test(pattern, *tests):
1561 1561 ... kind, pattern, matcher = _stringmatcher(pattern)
1562 1562 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
1563 1563
1564 1564 exact matching (no prefix):
1565 1565 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
1566 1566 ('literal', 'abcdefg', [False, False, True])
1567 1567
1568 1568 regex matching ('re:' prefix)
1569 1569 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
1570 1570 ('re', 'a.+b', [False, False, True])
1571 1571
1572 1572 force exact matches ('literal:' prefix)
1573 1573 >>> test('literal:re:foobar', 'foobar', 're:foobar')
1574 1574 ('literal', 're:foobar', [False, True])
1575 1575
1576 1576 unknown prefixes are ignored and treated as literals
1577 1577 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
1578 1578 ('literal', 'foo:bar', [False, False, True])
1579 1579 """
1580 1580 if pattern.startswith('re:'):
1581 1581 pattern = pattern[3:]
1582 1582 try:
1583 1583 regex = re.compile(pattern)
1584 1584 except re.error, e:
1585 1585 raise error.ParseError(_('invalid regular expression: %s')
1586 1586 % e)
1587 1587 return 're', pattern, regex.search
1588 1588 elif pattern.startswith('literal:'):
1589 1589 pattern = pattern[8:]
1590 1590 return 'literal', pattern, pattern.__eq__
1591 1591
1592 1592 def _substringmatcher(pattern):
1593 1593 kind, pattern, matcher = _stringmatcher(pattern)
1594 1594 if kind == 'literal':
1595 1595 matcher = lambda s: pattern in s
1596 1596 return kind, pattern, matcher
1597 1597
1598 1598 def tag(repo, subset, x):
1599 1599 """``tag([name])``
1600 1600 The specified tag by name, or all tagged revisions if no name is given.
1601 1601
1602 1602 If `name` starts with `re:`, the remainder of the name is treated as
1603 1603 a regular expression. To match a tag that actually starts with `re:`,
1604 1604 use the prefix `literal:`.
1605 1605 """
1606 1606 # i18n: "tag" is a keyword
1607 1607 args = getargs(x, 0, 1, _("tag takes one or no arguments"))
1608 1608 cl = repo.changelog
1609 1609 if args:
1610 1610 pattern = getstring(args[0],
1611 1611 # i18n: "tag" is a keyword
1612 1612 _('the argument to tag must be a string'))
1613 1613 kind, pattern, matcher = _stringmatcher(pattern)
1614 1614 if kind == 'literal':
1615 1615 # avoid resolving all tags
1616 1616 tn = repo._tagscache.tags.get(pattern, None)
1617 1617 if tn is None:
1618 1618 raise util.Abort(_("tag '%s' does not exist") % pattern)
1619 1619 s = set([repo[tn].rev()])
1620 1620 else:
1621 1621 s = set([cl.rev(n) for t, n in repo.tagslist() if matcher(t)])
1622 1622 else:
1623 1623 s = set([cl.rev(n) for t, n in repo.tagslist() if t != 'tip'])
1624 1624 return subset & s
1625 1625
1626 1626 def tagged(repo, subset, x):
1627 1627 return tag(repo, subset, x)
1628 1628
1629 1629 def unstable(repo, subset, x):
1630 1630 """``unstable()``
1631 1631 Non-obsolete changesets with obsolete ancestors.
1632 1632 """
1633 1633 # i18n: "unstable" is a keyword
1634 1634 getargs(x, 0, 0, _("unstable takes no arguments"))
1635 1635 unstables = obsmod.getrevs(repo, 'unstable')
1636 1636 return subset & unstables
1637 1637
1638 1638
1639 1639 def user(repo, subset, x):
1640 1640 """``user(string)``
1641 1641 User name contains string. The match is case-insensitive.
1642 1642
1643 1643 If `string` starts with `re:`, the remainder of the string is treated as
1644 1644 a regular expression. To match a user that actually contains `re:`, use
1645 1645 the prefix `literal:`.
1646 1646 """
1647 1647 return author(repo, subset, x)
1648 1648
1649 1649 # for internal use
1650 1650 def _list(repo, subset, x):
1651 1651 s = getstring(x, "internal error")
1652 1652 if not s:
1653 1653 return baseset()
1654 1654 ls = [repo[r].rev() for r in s.split('\0')]
1655 1655 s = subset.set()
1656 1656 return baseset([r for r in ls if r in s])
1657 1657
1658 1658 # for internal use
1659 1659 def _intlist(repo, subset, x):
1660 1660 s = getstring(x, "internal error")
1661 1661 if not s:
1662 1662 return baseset()
1663 1663 ls = [int(r) for r in s.split('\0')]
1664 1664 s = subset.set()
1665 1665 return baseset([r for r in ls if r in s])
1666 1666
1667 1667 # for internal use
1668 1668 def _hexlist(repo, subset, x):
1669 1669 s = getstring(x, "internal error")
1670 1670 if not s:
1671 1671 return baseset()
1672 1672 cl = repo.changelog
1673 1673 ls = [cl.rev(node.bin(r)) for r in s.split('\0')]
1674 1674 s = subset.set()
1675 1675 return baseset([r for r in ls if r in s])
1676 1676
1677 1677 symbols = {
1678 1678 "adds": adds,
1679 1679 "all": getall,
1680 1680 "ancestor": ancestor,
1681 1681 "ancestors": ancestors,
1682 1682 "_firstancestors": _firstancestors,
1683 1683 "author": author,
1684 1684 "only": only,
1685 1685 "bisect": bisect,
1686 1686 "bisected": bisected,
1687 1687 "bookmark": bookmark,
1688 1688 "branch": branch,
1689 1689 "branchpoint": branchpoint,
1690 1690 "bumped": bumped,
1691 1691 "bundle": bundle,
1692 1692 "children": children,
1693 1693 "closed": closed,
1694 1694 "contains": contains,
1695 1695 "converted": converted,
1696 1696 "date": date,
1697 1697 "desc": desc,
1698 1698 "descendants": descendants,
1699 1699 "_firstdescendants": _firstdescendants,
1700 1700 "destination": destination,
1701 1701 "divergent": divergent,
1702 1702 "draft": draft,
1703 1703 "extinct": extinct,
1704 1704 "extra": extra,
1705 1705 "file": hasfile,
1706 1706 "filelog": filelog,
1707 1707 "first": first,
1708 1708 "follow": follow,
1709 1709 "_followfirst": _followfirst,
1710 1710 "grep": grep,
1711 1711 "head": head,
1712 1712 "heads": heads,
1713 1713 "hidden": hidden,
1714 1714 "id": node_,
1715 1715 "keyword": keyword,
1716 1716 "last": last,
1717 1717 "limit": limit,
1718 1718 "_matchfiles": _matchfiles,
1719 1719 "max": maxrev,
1720 1720 "merge": merge,
1721 1721 "min": minrev,
1722 1722 "modifies": modifies,
1723 1723 "obsolete": obsolete,
1724 1724 "origin": origin,
1725 1725 "outgoing": outgoing,
1726 1726 "p1": p1,
1727 1727 "p2": p2,
1728 1728 "parents": parents,
1729 1729 "present": present,
1730 1730 "public": public,
1731 1731 "remote": remote,
1732 1732 "removes": removes,
1733 1733 "rev": rev,
1734 1734 "reverse": reverse,
1735 1735 "roots": roots,
1736 1736 "sort": sort,
1737 1737 "secret": secret,
1738 1738 "matching": matching,
1739 1739 "tag": tag,
1740 1740 "tagged": tagged,
1741 1741 "user": user,
1742 1742 "unstable": unstable,
1743 1743 "_list": _list,
1744 1744 "_intlist": _intlist,
1745 1745 "_hexlist": _hexlist,
1746 1746 }
1747 1747
1748 1748 # symbols which can't be used for a DoS attack for any given input
1749 1749 # (e.g. those which accept regexes as plain strings shouldn't be included)
1750 1750 # functions that just return a lot of changesets (like all) don't count here
1751 1751 safesymbols = set([
1752 1752 "adds",
1753 1753 "all",
1754 1754 "ancestor",
1755 1755 "ancestors",
1756 1756 "_firstancestors",
1757 1757 "author",
1758 1758 "bisect",
1759 1759 "bisected",
1760 1760 "bookmark",
1761 1761 "branch",
1762 1762 "branchpoint",
1763 1763 "bumped",
1764 1764 "bundle",
1765 1765 "children",
1766 1766 "closed",
1767 1767 "converted",
1768 1768 "date",
1769 1769 "desc",
1770 1770 "descendants",
1771 1771 "_firstdescendants",
1772 1772 "destination",
1773 1773 "divergent",
1774 1774 "draft",
1775 1775 "extinct",
1776 1776 "extra",
1777 1777 "file",
1778 1778 "filelog",
1779 1779 "first",
1780 1780 "follow",
1781 1781 "_followfirst",
1782 1782 "head",
1783 1783 "heads",
1784 1784 "hidden",
1785 1785 "id",
1786 1786 "keyword",
1787 1787 "last",
1788 1788 "limit",
1789 1789 "_matchfiles",
1790 1790 "max",
1791 1791 "merge",
1792 1792 "min",
1793 1793 "modifies",
1794 1794 "obsolete",
1795 1795 "origin",
1796 1796 "outgoing",
1797 1797 "p1",
1798 1798 "p2",
1799 1799 "parents",
1800 1800 "present",
1801 1801 "public",
1802 1802 "remote",
1803 1803 "removes",
1804 1804 "rev",
1805 1805 "reverse",
1806 1806 "roots",
1807 1807 "sort",
1808 1808 "secret",
1809 1809 "matching",
1810 1810 "tag",
1811 1811 "tagged",
1812 1812 "user",
1813 1813 "unstable",
1814 1814 "_list",
1815 1815 "_intlist",
1816 1816 "_hexlist",
1817 1817 ])
1818 1818
1819 1819 methods = {
1820 1820 "range": rangeset,
1821 1821 "dagrange": dagrange,
1822 1822 "string": stringset,
1823 1823 "symbol": symbolset,
1824 1824 "and": andset,
1825 1825 "or": orset,
1826 1826 "not": notset,
1827 1827 "list": listset,
1828 1828 "func": func,
1829 1829 "ancestor": ancestorspec,
1830 1830 "parent": parentspec,
1831 1831 "parentpost": p1,
1832 1832 }
1833 1833
1834 1834 def optimize(x, small):
1835 1835 if x is None:
1836 1836 return 0, x
1837 1837
1838 1838 smallbonus = 1
1839 1839 if small:
1840 1840 smallbonus = .5
1841 1841
1842 1842 op = x[0]
1843 1843 if op == 'minus':
1844 1844 return optimize(('and', x[1], ('not', x[2])), small)
1845 1845 elif op == 'dagrangepre':
1846 1846 return optimize(('func', ('symbol', 'ancestors'), x[1]), small)
1847 1847 elif op == 'dagrangepost':
1848 1848 return optimize(('func', ('symbol', 'descendants'), x[1]), small)
1849 1849 elif op == 'rangepre':
1850 1850 return optimize(('range', ('string', '0'), x[1]), small)
1851 1851 elif op == 'rangepost':
1852 1852 return optimize(('range', x[1], ('string', 'tip')), small)
1853 1853 elif op == 'negate':
1854 1854 return optimize(('string',
1855 1855 '-' + getstring(x[1], _("can't negate that"))), small)
1856 1856 elif op in 'string symbol negate':
1857 1857 return smallbonus, x # single revisions are small
1858 1858 elif op == 'and':
1859 1859 wa, ta = optimize(x[1], True)
1860 1860 wb, tb = optimize(x[2], True)
1861 1861
1862 1862 # (::x and not ::y)/(not ::y and ::x) have a fast path
1863 1863 def isonly(revs, bases):
1864 1864 return (
1865 1865 revs[0] == 'func'
1866 1866 and getstring(revs[1], _('not a symbol')) == 'ancestors'
1867 1867 and bases[0] == 'not'
1868 1868 and bases[1][0] == 'func'
1869 1869 and getstring(bases[1][1], _('not a symbol')) == 'ancestors')
1870 1870
1871 1871 w = min(wa, wb)
1872 1872 if isonly(ta, tb):
1873 1873 return w, ('func', ('symbol', 'only'), ('list', ta[2], tb[1][2]))
1874 1874 if isonly(tb, ta):
1875 1875 return w, ('func', ('symbol', 'only'), ('list', tb[2], ta[1][2]))
1876 1876
1877 1877 if wa > wb:
1878 1878 return w, (op, tb, ta)
1879 1879 return w, (op, ta, tb)
1880 1880 elif op == 'or':
1881 1881 wa, ta = optimize(x[1], False)
1882 1882 wb, tb = optimize(x[2], False)
1883 1883 if wb < wa:
1884 1884 wb, wa = wa, wb
1885 1885 return max(wa, wb), (op, ta, tb)
1886 1886 elif op == 'not':
1887 1887 o = optimize(x[1], not small)
1888 1888 return o[0], (op, o[1])
1889 1889 elif op == 'parentpost':
1890 1890 o = optimize(x[1], small)
1891 1891 return o[0], (op, o[1])
1892 1892 elif op == 'group':
1893 1893 return optimize(x[1], small)
1894 1894 elif op in 'dagrange range list parent ancestorspec':
1895 1895 if op == 'parent':
1896 1896 # x^:y means (x^) : y, not x ^ (:y)
1897 1897 post = ('parentpost', x[1])
1898 1898 if x[2][0] == 'dagrangepre':
1899 1899 return optimize(('dagrange', post, x[2][1]), small)
1900 1900 elif x[2][0] == 'rangepre':
1901 1901 return optimize(('range', post, x[2][1]), small)
1902 1902
1903 1903 wa, ta = optimize(x[1], small)
1904 1904 wb, tb = optimize(x[2], small)
1905 1905 return wa + wb, (op, ta, tb)
1906 1906 elif op == 'func':
1907 1907 f = getstring(x[1], _("not a symbol"))
1908 1908 wa, ta = optimize(x[2], small)
1909 1909 if f in ("author branch closed date desc file grep keyword "
1910 1910 "outgoing user"):
1911 1911 w = 10 # slow
1912 1912 elif f in "modifies adds removes":
1913 1913 w = 30 # slower
1914 1914 elif f == "contains":
1915 1915 w = 100 # very slow
1916 1916 elif f == "ancestor":
1917 1917 w = 1 * smallbonus
1918 1918 elif f in "reverse limit first _intlist":
1919 1919 w = 0
1920 1920 elif f in "sort":
1921 1921 w = 10 # assume most sorts look at changelog
1922 1922 else:
1923 1923 w = 1
1924 1924 return w + wa, (op, x[1], ta)
1925 1925 return 1, x
1926 1926
1927 1927 _aliasarg = ('func', ('symbol', '_aliasarg'))
1928 1928 def _getaliasarg(tree):
1929 1929 """If tree matches ('func', ('symbol', '_aliasarg'), ('string', X))
1930 1930 return X, None otherwise.
1931 1931 """
1932 1932 if (len(tree) == 3 and tree[:2] == _aliasarg
1933 1933 and tree[2][0] == 'string'):
1934 1934 return tree[2][1]
1935 1935 return None
1936 1936
1937 1937 def _checkaliasarg(tree, known=None):
1938 1938 """Check tree contains no _aliasarg construct or only ones which
1939 1939 value is in known. Used to avoid alias placeholders injection.
1940 1940 """
1941 1941 if isinstance(tree, tuple):
1942 1942 arg = _getaliasarg(tree)
1943 1943 if arg is not None and (not known or arg not in known):
1944 1944 raise error.ParseError(_("not a function: %s") % '_aliasarg')
1945 1945 for t in tree:
1946 1946 _checkaliasarg(t, known)
1947 1947
1948 1948 class revsetalias(object):
1949 1949 funcre = re.compile('^([^(]+)\(([^)]+)\)$')
1950 1950 args = None
1951 1951
1952 1952 def __init__(self, name, value):
1953 1953 '''Aliases like:
1954 1954
1955 1955 h = heads(default)
1956 1956 b($1) = ancestors($1) - ancestors(default)
1957 1957 '''
1958 1958 m = self.funcre.search(name)
1959 1959 if m:
1960 1960 self.name = m.group(1)
1961 1961 self.tree = ('func', ('symbol', m.group(1)))
1962 1962 self.args = [x.strip() for x in m.group(2).split(',')]
1963 1963 for arg in self.args:
1964 1964 # _aliasarg() is an unknown symbol only used separate
1965 1965 # alias argument placeholders from regular strings.
1966 1966 value = value.replace(arg, '_aliasarg(%r)' % (arg,))
1967 1967 else:
1968 1968 self.name = name
1969 1969 self.tree = ('symbol', name)
1970 1970
1971 1971 self.replacement, pos = parse(value)
1972 1972 if pos != len(value):
1973 1973 raise error.ParseError(_('invalid token'), pos)
1974 1974 # Check for placeholder injection
1975 1975 _checkaliasarg(self.replacement, self.args)
1976 1976
1977 1977 def _getalias(aliases, tree):
1978 1978 """If tree looks like an unexpanded alias, return it. Return None
1979 1979 otherwise.
1980 1980 """
1981 1981 if isinstance(tree, tuple) and tree:
1982 1982 if tree[0] == 'symbol' and len(tree) == 2:
1983 1983 name = tree[1]
1984 1984 alias = aliases.get(name)
1985 1985 if alias and alias.args is None and alias.tree == tree:
1986 1986 return alias
1987 1987 if tree[0] == 'func' and len(tree) > 1:
1988 1988 if tree[1][0] == 'symbol' and len(tree[1]) == 2:
1989 1989 name = tree[1][1]
1990 1990 alias = aliases.get(name)
1991 1991 if alias and alias.args is not None and alias.tree == tree[:2]:
1992 1992 return alias
1993 1993 return None
1994 1994
1995 1995 def _expandargs(tree, args):
1996 1996 """Replace _aliasarg instances with the substitution value of the
1997 1997 same name in args, recursively.
1998 1998 """
1999 1999 if not tree or not isinstance(tree, tuple):
2000 2000 return tree
2001 2001 arg = _getaliasarg(tree)
2002 2002 if arg is not None:
2003 2003 return args[arg]
2004 2004 return tuple(_expandargs(t, args) for t in tree)
2005 2005
2006 2006 def _expandaliases(aliases, tree, expanding, cache):
2007 2007 """Expand aliases in tree, recursively.
2008 2008
2009 2009 'aliases' is a dictionary mapping user defined aliases to
2010 2010 revsetalias objects.
2011 2011 """
2012 2012 if not isinstance(tree, tuple):
2013 2013 # Do not expand raw strings
2014 2014 return tree
2015 2015 alias = _getalias(aliases, tree)
2016 2016 if alias is not None:
2017 2017 if alias in expanding:
2018 2018 raise error.ParseError(_('infinite expansion of revset alias "%s" '
2019 2019 'detected') % alias.name)
2020 2020 expanding.append(alias)
2021 2021 if alias.name not in cache:
2022 2022 cache[alias.name] = _expandaliases(aliases, alias.replacement,
2023 2023 expanding, cache)
2024 2024 result = cache[alias.name]
2025 2025 expanding.pop()
2026 2026 if alias.args is not None:
2027 2027 l = getlist(tree[2])
2028 2028 if len(l) != len(alias.args):
2029 2029 raise error.ParseError(
2030 2030 _('invalid number of arguments: %s') % len(l))
2031 2031 l = [_expandaliases(aliases, a, [], cache) for a in l]
2032 2032 result = _expandargs(result, dict(zip(alias.args, l)))
2033 2033 else:
2034 2034 result = tuple(_expandaliases(aliases, t, expanding, cache)
2035 2035 for t in tree)
2036 2036 return result
2037 2037
2038 2038 def findaliases(ui, tree):
2039 2039 _checkaliasarg(tree)
2040 2040 aliases = {}
2041 2041 for k, v in ui.configitems('revsetalias'):
2042 2042 alias = revsetalias(k, v)
2043 2043 aliases[alias.name] = alias
2044 2044 return _expandaliases(aliases, tree, [], {})
2045 2045
2046 2046 def parse(spec, lookup=None):
2047 2047 p = parser.parser(tokenize, elements)
2048 2048 return p.parse(spec, lookup=lookup)
2049 2049
2050 2050 def match(ui, spec, repo=None):
2051 2051 if not spec:
2052 2052 raise error.ParseError(_("empty query"))
2053 2053 lookup = None
2054 2054 if repo:
2055 2055 lookup = repo.__contains__
2056 2056 tree, pos = parse(spec, lookup)
2057 2057 if (pos != len(spec)):
2058 2058 raise error.ParseError(_("invalid token"), pos)
2059 2059 if ui:
2060 2060 tree = findaliases(ui, tree)
2061 2061 weight, tree = optimize(tree, True)
2062 2062 def mfunc(repo, subset):
2063 2063 if util.safehasattr(subset, 'set'):
2064 2064 result = getset(repo, subset, tree)
2065 2065 else:
2066 2066 result = getset(repo, baseset(subset), tree)
2067 2067 return result
2068 2068 return mfunc
2069 2069
2070 2070 def formatspec(expr, *args):
2071 2071 '''
2072 2072 This is a convenience function for using revsets internally, and
2073 2073 escapes arguments appropriately. Aliases are intentionally ignored
2074 2074 so that intended expression behavior isn't accidentally subverted.
2075 2075
2076 2076 Supported arguments:
2077 2077
2078 2078 %r = revset expression, parenthesized
2079 2079 %d = int(arg), no quoting
2080 2080 %s = string(arg), escaped and single-quoted
2081 2081 %b = arg.branch(), escaped and single-quoted
2082 2082 %n = hex(arg), single-quoted
2083 2083 %% = a literal '%'
2084 2084
2085 2085 Prefixing the type with 'l' specifies a parenthesized list of that type.
2086 2086
2087 2087 >>> formatspec('%r:: and %lr', '10 or 11', ("this()", "that()"))
2088 2088 '(10 or 11):: and ((this()) or (that()))'
2089 2089 >>> formatspec('%d:: and not %d::', 10, 20)
2090 2090 '10:: and not 20::'
2091 2091 >>> formatspec('%ld or %ld', [], [1])
2092 2092 "_list('') or 1"
2093 2093 >>> formatspec('keyword(%s)', 'foo\\xe9')
2094 2094 "keyword('foo\\\\xe9')"
2095 2095 >>> b = lambda: 'default'
2096 2096 >>> b.branch = b
2097 2097 >>> formatspec('branch(%b)', b)
2098 2098 "branch('default')"
2099 2099 >>> formatspec('root(%ls)', ['a', 'b', 'c', 'd'])
2100 2100 "root(_list('a\\x00b\\x00c\\x00d'))"
2101 2101 '''
2102 2102
2103 2103 def quote(s):
2104 2104 return repr(str(s))
2105 2105
2106 2106 def argtype(c, arg):
2107 2107 if c == 'd':
2108 2108 return str(int(arg))
2109 2109 elif c == 's':
2110 2110 return quote(arg)
2111 2111 elif c == 'r':
2112 2112 parse(arg) # make sure syntax errors are confined
2113 2113 return '(%s)' % arg
2114 2114 elif c == 'n':
2115 2115 return quote(node.hex(arg))
2116 2116 elif c == 'b':
2117 2117 return quote(arg.branch())
2118 2118
2119 2119 def listexp(s, t):
2120 2120 l = len(s)
2121 2121 if l == 0:
2122 2122 return "_list('')"
2123 2123 elif l == 1:
2124 2124 return argtype(t, s[0])
2125 2125 elif t == 'd':
2126 2126 return "_intlist('%s')" % "\0".join(str(int(a)) for a in s)
2127 2127 elif t == 's':
2128 2128 return "_list('%s')" % "\0".join(s)
2129 2129 elif t == 'n':
2130 2130 return "_hexlist('%s')" % "\0".join(node.hex(a) for a in s)
2131 2131 elif t == 'b':
2132 2132 return "_list('%s')" % "\0".join(a.branch() for a in s)
2133 2133
2134 2134 m = l // 2
2135 2135 return '(%s or %s)' % (listexp(s[:m], t), listexp(s[m:], t))
2136 2136
2137 2137 ret = ''
2138 2138 pos = 0
2139 2139 arg = 0
2140 2140 while pos < len(expr):
2141 2141 c = expr[pos]
2142 2142 if c == '%':
2143 2143 pos += 1
2144 2144 d = expr[pos]
2145 2145 if d == '%':
2146 2146 ret += d
2147 2147 elif d in 'dsnbr':
2148 2148 ret += argtype(d, args[arg])
2149 2149 arg += 1
2150 2150 elif d == 'l':
2151 2151 # a list of some type
2152 2152 pos += 1
2153 2153 d = expr[pos]
2154 2154 ret += listexp(list(args[arg]), d)
2155 2155 arg += 1
2156 2156 else:
2157 2157 raise util.Abort('unexpected revspec format character %s' % d)
2158 2158 else:
2159 2159 ret += c
2160 2160 pos += 1
2161 2161
2162 2162 return ret
2163 2163
2164 2164 def prettyformat(tree):
2165 2165 def _prettyformat(tree, level, lines):
2166 2166 if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
2167 2167 lines.append((level, str(tree)))
2168 2168 else:
2169 2169 lines.append((level, '(%s' % tree[0]))
2170 2170 for s in tree[1:]:
2171 2171 _prettyformat(s, level + 1, lines)
2172 2172 lines[-1:] = [(lines[-1][0], lines[-1][1] + ')')]
2173 2173
2174 2174 lines = []
2175 2175 _prettyformat(tree, 0, lines)
2176 2176 output = '\n'.join((' '*l + s) for l, s in lines)
2177 2177 return output
2178 2178
2179 2179 def depth(tree):
2180 2180 if isinstance(tree, tuple):
2181 2181 return max(map(depth, tree)) + 1
2182 2182 else:
2183 2183 return 0
2184 2184
2185 2185 def funcsused(tree):
2186 2186 if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
2187 2187 return set()
2188 2188 else:
2189 2189 funcs = set()
2190 2190 for s in tree[1:]:
2191 2191 funcs |= funcsused(s)
2192 2192 if tree[0] == 'func':
2193 2193 funcs.add(tree[1][1])
2194 2194 return funcs
2195 2195
2196 2196 class abstractsmartset(object):
2197 2197
2198 2198 def __nonzero__(self):
2199 2199 """True if the smartset is not empty"""
2200 2200 raise NotImplementedError()
2201 2201
2202 2202 def __contains__(self, rev):
2203 2203 """provide fast membership testing"""
2204 2204 raise NotImplementedError()
2205 2205
2206 2206 def __set__(self):
2207 2207 """Returns a set or a smartset containing all the elements.
2208 2208
2209 2209 The returned structure should be the fastest option for membership
2210 2210 testing.
2211 2211
2212 2212 This is part of the mandatory API for smartset."""
2213 2213 raise NotImplementedError()
2214 2214
2215 2215 def __iter__(self):
2216 2216 """iterate the set in the order it is supposed to be iterated"""
2217 2217 raise NotImplementedError()
2218 2218
2219 2219 # Attributes containing a function to perform a fast iteration in a given
2220 2220 # direction. A smartset can have none, one, or both defined.
2221 2221 #
2222 2222 # Default value is None instead of a function returning None to avoid
2223 2223 # initializing an iterator just for testing if a fast method exists.
2224 2224 fastasc = None
2225 2225 fastdesc = None
2226 2226
2227 2227 def isascending(self):
2228 2228 """True if the set will iterate in ascending order"""
2229 2229 raise NotImplementedError()
2230 2230
2231 2231 def ascending(self):
2232 2232 """Sorts the set in ascending order (in place).
2233 2233
2234 2234 This is part of the mandatory API for smartset."""
2235 2235 self.sort()
2236 2236
2237 2237 def isdescending(self):
2238 2238 """True if the set will iterate in descending order"""
2239 2239 raise NotImplementedError()
2240 2240
2241 2241 def descending(self):
2242 2242 """Sorts the set in descending order (in place).
2243 2243
2244 2244 This is part of the mandatory API for smartset."""
2245 2245 self.sort(reverse=True)
2246 2246
2247 2247 def min(self):
2248 2248 """return the minimum element in the set"""
2249 2249 if self.fastasc is not None:
2250 2250 for r in self.fastasc():
2251 2251 return r
2252 2252 raise ValueError('arg is an empty sequence')
2253 2253 return min(self)
2254 2254
2255 2255 def max(self):
2256 2256 """return the maximum element in the set"""
2257 2257 if self.fastdesc is not None:
2258 2258 for r in self.fastdesc():
2259 2259 return r
2260 2260 raise ValueError('arg is an empty sequence')
2261 2261 return max(self)
2262 2262
2263 2263 def first(self):
2264 2264 """return the first element in the set (user iteration perspective)
2265 2265
2266 2266 Return None if the set is empty"""
2267 2267 raise NotImplementedError()
2268 2268
2269 2269 def last(self):
2270 2270 """return the last element in the set (user iteration perspective)
2271 2271
2272 2272 Return None if the set is empty"""
2273 2273 raise NotImplementedError()
2274 2274
2275 2275 def reverse(self):
2276 2276 """reverse the expected iteration order"""
2277 2277 raise NotImplementedError()
2278 2278
2279 2279 def sort(self, reverse=True):
2280 2280 """get the set to iterate in an ascending or descending order"""
2281 2281 raise NotImplementedError()
2282 2282
2283 2283 def __and__(self, other):
2284 2284 """Returns a new object with the intersection of the two collections.
2285 2285
2286 2286 This is part of the mandatory API for smartset."""
2287 2287 return self.filter(other.__contains__)
2288 2288
2289 2289 def __add__(self, other):
2290 2290 """Returns a new object with the union of the two collections.
2291 2291
2292 2292 This is part of the mandatory API for smartset."""
2293 2293 kwargs = {}
2294 2294 if self.isascending() and other.isascending():
2295 2295 kwargs['ascending'] = True
2296 2296 if self.isdescending() and other.isdescending():
2297 2297 kwargs['ascending'] = False
2298 2298 return addset(self, other, **kwargs)
2299 2299
2300 2300 def __sub__(self, other):
2301 2301 """Returns a new object with the substraction of the two collections.
2302 2302
2303 2303 This is part of the mandatory API for smartset."""
2304 2304 c = other.__contains__
2305 2305 return self.filter(lambda r: not c(r))
2306 2306
2307 2307 def filter(self, condition):
2308 2308 """Returns this smartset filtered by condition as a new smartset.
2309 2309
2310 2310 `condition` is a callable which takes a revision number and returns a
2311 2311 boolean.
2312 2312
2313 2313 This is part of the mandatory API for smartset."""
2314 2314 kwargs = {}
2315 2315 if self.isascending():
2316 2316 kwargs['ascending'] = True
2317 2317 elif self.isdescending():
2318 2318 kwargs['ascending'] = False
2319 2319 return filteredset(self, condition, **kwargs)
2320 2320
2321 2321 class baseset(abstractsmartset):
2322 2322 """Basic data structure that represents a revset and contains the basic
2323 2323 operation that it should be able to perform.
2324 2324
2325 2325 Every method in this class should be implemented by any smartset class.
2326 2326 """
2327 2327 def __init__(self, data=()):
2328 2328 if not isinstance(data, list):
2329 2329 data = list(data)
2330 2330 self._list = data
2331 2331 self._set = None
2332 2332 self._ascending = None
2333 2333
2334 2334 @util.propertycache
2335 2335 def _asclist(self):
2336 2336 asclist = self._list[:]
2337 2337 asclist.sort()
2338 2338 return asclist
2339 2339
2340 2340 def __iter__(self):
2341 2341 if self._ascending is None:
2342 2342 return iter(self._list)
2343 2343 elif self._ascending:
2344 2344 return iter(self._asclist)
2345 2345 else:
2346 2346 return reversed(self._asclist)
2347 2347
2348 2348 def fastasc(self):
2349 2349 return iter(self._asclist)
2350 2350
2351 2351 def fastdesc(self):
2352 2352 return reversed(self._asclist)
2353 2353
2354 2354 def set(self):
2355 2355 """Returns a set or a smartset containing all the elements.
2356 2356
2357 2357 The returned structure should be the fastest option for membership
2358 2358 testing.
2359 2359
2360 2360 This is part of the mandatory API for smartset."""
2361 2361 if not self._set:
2362 2362 self._set = set(self)
2363 2363 return self._set
2364 2364
2365 2365 @util.propertycache
2366 2366 def __contains__(self):
2367 2367 return self.set().__contains__
2368 2368
2369 2369 def __nonzero__(self):
2370 2370 return bool(self._list)
2371 2371
2372 2372 def sort(self, reverse=False):
2373 2373 self._ascending = not bool(reverse)
2374 2374
2375 2375 def reverse(self):
2376 2376 if self._ascending is None:
2377 2377 self._list.reverse()
2378 2378 else:
2379 2379 self._ascending = not self._ascending
2380 2380
2381 2381 def __len__(self):
2382 2382 return len(self._list)
2383 2383
2384 2384 def __sub__(self, other):
2385 2385 """Returns a new object with the substraction of the two collections.
2386 2386
2387 2387 This is part of the mandatory API for smartset."""
2388 2388 # If we are operating on 2 baseset, do the computation now since all
2389 2389 # data is available. The alternative is to involve a filteredset, which
2390 2390 # may be slow.
2391 2391 if isinstance(other, baseset):
2392 2392 other = other.set()
2393 2393 return baseset([x for x in self if x not in other])
2394 2394
2395 2395 return self.filter(lambda x: x not in other)
2396 2396
2397 2397 def __and__(self, other):
2398 2398 """Returns a new object with the intersection of the two collections.
2399 2399
2400 2400 This is part of the mandatory API for smartset."""
2401 2401 return baseset([y for y in self if y in other])
2402 2402
2403 2403 def __add__(self, other):
2404 2404 """Returns a new object with the union of the two collections.
2405 2405
2406 2406 This is part of the mandatory API for smartset."""
2407 2407 s = self.set()
2408 2408 l = [r for r in other if r not in s]
2409 2409 return baseset(list(self) + l)
2410 2410
2411 2411 def isascending(self):
2412 2412 """Returns True if the collection is ascending order, False if not.
2413 2413
2414 2414 This is part of the mandatory API for smartset."""
2415 2415 return self._ascending is not None and self._ascending
2416 2416
2417 2417 def isdescending(self):
2418 2418 """Returns True if the collection is descending order, False if not.
2419 2419
2420 2420 This is part of the mandatory API for smartset."""
2421 2421 return self._ascending is not None and not self._ascending
2422 2422
2423 2423 def first(self):
2424 2424 if self:
2425 2425 if self._ascending is None:
2426 2426 return self._list[0]
2427 2427 elif self._ascending:
2428 2428 return self._asclist[0]
2429 2429 else:
2430 2430 return self._asclist[-1]
2431 2431 return None
2432 2432
2433 2433 def last(self):
2434 2434 if self:
2435 2435 if self._ascending is None:
2436 2436 return self._list[-1]
2437 2437 elif self._ascending:
2438 2438 return self._asclist[-1]
2439 2439 else:
2440 2440 return self._asclist[0]
2441 2441 return None
2442 2442
2443 2443 class filteredset(abstractsmartset):
2444 2444 """Duck type for baseset class which iterates lazily over the revisions in
2445 2445 the subset and contains a function which tests for membership in the
2446 2446 revset
2447 2447 """
2448 2448 def __init__(self, subset, condition=lambda x: True, ascending=None):
2449 2449 """
2450 2450 condition: a function that decide whether a revision in the subset
2451 2451 belongs to the revset or not.
2452 2452 """
2453 2453 self._subset = subset
2454 2454 self._condition = condition
2455 2455 self._cache = {}
2456 2456 if ascending is not None:
2457 2457 ascending = bool(ascending)
2458 2458 self._ascending = ascending
2459 2459
2460 2460 def __contains__(self, x):
2461 2461 c = self._cache
2462 2462 if x not in c:
2463 2463 v = c[x] = x in self._subset and self._condition(x)
2464 2464 return v
2465 2465 return c[x]
2466 2466
2467 2467 def __iter__(self):
2468 2468 return self._iterfilter(self._subset)
2469 2469
2470 2470 def _iterfilter(self, it):
2471 2471 cond = self._condition
2472 2472 for x in it:
2473 2473 if cond(x):
2474 2474 yield x
2475 2475
2476 2476 @property
2477 2477 def fastasc(self):
2478 2478 it = self._subset.fastasc
2479 2479 if it is None:
2480 2480 return None
2481 2481 return lambda: self._iterfilter(it())
2482 2482
2483 2483 @property
2484 2484 def fastdesc(self):
2485 2485 it = self._subset.fastdesc
2486 2486 if it is None:
2487 2487 return None
2488 2488 return lambda: self._iterfilter(it())
2489 2489
2490 2490 def __nonzero__(self):
2491 2491 for r in self:
2492 2492 return True
2493 2493 return False
2494 2494
2495 2495 def __len__(self):
2496 2496 # Basic implementation to be changed in future patches.
2497 2497 l = baseset([r for r in self])
2498 2498 return len(l)
2499 2499
2500 2500 def __getitem__(self, x):
2501 2501 # Basic implementation to be changed in future patches.
2502 2502 l = baseset([r for r in self])
2503 2503 return l[x]
2504 2504
2505 2505 def sort(self, reverse=False):
2506 2506 if self._ascending is None:
2507 2507 if not util.safehasattr(self._subset, 'sort'):
2508 2508 self._subset = baseset(self._subset)
2509 2509 self._subset.sort(reverse=reverse)
2510 2510 self._ascending = not reverse
2511 2511 elif bool(reverse) == self._ascending:
2512 2512 self.reverse()
2513 2513
2514 2514 def reverse(self):
2515 2515 self._subset.reverse()
2516 2516 if self._ascending is not None:
2517 2517 self._ascending = not self._ascending
2518 2518
2519 2519 def set(self):
2520 2520 return set([r for r in self])
2521 2521
2522 2522 def isascending(self):
2523 2523 return self._ascending is not None and self._ascending
2524 2524
2525 2525 def isdescending(self):
2526 2526 return self._ascending is not None and not self._ascending
2527 2527
2528 2528 def first(self):
2529 2529 for x in self:
2530 2530 return x
2531 2531 return None
2532 2532
2533 2533 def last(self):
2534 2534 it = None
2535 2535 if self._ascending is not None:
2536 2536 if self._ascending:
2537 2537 it = self.fastdesc
2538 2538 else:
2539 2539 it = self.fastasc
2540 2540 if it is None:
2541 2541 # slowly consume everything. This needs improvement
2542 2542 it = lambda: reversed(list(self))
2543 2543 for x in it():
2544 2544 return x
2545 2545 return None
2546 2546
2547 2547 class addset(abstractsmartset):
2548 2548 """Represent the addition of two sets
2549 2549
2550 2550 Wrapper structure for lazily adding two structures without losing much
2551 2551 performance on the __contains__ method
2552 2552
2553 2553 If the ascending attribute is set, that means the two structures are
2554 2554 ordered in either an ascending or descending way. Therefore, we can add
2555 2555 them maintaining the order by iterating over both at the same time
2556 2556 """
2557 2557 def __init__(self, revs1, revs2, ascending=None):
2558 2558 self._r1 = revs1
2559 2559 self._r2 = revs2
2560 2560 self._iter = None
2561 2561 self._ascending = ascending
2562 2562 self._genlist = None
2563 2563
2564 2564 def __len__(self):
2565 2565 return len(self._list)
2566 2566
2567 2567 def __nonzero__(self):
2568 2568 return bool(self._r1 or self._r2)
2569 2569
2570 2570 @util.propertycache
2571 2571 def _list(self):
2572 2572 if not self._genlist:
2573 2573 self._genlist = baseset(self._iterator())
2574 2574 return self._genlist
2575 2575
2576 2576 def _iterator(self):
2577 2577 """Iterate over both collections without repeating elements
2578 2578
2579 2579 If the ascending attribute is not set, iterate over the first one and
2580 2580 then over the second one checking for membership on the first one so we
2581 2581 dont yield any duplicates.
2582 2582
2583 2583 If the ascending attribute is set, iterate over both collections at the
2584 2584 same time, yielding only one value at a time in the given order.
2585 2585 """
2586 2586 if self._ascending is None:
2587 2587 def gen():
2588 2588 for r in self._r1:
2589 2589 yield r
2590 2590 s = self._r1.set()
2591 2591 for r in self._r2:
2592 2592 if r not in s:
2593 2593 yield r
2594 2594 gen = gen()
2595 2595 else:
2596 2596 iter1 = iter(self._r1)
2597 2597 iter2 = iter(self._r2)
2598 2598 gen = self._iterordered(self._ascending, iter1, iter2)
2599 2599 return gen
2600 2600
2601 2601 def __iter__(self):
2602 2602 if self._genlist:
2603 2603 return iter(self._genlist)
2604 2604 return iter(self._iterator())
2605 2605
2606 2606 @property
2607 2607 def fastasc(self):
2608 2608 iter1 = self._r1.fastasc
2609 2609 iter2 = self._r2.fastasc
2610 2610 if None in (iter1, iter2):
2611 2611 return None
2612 2612 return lambda: self._iterordered(True, iter1(), iter2())
2613 2613
2614 2614 @property
2615 2615 def fastdesc(self):
2616 2616 iter1 = self._r1.fastdesc
2617 2617 iter2 = self._r2.fastdesc
2618 2618 if None in (iter1, iter2):
2619 2619 return None
2620 2620 return lambda: self._iterordered(False, iter1(), iter2())
2621 2621
2622 2622 def _iterordered(self, ascending, iter1, iter2):
2623 2623 """produce an ordered iteration from two iterators with the same order
2624 2624
2625 2625 The ascending is used to indicated the iteration direction.
2626 2626 """
2627 2627 choice = max
2628 2628 if ascending:
2629 2629 choice = min
2630 2630
2631 2631 val1 = None
2632 2632 val2 = None
2633 2633
2634 2634 choice = max
2635 2635 if ascending:
2636 2636 choice = min
2637 2637 try:
2638 2638 # Consume both iterators in an ordered way until one is
2639 2639 # empty
2640 2640 while True:
2641 2641 if val1 is None:
2642 2642 val1 = iter1.next()
2643 2643 if val2 is None:
2644 2644 val2 = iter2.next()
2645 2645 next = choice(val1, val2)
2646 2646 yield next
2647 2647 if val1 == next:
2648 2648 val1 = None
2649 2649 if val2 == next:
2650 2650 val2 = None
2651 2651 except StopIteration:
2652 2652 # Flush any remaining values and consume the other one
2653 2653 it = iter2
2654 2654 if val1 is not None:
2655 2655 yield val1
2656 2656 it = iter1
2657 2657 elif val2 is not None:
2658 2658 # might have been equality and both are empty
2659 2659 yield val2
2660 2660 for val in it:
2661 2661 yield val
2662 2662
2663 2663 def __contains__(self, x):
2664 2664 return x in self._r1 or x in self._r2
2665 2665
2666 2666 def set(self):
2667 2667 return self
2668 2668
2669 2669 def sort(self, reverse=False):
2670 2670 """Sort the added set
2671 2671
2672 2672 For this we use the cached list with all the generated values and if we
2673 2673 know they are ascending or descending we can sort them in a smart way.
2674 2674 """
2675 2675 if self._ascending is None:
2676 2676 self._list.sort(reverse=reverse)
2677 2677 self._ascending = not reverse
2678 2678 else:
2679 2679 if bool(self._ascending) == bool(reverse):
2680 2680 self.reverse()
2681 2681
2682 2682 def isascending(self):
2683 2683 return self._ascending is not None and self._ascending
2684 2684
2685 2685 def isdescending(self):
2686 2686 return self._ascending is not None and not self._ascending
2687 2687
2688 2688 def reverse(self):
2689 2689 self._list.reverse()
2690 2690 if self._ascending is not None:
2691 2691 self._ascending = not self._ascending
2692 2692
2693 2693 def first(self):
2694 2694 if self:
2695 2695 return self._list.first()
2696 2696 return None
2697 2697
2698 2698 def last(self):
2699 2699 if self:
2700 2700 return self._list.last()
2701 2701 return None
2702 2702
2703 2703 class generatorset(abstractsmartset):
2704 2704 """Wrap a generator for lazy iteration
2705 2705
2706 2706 Wrapper structure for generators that provides lazy membership and can
2707 2707 be iterated more than once.
2708 2708 When asked for membership it generates values until either it finds the
2709 2709 requested one or has gone through all the elements in the generator
2710 2710 """
2711 2711 def __init__(self, gen, iterasc=None):
2712 2712 """
2713 2713 gen: a generator producing the values for the generatorset.
2714 2714 """
2715 2715 self._gen = gen
2716 2716 self._asclist = None
2717 2717 self._cache = {}
2718 2718 self._genlist = []
2719 2719 self._finished = False
2720 2720 self._ascending = True
2721 2721 if iterasc is not None:
2722 2722 if iterasc:
2723 2723 self.fastasc = self._iterator
2724 2724 self.__contains__ = self._asccontains
2725 2725 else:
2726 2726 self.fastdesc = self._iterator
2727 2727 self.__contains__ = self._desccontains
2728 2728
2729 2729 def __nonzero__(self):
2730 2730 for r in self:
2731 2731 return True
2732 2732 return False
2733 2733
2734 2734 def __contains__(self, x):
2735 2735 if x in self._cache:
2736 2736 return self._cache[x]
2737 2737
2738 2738 # Use new values only, as existing values would be cached.
2739 2739 for l in self._consumegen():
2740 2740 if l == x:
2741 2741 return True
2742 2742
2743 2743 self._cache[x] = False
2744 2744 return False
2745 2745
2746 2746 def _asccontains(self, x):
2747 2747 """version of contains optimised for ascending generator"""
2748 2748 if x in self._cache:
2749 2749 return self._cache[x]
2750 2750
2751 2751 # Use new values only, as existing values would be cached.
2752 2752 for l in self._consumegen():
2753 2753 if l == x:
2754 2754 return True
2755 2755 if l > x:
2756 2756 break
2757 2757
2758 2758 self._cache[x] = False
2759 2759 return False
2760 2760
2761 2761 def _desccontains(self, x):
2762 2762 """version of contains optimised for descending generator"""
2763 2763 if x in self._cache:
2764 2764 return self._cache[x]
2765 2765
2766 2766 # Use new values only, as existing values would be cached.
2767 2767 for l in self._consumegen():
2768 2768 if l == x:
2769 2769 return True
2770 2770 if l < x:
2771 2771 break
2772 2772
2773 2773 self._cache[x] = False
2774 2774 return False
2775 2775
2776 2776 def __iter__(self):
2777 2777 if self._ascending:
2778 2778 it = self.fastasc
2779 2779 else:
2780 2780 it = self.fastdesc
2781 2781 if it is not None:
2782 2782 return it()
2783 2783 # we need to consume the iterator
2784 2784 for x in self._consumegen():
2785 2785 pass
2786 2786 # recall the same code
2787 2787 return iter(self)
2788 2788
2789 2789 def _iterator(self):
2790 2790 if self._finished:
2791 2791 return iter(self._genlist)
2792 2792
2793 2793 # We have to use this complex iteration strategy to allow multiple
2794 2794 # iterations at the same time. We need to be able to catch revision
2795 2795 # removed from `consumegen` and added to genlist in another instance.
2796 2796 #
2797 2797 # Getting rid of it would provide an about 15% speed up on this
2798 2798 # iteration.
2799 2799 genlist = self._genlist
2800 2800 nextrev = self._consumegen().next
2801 2801 _len = len # cache global lookup
2802 2802 def gen():
2803 2803 i = 0
2804 2804 while True:
2805 2805 if i < _len(genlist):
2806 2806 yield genlist[i]
2807 2807 else:
2808 2808 yield nextrev()
2809 2809 i += 1
2810 2810 return gen()
2811 2811
2812 2812 def _consumegen(self):
2813 2813 cache = self._cache
2814 2814 genlist = self._genlist.append
2815 2815 for item in self._gen:
2816 2816 cache[item] = True
2817 2817 genlist(item)
2818 2818 yield item
2819 2819 if not self._finished:
2820 2820 self._finished = True
2821 2821 asc = self._genlist[:]
2822 2822 asc.sort()
2823 2823 self._asclist = asc
2824 2824 self.fastasc = asc.__iter__
2825 2825 self.fastdesc = asc.__reversed__
2826 2826
2827 2827 def set(self):
2828 2828 return self
2829 2829
2830 2830 def sort(self, reverse=False):
2831 2831 self._ascending = not reverse
2832 2832
2833 2833 def reverse(self):
2834 2834 self._ascending = not self._ascending
2835 2835
2836 2836 def isascending(self):
2837 2837 return self._ascending
2838 2838
2839 2839 def isdescending(self):
2840 2840 return not self._ascending
2841 2841
2842 2842 def first(self):
2843 2843 if self._ascending:
2844 2844 it = self.fastasc
2845 2845 else:
2846 2846 it = self.fastdesc
2847 2847 if it is None:
2848 2848 # we need to consume all and try again
2849 2849 for x in self._consumegen():
2850 2850 pass
2851 2851 return self.first()
2852 2852 if self:
2853 2853 return it.next()
2854 2854 return None
2855 2855
2856 2856 def last(self):
2857 2857 if self._ascending:
2858 2858 it = self.fastdesc
2859 2859 else:
2860 2860 it = self.fastasc
2861 2861 if it is None:
2862 2862 # we need to consume all and try again
2863 2863 for x in self._consumegen():
2864 2864 pass
2865 2865 return self.first()
2866 2866 if self:
2867 2867 return it.next()
2868 2868 return None
2869 2869
2870 2870 def spanset(repo, start=None, end=None):
2871 2871 """factory function to dispatch between fullreposet and actual spanset
2872 2872
2873 2873 Feel free to update all spanset call sites and kill this function at some
2874 2874 point.
2875 2875 """
2876 2876 if start is None and end is None:
2877 2877 return fullreposet(repo)
2878 2878 return _spanset(repo, start, end)
2879 2879
2880 2880
2881 2881 class _spanset(abstractsmartset):
2882 2882 """Duck type for baseset class which represents a range of revisions and
2883 2883 can work lazily and without having all the range in memory
2884 2884
2885 2885 Note that spanset(x, y) behave almost like xrange(x, y) except for two
2886 2886 notable points:
2887 2887 - when x < y it will be automatically descending,
2888 2888 - revision filtered with this repoview will be skipped.
2889 2889
2890 2890 """
2891 2891 def __init__(self, repo, start=0, end=None):
2892 2892 """
2893 2893 start: first revision included the set
2894 2894 (default to 0)
2895 2895 end: first revision excluded (last+1)
2896 2896 (default to len(repo)
2897 2897
2898 2898 Spanset will be descending if `end` < `start`.
2899 2899 """
2900 2900 if end is None:
2901 2901 end = len(repo)
2902 2902 self._ascending = start <= end
2903 2903 if not self._ascending:
2904 2904 start, end = end + 1, start +1
2905 2905 self._start = start
2906 2906 self._end = end
2907 2907 self._hiddenrevs = repo.changelog.filteredrevs
2908 2908
2909 2909 def sort(self, reverse=False):
2910 2910 self._ascending = not reverse
2911 2911
2912 2912 def reverse(self):
2913 2913 self._ascending = not self._ascending
2914 2914
2915 2915 def _iterfilter(self, iterrange):
2916 2916 s = self._hiddenrevs
2917 2917 for r in iterrange:
2918 2918 if r not in s:
2919 2919 yield r
2920 2920
2921 2921 def __iter__(self):
2922 2922 if self._ascending:
2923 2923 return self.fastasc()
2924 2924 else:
2925 2925 return self.fastdesc()
2926 2926
2927 2927 def fastasc(self):
2928 2928 iterrange = xrange(self._start, self._end)
2929 2929 if self._hiddenrevs:
2930 2930 return self._iterfilter(iterrange)
2931 2931 return iter(iterrange)
2932 2932
2933 2933 def fastdesc(self):
2934 2934 iterrange = xrange(self._end - 1, self._start - 1, -1)
2935 2935 if self._hiddenrevs:
2936 2936 return self._iterfilter(iterrange)
2937 2937 return iter(iterrange)
2938 2938
2939 2939 def __contains__(self, rev):
2940 2940 hidden = self._hiddenrevs
2941 2941 return ((self._start <= rev < self._end)
2942 2942 and not (hidden and rev in hidden))
2943 2943
2944 2944 def __nonzero__(self):
2945 2945 for r in self:
2946 2946 return True
2947 2947 return False
2948 2948
2949 2949 def __len__(self):
2950 2950 if not self._hiddenrevs:
2951 2951 return abs(self._end - self._start)
2952 2952 else:
2953 2953 count = 0
2954 2954 start = self._start
2955 2955 end = self._end
2956 2956 for rev in self._hiddenrevs:
2957 2957 if (end < rev <= start) or (start <= rev < end):
2958 2958 count += 1
2959 2959 return abs(self._end - self._start) - count
2960 2960
2961 2961 def __getitem__(self, x):
2962 2962 # Basic implementation to be changed in future patches.
2963 2963 l = baseset([r for r in self])
2964 2964 return l[x]
2965 2965
2966 2966 def set(self):
2967 2967 return self
2968 2968
2969 2969 def isascending(self):
2970 2970 return self._start <= self._end
2971 2971
2972 2972 def isdescending(self):
2973 2973 return self._start >= self._end
2974 2974
2975 2975 def first(self):
2976 2976 if self._ascending:
2977 2977 it = self.fastasc
2978 2978 else:
2979 2979 it = self.fastdesc
2980 2980 for x in it():
2981 2981 return x
2982 2982 return None
2983 2983
2984 2984 def last(self):
2985 2985 if self._ascending:
2986 2986 it = self.fastdesc
2987 2987 else:
2988 2988 it = self.fastasc
2989 2989 for x in it():
2990 2990 return x
2991 2991 return None
2992 2992
2993 2993 class fullreposet(_spanset):
2994 2994 """a set containing all revisions in the repo
2995 2995
2996 2996 This class exists to host special optimisation.
2997 2997 """
2998 2998
2999 2999 def __init__(self, repo):
3000 3000 super(fullreposet, self).__init__(repo)
3001 3001
3002 3002 def __and__(self, other):
3003 3003 """fullrepo & other -> other
3004 3004
3005 3005 As self contains the whole repo, all of the other set should also be in
3006 3006 self. Therefor `self & other = other`.
3007 3007
3008 3008 This boldly assumes the other contains valid revs only.
3009 3009 """
3010 3010 # other not a smartset, make is so
3011 3011 if not util.safehasattr(other, 'set'):
3012 3012 # filter out hidden revision
3013 3013 # (this boldly assumes all smartset are pure)
3014 3014 #
3015 3015 # `other` was used with "&", let's assume this is a set like
3016 3016 # object.
3017 3017 other = baseset(other - self._hiddenrevs)
3018 3018 elif not util.safehasattr(other, 'ascending'):
3019 3019 # "other" is generatorset not a real smart set
3020 3020 # we fallback to the old way (sad kitten)
3021 3021 return super(fullreposet, self).__and__(other)
3022 3022
3023 # preserve order:
3024 #
3025 # this is probably useless and harmful in multiple cases but matches
3026 # the current behavior.
3027 3023 if self.isascending():
3028 other.ascending()
3024 other.sort()
3029 3025 else:
3030 other.descending()
3026 other.sort(reverse)
3031 3027 return other
3032 3028
3033 3029 # tell hggettext to extract docstrings from these functions:
3034 3030 i18nfunctions = symbols.values()
General Comments 0
You need to be logged in to leave comments. Login now