##// END OF EJS Templates
revset: refactor parents() into a single return point...
Pierre-Yves David -
r22496:35af9361 default
parent child Browse files
Show More
@@ -1,2851 +1,2850 b''
1 1 # revset.py - revision set queries for mercurial
2 2 #
3 3 # Copyright 2010 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 import re
9 9 import parser, util, error, discovery, hbisect, phases
10 10 import node
11 11 import heapq
12 12 import match as matchmod
13 13 import ancestor as ancestormod
14 14 from i18n import _
15 15 import encoding
16 16 import obsolete as obsmod
17 17 import pathutil
18 18 import repoview
19 19
20 20 def _revancestors(repo, revs, followfirst):
21 21 """Like revlog.ancestors(), but supports followfirst."""
22 22 cut = followfirst and 1 or None
23 23 cl = repo.changelog
24 24
25 25 def iterate():
26 26 revqueue, revsnode = None, None
27 27 h = []
28 28
29 29 revs.descending()
30 30 revqueue = util.deque(revs)
31 31 if revqueue:
32 32 revsnode = revqueue.popleft()
33 33 heapq.heappush(h, -revsnode)
34 34
35 35 seen = set([node.nullrev])
36 36 while h:
37 37 current = -heapq.heappop(h)
38 38 if current not in seen:
39 39 if revsnode and current == revsnode:
40 40 if revqueue:
41 41 revsnode = revqueue.popleft()
42 42 heapq.heappush(h, -revsnode)
43 43 seen.add(current)
44 44 yield current
45 45 for parent in cl.parentrevs(current)[:cut]:
46 46 if parent != node.nullrev:
47 47 heapq.heappush(h, -parent)
48 48
49 49 return _descgeneratorset(iterate())
50 50
51 51 def _revdescendants(repo, revs, followfirst):
52 52 """Like revlog.descendants() but supports followfirst."""
53 53 cut = followfirst and 1 or None
54 54
55 55 def iterate():
56 56 cl = repo.changelog
57 57 first = min(revs)
58 58 nullrev = node.nullrev
59 59 if first == nullrev:
60 60 # Are there nodes with a null first parent and a non-null
61 61 # second one? Maybe. Do we care? Probably not.
62 62 for i in cl:
63 63 yield i
64 64 else:
65 65 seen = set(revs)
66 66 for i in cl.revs(first + 1):
67 67 for x in cl.parentrevs(i)[:cut]:
68 68 if x != nullrev and x in seen:
69 69 seen.add(i)
70 70 yield i
71 71 break
72 72
73 73 return _ascgeneratorset(iterate())
74 74
75 75 def _revsbetween(repo, roots, heads):
76 76 """Return all paths between roots and heads, inclusive of both endpoint
77 77 sets."""
78 78 if not roots:
79 79 return baseset([])
80 80 parentrevs = repo.changelog.parentrevs
81 81 visit = list(heads)
82 82 reachable = set()
83 83 seen = {}
84 84 minroot = min(roots)
85 85 roots = set(roots)
86 86 # open-code the post-order traversal due to the tiny size of
87 87 # sys.getrecursionlimit()
88 88 while visit:
89 89 rev = visit.pop()
90 90 if rev in roots:
91 91 reachable.add(rev)
92 92 parents = parentrevs(rev)
93 93 seen[rev] = parents
94 94 for parent in parents:
95 95 if parent >= minroot and parent not in seen:
96 96 visit.append(parent)
97 97 if not reachable:
98 98 return baseset([])
99 99 for rev in sorted(seen):
100 100 for parent in seen[rev]:
101 101 if parent in reachable:
102 102 reachable.add(rev)
103 103 return baseset(sorted(reachable))
104 104
105 105 elements = {
106 106 "(": (20, ("group", 1, ")"), ("func", 1, ")")),
107 107 "~": (18, None, ("ancestor", 18)),
108 108 "^": (18, None, ("parent", 18), ("parentpost", 18)),
109 109 "-": (5, ("negate", 19), ("minus", 5)),
110 110 "::": (17, ("dagrangepre", 17), ("dagrange", 17),
111 111 ("dagrangepost", 17)),
112 112 "..": (17, ("dagrangepre", 17), ("dagrange", 17),
113 113 ("dagrangepost", 17)),
114 114 ":": (15, ("rangepre", 15), ("range", 15), ("rangepost", 15)),
115 115 "not": (10, ("not", 10)),
116 116 "!": (10, ("not", 10)),
117 117 "and": (5, None, ("and", 5)),
118 118 "&": (5, None, ("and", 5)),
119 119 "or": (4, None, ("or", 4)),
120 120 "|": (4, None, ("or", 4)),
121 121 "+": (4, None, ("or", 4)),
122 122 ",": (2, None, ("list", 2)),
123 123 ")": (0, None, None),
124 124 "symbol": (0, ("symbol",), None),
125 125 "string": (0, ("string",), None),
126 126 "end": (0, None, None),
127 127 }
128 128
129 129 keywords = set(['and', 'or', 'not'])
130 130
131 131 def tokenize(program, lookup=None):
132 132 '''
133 133 Parse a revset statement into a stream of tokens
134 134
135 135 Check that @ is a valid unquoted token character (issue3686):
136 136 >>> list(tokenize("@::"))
137 137 [('symbol', '@', 0), ('::', None, 1), ('end', None, 3)]
138 138
139 139 '''
140 140
141 141 pos, l = 0, len(program)
142 142 while pos < l:
143 143 c = program[pos]
144 144 if c.isspace(): # skip inter-token whitespace
145 145 pass
146 146 elif c == ':' and program[pos:pos + 2] == '::': # look ahead carefully
147 147 yield ('::', None, pos)
148 148 pos += 1 # skip ahead
149 149 elif c == '.' and program[pos:pos + 2] == '..': # look ahead carefully
150 150 yield ('..', None, pos)
151 151 pos += 1 # skip ahead
152 152 elif c in "():,-|&+!~^": # handle simple operators
153 153 yield (c, None, pos)
154 154 elif (c in '"\'' or c == 'r' and
155 155 program[pos:pos + 2] in ("r'", 'r"')): # handle quoted strings
156 156 if c == 'r':
157 157 pos += 1
158 158 c = program[pos]
159 159 decode = lambda x: x
160 160 else:
161 161 decode = lambda x: x.decode('string-escape')
162 162 pos += 1
163 163 s = pos
164 164 while pos < l: # find closing quote
165 165 d = program[pos]
166 166 if d == '\\': # skip over escaped characters
167 167 pos += 2
168 168 continue
169 169 if d == c:
170 170 yield ('string', decode(program[s:pos]), s)
171 171 break
172 172 pos += 1
173 173 else:
174 174 raise error.ParseError(_("unterminated string"), s)
175 175 # gather up a symbol/keyword
176 176 elif c.isalnum() or c in '._@' or ord(c) > 127:
177 177 s = pos
178 178 pos += 1
179 179 while pos < l: # find end of symbol
180 180 d = program[pos]
181 181 if not (d.isalnum() or d in "-._/@" or ord(d) > 127):
182 182 break
183 183 if d == '.' and program[pos - 1] == '.': # special case for ..
184 184 pos -= 1
185 185 break
186 186 pos += 1
187 187 sym = program[s:pos]
188 188 if sym in keywords: # operator keywords
189 189 yield (sym, None, s)
190 190 elif '-' in sym:
191 191 # some jerk gave us foo-bar-baz, try to check if it's a symbol
192 192 if lookup and lookup(sym):
193 193 # looks like a real symbol
194 194 yield ('symbol', sym, s)
195 195 else:
196 196 # looks like an expression
197 197 parts = sym.split('-')
198 198 for p in parts[:-1]:
199 199 if p: # possible consecutive -
200 200 yield ('symbol', p, s)
201 201 s += len(p)
202 202 yield ('-', None, pos)
203 203 s += 1
204 204 if parts[-1]: # possible trailing -
205 205 yield ('symbol', parts[-1], s)
206 206 else:
207 207 yield ('symbol', sym, s)
208 208 pos -= 1
209 209 else:
210 210 raise error.ParseError(_("syntax error"), pos)
211 211 pos += 1
212 212 yield ('end', None, pos)
213 213
214 214 # helpers
215 215
216 216 def getstring(x, err):
217 217 if x and (x[0] == 'string' or x[0] == 'symbol'):
218 218 return x[1]
219 219 raise error.ParseError(err)
220 220
221 221 def getlist(x):
222 222 if not x:
223 223 return []
224 224 if x[0] == 'list':
225 225 return getlist(x[1]) + [x[2]]
226 226 return [x]
227 227
228 228 def getargs(x, min, max, err):
229 229 l = getlist(x)
230 230 if len(l) < min or (max >= 0 and len(l) > max):
231 231 raise error.ParseError(err)
232 232 return l
233 233
234 234 def getset(repo, subset, x):
235 235 if not x:
236 236 raise error.ParseError(_("missing argument"))
237 237 s = methods[x[0]](repo, subset, *x[1:])
238 238 if util.safehasattr(s, 'set'):
239 239 return s
240 240 return baseset(s)
241 241
242 242 def _getrevsource(repo, r):
243 243 extra = repo[r].extra()
244 244 for label in ('source', 'transplant_source', 'rebase_source'):
245 245 if label in extra:
246 246 try:
247 247 return repo[extra[label]].rev()
248 248 except error.RepoLookupError:
249 249 pass
250 250 return None
251 251
252 252 # operator methods
253 253
254 254 def stringset(repo, subset, x):
255 255 x = repo[x].rev()
256 256 if x == -1 and len(subset) == len(repo):
257 257 return baseset([-1])
258 258 if len(subset) == len(repo) or x in subset:
259 259 return baseset([x])
260 260 return baseset([])
261 261
262 262 def symbolset(repo, subset, x):
263 263 if x in symbols:
264 264 raise error.ParseError(_("can't use %s here") % x)
265 265 return stringset(repo, subset, x)
266 266
267 267 def rangeset(repo, subset, x, y):
268 268 cl = baseset(repo.changelog)
269 269 m = getset(repo, cl, x)
270 270 n = getset(repo, cl, y)
271 271
272 272 if not m or not n:
273 273 return baseset([])
274 274 m, n = m[0], n[-1]
275 275
276 276 if m < n:
277 277 r = spanset(repo, m, n + 1)
278 278 else:
279 279 r = spanset(repo, m, n - 1)
280 280 return r & subset
281 281
282 282 def dagrange(repo, subset, x, y):
283 283 r = spanset(repo)
284 284 xs = _revsbetween(repo, getset(repo, r, x), getset(repo, r, y))
285 285 s = subset.set()
286 286 return xs.filter(s.__contains__)
287 287
288 288 def andset(repo, subset, x, y):
289 289 return getset(repo, getset(repo, subset, x), y)
290 290
291 291 def orset(repo, subset, x, y):
292 292 xl = getset(repo, subset, x)
293 293 yl = getset(repo, subset - xl, y)
294 294 return xl + yl
295 295
296 296 def notset(repo, subset, x):
297 297 return subset - getset(repo, subset, x)
298 298
299 299 def listset(repo, subset, a, b):
300 300 raise error.ParseError(_("can't use a list in this context"))
301 301
302 302 def func(repo, subset, a, b):
303 303 if a[0] == 'symbol' and a[1] in symbols:
304 304 return symbols[a[1]](repo, subset, b)
305 305 raise error.ParseError(_("not a function: %s") % a[1])
306 306
307 307 # functions
308 308
309 309 def adds(repo, subset, x):
310 310 """``adds(pattern)``
311 311 Changesets that add a file matching pattern.
312 312
313 313 The pattern without explicit kind like ``glob:`` is expected to be
314 314 relative to the current directory and match against a file or a
315 315 directory.
316 316 """
317 317 # i18n: "adds" is a keyword
318 318 pat = getstring(x, _("adds requires a pattern"))
319 319 return checkstatus(repo, subset, pat, 1)
320 320
321 321 def ancestor(repo, subset, x):
322 322 """``ancestor(*changeset)``
323 323 A greatest common ancestor of the changesets.
324 324
325 325 Accepts 0 or more changesets.
326 326 Will return empty list when passed no args.
327 327 Greatest common ancestor of a single changeset is that changeset.
328 328 """
329 329 # i18n: "ancestor" is a keyword
330 330 l = getlist(x)
331 331 rl = spanset(repo)
332 332 anc = None
333 333
334 334 # (getset(repo, rl, i) for i in l) generates a list of lists
335 335 for revs in (getset(repo, rl, i) for i in l):
336 336 for r in revs:
337 337 if anc is None:
338 338 anc = repo[r]
339 339 else:
340 340 anc = anc.ancestor(repo[r])
341 341
342 342 if anc is not None and anc.rev() in subset:
343 343 return baseset([anc.rev()])
344 344 return baseset([])
345 345
346 346 def _ancestors(repo, subset, x, followfirst=False):
347 347 args = getset(repo, spanset(repo), x)
348 348 if not args:
349 349 return baseset([])
350 350 s = _revancestors(repo, args, followfirst)
351 351 return subset.filter(s.__contains__)
352 352
353 353 def ancestors(repo, subset, x):
354 354 """``ancestors(set)``
355 355 Changesets that are ancestors of a changeset in set.
356 356 """
357 357 return _ancestors(repo, subset, x)
358 358
359 359 def _firstancestors(repo, subset, x):
360 360 # ``_firstancestors(set)``
361 361 # Like ``ancestors(set)`` but follows only the first parents.
362 362 return _ancestors(repo, subset, x, followfirst=True)
363 363
364 364 def ancestorspec(repo, subset, x, n):
365 365 """``set~n``
366 366 Changesets that are the Nth ancestor (first parents only) of a changeset
367 367 in set.
368 368 """
369 369 try:
370 370 n = int(n[1])
371 371 except (TypeError, ValueError):
372 372 raise error.ParseError(_("~ expects a number"))
373 373 ps = set()
374 374 cl = repo.changelog
375 375 for r in getset(repo, baseset(cl), x):
376 376 for i in range(n):
377 377 r = cl.parentrevs(r)[0]
378 378 ps.add(r)
379 379 return subset.filter(ps.__contains__)
380 380
381 381 def author(repo, subset, x):
382 382 """``author(string)``
383 383 Alias for ``user(string)``.
384 384 """
385 385 # i18n: "author" is a keyword
386 386 n = encoding.lower(getstring(x, _("author requires a string")))
387 387 kind, pattern, matcher = _substringmatcher(n)
388 388 return subset.filter(lambda x: matcher(encoding.lower(repo[x].user())))
389 389
390 390 def only(repo, subset, x):
391 391 """``only(set, [set])``
392 392 Changesets that are ancestors of the first set that are not ancestors
393 393 of any other head in the repo. If a second set is specified, the result
394 394 is ancestors of the first set that are not ancestors of the second set
395 395 (i.e. ::<set1> - ::<set2>).
396 396 """
397 397 cl = repo.changelog
398 398 # i18n: "only" is a keyword
399 399 args = getargs(x, 1, 2, _('only takes one or two arguments'))
400 400 include = getset(repo, spanset(repo), args[0]).set()
401 401 if len(args) == 1:
402 402 if len(include) == 0:
403 403 return baseset([])
404 404
405 405 descendants = set(_revdescendants(repo, include, False))
406 406 exclude = [rev for rev in cl.headrevs()
407 407 if not rev in descendants and not rev in include]
408 408 else:
409 409 exclude = getset(repo, spanset(repo), args[1])
410 410
411 411 results = set(ancestormod.missingancestors(include, exclude, cl.parentrevs))
412 412 return lazyset(subset, results.__contains__)
413 413
414 414 def bisect(repo, subset, x):
415 415 """``bisect(string)``
416 416 Changesets marked in the specified bisect status:
417 417
418 418 - ``good``, ``bad``, ``skip``: csets explicitly marked as good/bad/skip
419 419 - ``goods``, ``bads`` : csets topologically good/bad
420 420 - ``range`` : csets taking part in the bisection
421 421 - ``pruned`` : csets that are goods, bads or skipped
422 422 - ``untested`` : csets whose fate is yet unknown
423 423 - ``ignored`` : csets ignored due to DAG topology
424 424 - ``current`` : the cset currently being bisected
425 425 """
426 426 # i18n: "bisect" is a keyword
427 427 status = getstring(x, _("bisect requires a string")).lower()
428 428 state = set(hbisect.get(repo, status))
429 429 return subset.filter(state.__contains__)
430 430
431 431 # Backward-compatibility
432 432 # - no help entry so that we do not advertise it any more
433 433 def bisected(repo, subset, x):
434 434 return bisect(repo, subset, x)
435 435
436 436 def bookmark(repo, subset, x):
437 437 """``bookmark([name])``
438 438 The named bookmark or all bookmarks.
439 439
440 440 If `name` starts with `re:`, the remainder of the name is treated as
441 441 a regular expression. To match a bookmark that actually starts with `re:`,
442 442 use the prefix `literal:`.
443 443 """
444 444 # i18n: "bookmark" is a keyword
445 445 args = getargs(x, 0, 1, _('bookmark takes one or no arguments'))
446 446 if args:
447 447 bm = getstring(args[0],
448 448 # i18n: "bookmark" is a keyword
449 449 _('the argument to bookmark must be a string'))
450 450 kind, pattern, matcher = _stringmatcher(bm)
451 451 if kind == 'literal':
452 452 bmrev = repo._bookmarks.get(pattern, None)
453 453 if not bmrev:
454 454 raise util.Abort(_("bookmark '%s' does not exist") % bm)
455 455 bmrev = repo[bmrev].rev()
456 456 return subset.filter(lambda r: r == bmrev)
457 457 else:
458 458 matchrevs = set()
459 459 for name, bmrev in repo._bookmarks.iteritems():
460 460 if matcher(name):
461 461 matchrevs.add(bmrev)
462 462 if not matchrevs:
463 463 raise util.Abort(_("no bookmarks exist that match '%s'")
464 464 % pattern)
465 465 bmrevs = set()
466 466 for bmrev in matchrevs:
467 467 bmrevs.add(repo[bmrev].rev())
468 468 return subset & bmrevs
469 469
470 470 bms = set([repo[r].rev()
471 471 for r in repo._bookmarks.values()])
472 472 return subset.filter(bms.__contains__)
473 473
474 474 def branch(repo, subset, x):
475 475 """``branch(string or set)``
476 476 All changesets belonging to the given branch or the branches of the given
477 477 changesets.
478 478
479 479 If `string` starts with `re:`, the remainder of the name is treated as
480 480 a regular expression. To match a branch that actually starts with `re:`,
481 481 use the prefix `literal:`.
482 482 """
483 483 try:
484 484 b = getstring(x, '')
485 485 except error.ParseError:
486 486 # not a string, but another revspec, e.g. tip()
487 487 pass
488 488 else:
489 489 kind, pattern, matcher = _stringmatcher(b)
490 490 if kind == 'literal':
491 491 # note: falls through to the revspec case if no branch with
492 492 # this name exists
493 493 if pattern in repo.branchmap():
494 494 return subset.filter(lambda r: matcher(repo[r].branch()))
495 495 else:
496 496 return subset.filter(lambda r: matcher(repo[r].branch()))
497 497
498 498 s = getset(repo, spanset(repo), x)
499 499 b = set()
500 500 for r in s:
501 501 b.add(repo[r].branch())
502 502 s = s.set()
503 503 return subset.filter(lambda r: r in s or repo[r].branch() in b)
504 504
505 505 def bumped(repo, subset, x):
506 506 """``bumped()``
507 507 Mutable changesets marked as successors of public changesets.
508 508
509 509 Only non-public and non-obsolete changesets can be `bumped`.
510 510 """
511 511 # i18n: "bumped" is a keyword
512 512 getargs(x, 0, 0, _("bumped takes no arguments"))
513 513 bumped = obsmod.getrevs(repo, 'bumped')
514 514 return subset & bumped
515 515
516 516 def bundle(repo, subset, x):
517 517 """``bundle()``
518 518 Changesets in the bundle.
519 519
520 520 Bundle must be specified by the -R option."""
521 521
522 522 try:
523 523 bundlerevs = repo.changelog.bundlerevs
524 524 except AttributeError:
525 525 raise util.Abort(_("no bundle provided - specify with -R"))
526 526 return subset & bundlerevs
527 527
528 528 def checkstatus(repo, subset, pat, field):
529 529 hasset = matchmod.patkind(pat) == 'set'
530 530
531 531 def matches(x):
532 532 m = None
533 533 fname = None
534 534 c = repo[x]
535 535 if not m or hasset:
536 536 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
537 537 if not m.anypats() and len(m.files()) == 1:
538 538 fname = m.files()[0]
539 539 if fname is not None:
540 540 if fname not in c.files():
541 541 return False
542 542 else:
543 543 for f in c.files():
544 544 if m(f):
545 545 break
546 546 else:
547 547 return False
548 548 files = repo.status(c.p1().node(), c.node())[field]
549 549 if fname is not None:
550 550 if fname in files:
551 551 return True
552 552 else:
553 553 for f in files:
554 554 if m(f):
555 555 return True
556 556
557 557 return subset.filter(matches)
558 558
559 559 def _children(repo, narrow, parentset):
560 560 cs = set()
561 561 if not parentset:
562 562 return baseset(cs)
563 563 pr = repo.changelog.parentrevs
564 564 minrev = min(parentset)
565 565 for r in narrow:
566 566 if r <= minrev:
567 567 continue
568 568 for p in pr(r):
569 569 if p in parentset:
570 570 cs.add(r)
571 571 return baseset(cs)
572 572
573 573 def children(repo, subset, x):
574 574 """``children(set)``
575 575 Child changesets of changesets in set.
576 576 """
577 577 s = getset(repo, baseset(repo), x).set()
578 578 cs = _children(repo, subset, s)
579 579 return subset & cs
580 580
581 581 def closed(repo, subset, x):
582 582 """``closed()``
583 583 Changeset is closed.
584 584 """
585 585 # i18n: "closed" is a keyword
586 586 getargs(x, 0, 0, _("closed takes no arguments"))
587 587 return subset.filter(lambda r: repo[r].closesbranch())
588 588
589 589 def contains(repo, subset, x):
590 590 """``contains(pattern)``
591 591 The revision's manifest contains a file matching pattern (but might not
592 592 modify it). See :hg:`help patterns` for information about file patterns.
593 593
594 594 The pattern without explicit kind like ``glob:`` is expected to be
595 595 relative to the current directory and match against a file exactly
596 596 for efficiency.
597 597 """
598 598 # i18n: "contains" is a keyword
599 599 pat = getstring(x, _("contains requires a pattern"))
600 600
601 601 def matches(x):
602 602 if not matchmod.patkind(pat):
603 603 pats = pathutil.canonpath(repo.root, repo.getcwd(), pat)
604 604 if pats in repo[x]:
605 605 return True
606 606 else:
607 607 c = repo[x]
608 608 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
609 609 for f in c.manifest():
610 610 if m(f):
611 611 return True
612 612 return False
613 613
614 614 return subset.filter(matches)
615 615
616 616 def converted(repo, subset, x):
617 617 """``converted([id])``
618 618 Changesets converted from the given identifier in the old repository if
619 619 present, or all converted changesets if no identifier is specified.
620 620 """
621 621
622 622 # There is exactly no chance of resolving the revision, so do a simple
623 623 # string compare and hope for the best
624 624
625 625 rev = None
626 626 # i18n: "converted" is a keyword
627 627 l = getargs(x, 0, 1, _('converted takes one or no arguments'))
628 628 if l:
629 629 # i18n: "converted" is a keyword
630 630 rev = getstring(l[0], _('converted requires a revision'))
631 631
632 632 def _matchvalue(r):
633 633 source = repo[r].extra().get('convert_revision', None)
634 634 return source is not None and (rev is None or source.startswith(rev))
635 635
636 636 return subset.filter(lambda r: _matchvalue(r))
637 637
638 638 def date(repo, subset, x):
639 639 """``date(interval)``
640 640 Changesets within the interval, see :hg:`help dates`.
641 641 """
642 642 # i18n: "date" is a keyword
643 643 ds = getstring(x, _("date requires a string"))
644 644 dm = util.matchdate(ds)
645 645 return subset.filter(lambda x: dm(repo[x].date()[0]))
646 646
647 647 def desc(repo, subset, x):
648 648 """``desc(string)``
649 649 Search commit message for string. The match is case-insensitive.
650 650 """
651 651 # i18n: "desc" is a keyword
652 652 ds = encoding.lower(getstring(x, _("desc requires a string")))
653 653
654 654 def matches(x):
655 655 c = repo[x]
656 656 return ds in encoding.lower(c.description())
657 657
658 658 return subset.filter(matches)
659 659
660 660 def _descendants(repo, subset, x, followfirst=False):
661 661 args = getset(repo, spanset(repo), x)
662 662 if not args:
663 663 return baseset([])
664 664 s = _revdescendants(repo, args, followfirst)
665 665
666 666 # Both sets need to be ascending in order to lazily return the union
667 667 # in the correct order.
668 668 args.ascending()
669 669 result = (orderedlazyset(s, subset.__contains__, ascending=True) +
670 670 orderedlazyset(args, subset.__contains__, ascending=True))
671 671
672 672 # Wrap result in a lazyset since it's an _addset, which doesn't implement
673 673 # all the necessary functions to be consumed by callers.
674 674 return orderedlazyset(result, lambda r: True, ascending=True)
675 675
676 676 def descendants(repo, subset, x):
677 677 """``descendants(set)``
678 678 Changesets which are descendants of changesets in set.
679 679 """
680 680 return _descendants(repo, subset, x)
681 681
682 682 def _firstdescendants(repo, subset, x):
683 683 # ``_firstdescendants(set)``
684 684 # Like ``descendants(set)`` but follows only the first parents.
685 685 return _descendants(repo, subset, x, followfirst=True)
686 686
687 687 def destination(repo, subset, x):
688 688 """``destination([set])``
689 689 Changesets that were created by a graft, transplant or rebase operation,
690 690 with the given revisions specified as the source. Omitting the optional set
691 691 is the same as passing all().
692 692 """
693 693 if x is not None:
694 694 args = getset(repo, spanset(repo), x).set()
695 695 else:
696 696 args = getall(repo, spanset(repo), x).set()
697 697
698 698 dests = set()
699 699
700 700 # subset contains all of the possible destinations that can be returned, so
701 701 # iterate over them and see if their source(s) were provided in the args.
702 702 # Even if the immediate src of r is not in the args, src's source (or
703 703 # further back) may be. Scanning back further than the immediate src allows
704 704 # transitive transplants and rebases to yield the same results as transitive
705 705 # grafts.
706 706 for r in subset:
707 707 src = _getrevsource(repo, r)
708 708 lineage = None
709 709
710 710 while src is not None:
711 711 if lineage is None:
712 712 lineage = list()
713 713
714 714 lineage.append(r)
715 715
716 716 # The visited lineage is a match if the current source is in the arg
717 717 # set. Since every candidate dest is visited by way of iterating
718 718 # subset, any dests further back in the lineage will be tested by a
719 719 # different iteration over subset. Likewise, if the src was already
720 720 # selected, the current lineage can be selected without going back
721 721 # further.
722 722 if src in args or src in dests:
723 723 dests.update(lineage)
724 724 break
725 725
726 726 r = src
727 727 src = _getrevsource(repo, r)
728 728
729 729 return subset.filter(dests.__contains__)
730 730
731 731 def divergent(repo, subset, x):
732 732 """``divergent()``
733 733 Final successors of changesets with an alternative set of final successors.
734 734 """
735 735 # i18n: "divergent" is a keyword
736 736 getargs(x, 0, 0, _("divergent takes no arguments"))
737 737 divergent = obsmod.getrevs(repo, 'divergent')
738 738 return subset.filter(divergent.__contains__)
739 739
740 740 def draft(repo, subset, x):
741 741 """``draft()``
742 742 Changeset in draft phase."""
743 743 # i18n: "draft" is a keyword
744 744 getargs(x, 0, 0, _("draft takes no arguments"))
745 745 pc = repo._phasecache
746 746 return subset.filter(lambda r: pc.phase(repo, r) == phases.draft)
747 747
748 748 def extinct(repo, subset, x):
749 749 """``extinct()``
750 750 Obsolete changesets with obsolete descendants only.
751 751 """
752 752 # i18n: "extinct" is a keyword
753 753 getargs(x, 0, 0, _("extinct takes no arguments"))
754 754 extincts = obsmod.getrevs(repo, 'extinct')
755 755 return subset & extincts
756 756
757 757 def extra(repo, subset, x):
758 758 """``extra(label, [value])``
759 759 Changesets with the given label in the extra metadata, with the given
760 760 optional value.
761 761
762 762 If `value` starts with `re:`, the remainder of the value is treated as
763 763 a regular expression. To match a value that actually starts with `re:`,
764 764 use the prefix `literal:`.
765 765 """
766 766
767 767 # i18n: "extra" is a keyword
768 768 l = getargs(x, 1, 2, _('extra takes at least 1 and at most 2 arguments'))
769 769 # i18n: "extra" is a keyword
770 770 label = getstring(l[0], _('first argument to extra must be a string'))
771 771 value = None
772 772
773 773 if len(l) > 1:
774 774 # i18n: "extra" is a keyword
775 775 value = getstring(l[1], _('second argument to extra must be a string'))
776 776 kind, value, matcher = _stringmatcher(value)
777 777
778 778 def _matchvalue(r):
779 779 extra = repo[r].extra()
780 780 return label in extra and (value is None or matcher(extra[label]))
781 781
782 782 return subset.filter(lambda r: _matchvalue(r))
783 783
784 784 def filelog(repo, subset, x):
785 785 """``filelog(pattern)``
786 786 Changesets connected to the specified filelog.
787 787
788 788 For performance reasons, visits only revisions mentioned in the file-level
789 789 filelog, rather than filtering through all changesets (much faster, but
790 790 doesn't include deletes or duplicate changes). For a slower, more accurate
791 791 result, use ``file()``.
792 792
793 793 The pattern without explicit kind like ``glob:`` is expected to be
794 794 relative to the current directory and match against a file exactly
795 795 for efficiency.
796 796 """
797 797
798 798 # i18n: "filelog" is a keyword
799 799 pat = getstring(x, _("filelog requires a pattern"))
800 800 s = set()
801 801
802 802 if not matchmod.patkind(pat):
803 803 f = pathutil.canonpath(repo.root, repo.getcwd(), pat)
804 804 fl = repo.file(f)
805 805 for fr in fl:
806 806 s.add(fl.linkrev(fr))
807 807 else:
808 808 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=repo[None])
809 809 for f in repo[None]:
810 810 if m(f):
811 811 fl = repo.file(f)
812 812 for fr in fl:
813 813 s.add(fl.linkrev(fr))
814 814
815 815 return subset.filter(s.__contains__)
816 816
817 817 def first(repo, subset, x):
818 818 """``first(set, [n])``
819 819 An alias for limit().
820 820 """
821 821 return limit(repo, subset, x)
822 822
823 823 def _follow(repo, subset, x, name, followfirst=False):
824 824 l = getargs(x, 0, 1, _("%s takes no arguments or a filename") % name)
825 825 c = repo['.']
826 826 if l:
827 827 x = getstring(l[0], _("%s expected a filename") % name)
828 828 if x in c:
829 829 cx = c[x]
830 830 s = set(ctx.rev() for ctx in cx.ancestors(followfirst=followfirst))
831 831 # include the revision responsible for the most recent version
832 832 s.add(cx.linkrev())
833 833 else:
834 834 return baseset([])
835 835 else:
836 836 s = _revancestors(repo, baseset([c.rev()]), followfirst)
837 837
838 838 return subset.filter(s.__contains__)
839 839
840 840 def follow(repo, subset, x):
841 841 """``follow([file])``
842 842 An alias for ``::.`` (ancestors of the working copy's first parent).
843 843 If a filename is specified, the history of the given file is followed,
844 844 including copies.
845 845 """
846 846 return _follow(repo, subset, x, 'follow')
847 847
848 848 def _followfirst(repo, subset, x):
849 849 # ``followfirst([file])``
850 850 # Like ``follow([file])`` but follows only the first parent of
851 851 # every revision or file revision.
852 852 return _follow(repo, subset, x, '_followfirst', followfirst=True)
853 853
854 854 def getall(repo, subset, x):
855 855 """``all()``
856 856 All changesets, the same as ``0:tip``.
857 857 """
858 858 # i18n: "all" is a keyword
859 859 getargs(x, 0, 0, _("all takes no arguments"))
860 860 return subset
861 861
862 862 def grep(repo, subset, x):
863 863 """``grep(regex)``
864 864 Like ``keyword(string)`` but accepts a regex. Use ``grep(r'...')``
865 865 to ensure special escape characters are handled correctly. Unlike
866 866 ``keyword(string)``, the match is case-sensitive.
867 867 """
868 868 try:
869 869 # i18n: "grep" is a keyword
870 870 gr = re.compile(getstring(x, _("grep requires a string")))
871 871 except re.error, e:
872 872 raise error.ParseError(_('invalid match pattern: %s') % e)
873 873
874 874 def matches(x):
875 875 c = repo[x]
876 876 for e in c.files() + [c.user(), c.description()]:
877 877 if gr.search(e):
878 878 return True
879 879 return False
880 880
881 881 return subset.filter(matches)
882 882
883 883 def _matchfiles(repo, subset, x):
884 884 # _matchfiles takes a revset list of prefixed arguments:
885 885 #
886 886 # [p:foo, i:bar, x:baz]
887 887 #
888 888 # builds a match object from them and filters subset. Allowed
889 889 # prefixes are 'p:' for regular patterns, 'i:' for include
890 890 # patterns and 'x:' for exclude patterns. Use 'r:' prefix to pass
891 891 # a revision identifier, or the empty string to reference the
892 892 # working directory, from which the match object is
893 893 # initialized. Use 'd:' to set the default matching mode, default
894 894 # to 'glob'. At most one 'r:' and 'd:' argument can be passed.
895 895
896 896 # i18n: "_matchfiles" is a keyword
897 897 l = getargs(x, 1, -1, _("_matchfiles requires at least one argument"))
898 898 pats, inc, exc = [], [], []
899 899 hasset = False
900 900 rev, default = None, None
901 901 for arg in l:
902 902 # i18n: "_matchfiles" is a keyword
903 903 s = getstring(arg, _("_matchfiles requires string arguments"))
904 904 prefix, value = s[:2], s[2:]
905 905 if prefix == 'p:':
906 906 pats.append(value)
907 907 elif prefix == 'i:':
908 908 inc.append(value)
909 909 elif prefix == 'x:':
910 910 exc.append(value)
911 911 elif prefix == 'r:':
912 912 if rev is not None:
913 913 # i18n: "_matchfiles" is a keyword
914 914 raise error.ParseError(_('_matchfiles expected at most one '
915 915 'revision'))
916 916 rev = value
917 917 elif prefix == 'd:':
918 918 if default is not None:
919 919 # i18n: "_matchfiles" is a keyword
920 920 raise error.ParseError(_('_matchfiles expected at most one '
921 921 'default mode'))
922 922 default = value
923 923 else:
924 924 # i18n: "_matchfiles" is a keyword
925 925 raise error.ParseError(_('invalid _matchfiles prefix: %s') % prefix)
926 926 if not hasset and matchmod.patkind(value) == 'set':
927 927 hasset = True
928 928 if not default:
929 929 default = 'glob'
930 930
931 931 def matches(x):
932 932 m = None
933 933 c = repo[x]
934 934 if not m or (hasset and rev is None):
935 935 ctx = c
936 936 if rev is not None:
937 937 ctx = repo[rev or None]
938 938 m = matchmod.match(repo.root, repo.getcwd(), pats, include=inc,
939 939 exclude=exc, ctx=ctx, default=default)
940 940 for f in c.files():
941 941 if m(f):
942 942 return True
943 943 return False
944 944
945 945 return subset.filter(matches)
946 946
947 947 def hasfile(repo, subset, x):
948 948 """``file(pattern)``
949 949 Changesets affecting files matched by pattern.
950 950
951 951 For a faster but less accurate result, consider using ``filelog()``
952 952 instead.
953 953
954 954 This predicate uses ``glob:`` as the default kind of pattern.
955 955 """
956 956 # i18n: "file" is a keyword
957 957 pat = getstring(x, _("file requires a pattern"))
958 958 return _matchfiles(repo, subset, ('string', 'p:' + pat))
959 959
960 960 def head(repo, subset, x):
961 961 """``head()``
962 962 Changeset is a named branch head.
963 963 """
964 964 # i18n: "head" is a keyword
965 965 getargs(x, 0, 0, _("head takes no arguments"))
966 966 hs = set()
967 967 for b, ls in repo.branchmap().iteritems():
968 968 hs.update(repo[h].rev() for h in ls)
969 969 return baseset(hs).filter(subset.__contains__)
970 970
971 971 def heads(repo, subset, x):
972 972 """``heads(set)``
973 973 Members of set with no children in set.
974 974 """
975 975 s = getset(repo, subset, x)
976 976 ps = parents(repo, subset, x)
977 977 return s - ps
978 978
979 979 def hidden(repo, subset, x):
980 980 """``hidden()``
981 981 Hidden changesets.
982 982 """
983 983 # i18n: "hidden" is a keyword
984 984 getargs(x, 0, 0, _("hidden takes no arguments"))
985 985 hiddenrevs = repoview.filterrevs(repo, 'visible')
986 986 return subset & hiddenrevs
987 987
988 988 def keyword(repo, subset, x):
989 989 """``keyword(string)``
990 990 Search commit message, user name, and names of changed files for
991 991 string. The match is case-insensitive.
992 992 """
993 993 # i18n: "keyword" is a keyword
994 994 kw = encoding.lower(getstring(x, _("keyword requires a string")))
995 995
996 996 def matches(r):
997 997 c = repo[r]
998 998 return util.any(kw in encoding.lower(t) for t in c.files() + [c.user(),
999 999 c.description()])
1000 1000
1001 1001 return subset.filter(matches)
1002 1002
1003 1003 def limit(repo, subset, x):
1004 1004 """``limit(set, [n])``
1005 1005 First n members of set, defaulting to 1.
1006 1006 """
1007 1007 # i18n: "limit" is a keyword
1008 1008 l = getargs(x, 1, 2, _("limit requires one or two arguments"))
1009 1009 try:
1010 1010 lim = 1
1011 1011 if len(l) == 2:
1012 1012 # i18n: "limit" is a keyword
1013 1013 lim = int(getstring(l[1], _("limit requires a number")))
1014 1014 except (TypeError, ValueError):
1015 1015 # i18n: "limit" is a keyword
1016 1016 raise error.ParseError(_("limit expects a number"))
1017 1017 ss = subset.set()
1018 1018 os = getset(repo, spanset(repo), l[0])
1019 1019 bs = baseset([])
1020 1020 it = iter(os)
1021 1021 for x in xrange(lim):
1022 1022 try:
1023 1023 y = it.next()
1024 1024 if y in ss:
1025 1025 bs.append(y)
1026 1026 except (StopIteration):
1027 1027 break
1028 1028 return bs
1029 1029
1030 1030 def last(repo, subset, x):
1031 1031 """``last(set, [n])``
1032 1032 Last n members of set, defaulting to 1.
1033 1033 """
1034 1034 # i18n: "last" is a keyword
1035 1035 l = getargs(x, 1, 2, _("last requires one or two arguments"))
1036 1036 try:
1037 1037 lim = 1
1038 1038 if len(l) == 2:
1039 1039 # i18n: "last" is a keyword
1040 1040 lim = int(getstring(l[1], _("last requires a number")))
1041 1041 except (TypeError, ValueError):
1042 1042 # i18n: "last" is a keyword
1043 1043 raise error.ParseError(_("last expects a number"))
1044 1044 ss = subset.set()
1045 1045 os = getset(repo, spanset(repo), l[0])
1046 1046 os.reverse()
1047 1047 bs = baseset([])
1048 1048 it = iter(os)
1049 1049 for x in xrange(lim):
1050 1050 try:
1051 1051 y = it.next()
1052 1052 if y in ss:
1053 1053 bs.append(y)
1054 1054 except (StopIteration):
1055 1055 break
1056 1056 return bs
1057 1057
1058 1058 def maxrev(repo, subset, x):
1059 1059 """``max(set)``
1060 1060 Changeset with highest revision number in set.
1061 1061 """
1062 1062 os = getset(repo, spanset(repo), x)
1063 1063 if os:
1064 1064 m = os.max()
1065 1065 if m in subset:
1066 1066 return baseset([m])
1067 1067 return baseset([])
1068 1068
1069 1069 def merge(repo, subset, x):
1070 1070 """``merge()``
1071 1071 Changeset is a merge changeset.
1072 1072 """
1073 1073 # i18n: "merge" is a keyword
1074 1074 getargs(x, 0, 0, _("merge takes no arguments"))
1075 1075 cl = repo.changelog
1076 1076 return subset.filter(lambda r: cl.parentrevs(r)[1] != -1)
1077 1077
1078 1078 def branchpoint(repo, subset, x):
1079 1079 """``branchpoint()``
1080 1080 Changesets with more than one child.
1081 1081 """
1082 1082 # i18n: "branchpoint" is a keyword
1083 1083 getargs(x, 0, 0, _("branchpoint takes no arguments"))
1084 1084 cl = repo.changelog
1085 1085 if not subset:
1086 1086 return baseset([])
1087 1087 baserev = min(subset)
1088 1088 parentscount = [0]*(len(repo) - baserev)
1089 1089 for r in cl.revs(start=baserev + 1):
1090 1090 for p in cl.parentrevs(r):
1091 1091 if p >= baserev:
1092 1092 parentscount[p - baserev] += 1
1093 1093 return subset.filter(lambda r: parentscount[r - baserev] > 1)
1094 1094
1095 1095 def minrev(repo, subset, x):
1096 1096 """``min(set)``
1097 1097 Changeset with lowest revision number in set.
1098 1098 """
1099 1099 os = getset(repo, spanset(repo), x)
1100 1100 if os:
1101 1101 m = os.min()
1102 1102 if m in subset:
1103 1103 return baseset([m])
1104 1104 return baseset([])
1105 1105
1106 1106 def modifies(repo, subset, x):
1107 1107 """``modifies(pattern)``
1108 1108 Changesets modifying files matched by pattern.
1109 1109
1110 1110 The pattern without explicit kind like ``glob:`` is expected to be
1111 1111 relative to the current directory and match against a file or a
1112 1112 directory.
1113 1113 """
1114 1114 # i18n: "modifies" is a keyword
1115 1115 pat = getstring(x, _("modifies requires a pattern"))
1116 1116 return checkstatus(repo, subset, pat, 0)
1117 1117
1118 1118 def node_(repo, subset, x):
1119 1119 """``id(string)``
1120 1120 Revision non-ambiguously specified by the given hex string prefix.
1121 1121 """
1122 1122 # i18n: "id" is a keyword
1123 1123 l = getargs(x, 1, 1, _("id requires one argument"))
1124 1124 # i18n: "id" is a keyword
1125 1125 n = getstring(l[0], _("id requires a string"))
1126 1126 if len(n) == 40:
1127 1127 rn = repo[n].rev()
1128 1128 else:
1129 1129 rn = None
1130 1130 pm = repo.changelog._partialmatch(n)
1131 1131 if pm is not None:
1132 1132 rn = repo.changelog.rev(pm)
1133 1133
1134 1134 return subset.filter(lambda r: r == rn)
1135 1135
1136 1136 def obsolete(repo, subset, x):
1137 1137 """``obsolete()``
1138 1138 Mutable changeset with a newer version."""
1139 1139 # i18n: "obsolete" is a keyword
1140 1140 getargs(x, 0, 0, _("obsolete takes no arguments"))
1141 1141 obsoletes = obsmod.getrevs(repo, 'obsolete')
1142 1142 return subset & obsoletes
1143 1143
1144 1144 def origin(repo, subset, x):
1145 1145 """``origin([set])``
1146 1146 Changesets that were specified as a source for the grafts, transplants or
1147 1147 rebases that created the given revisions. Omitting the optional set is the
1148 1148 same as passing all(). If a changeset created by these operations is itself
1149 1149 specified as a source for one of these operations, only the source changeset
1150 1150 for the first operation is selected.
1151 1151 """
1152 1152 if x is not None:
1153 1153 args = getset(repo, spanset(repo), x).set()
1154 1154 else:
1155 1155 args = getall(repo, spanset(repo), x).set()
1156 1156
1157 1157 def _firstsrc(rev):
1158 1158 src = _getrevsource(repo, rev)
1159 1159 if src is None:
1160 1160 return None
1161 1161
1162 1162 while True:
1163 1163 prev = _getrevsource(repo, src)
1164 1164
1165 1165 if prev is None:
1166 1166 return src
1167 1167 src = prev
1168 1168
1169 1169 o = set([_firstsrc(r) for r in args])
1170 1170 return subset.filter(o.__contains__)
1171 1171
1172 1172 def outgoing(repo, subset, x):
1173 1173 """``outgoing([path])``
1174 1174 Changesets not found in the specified destination repository, or the
1175 1175 default push location.
1176 1176 """
1177 1177 import hg # avoid start-up nasties
1178 1178 # i18n: "outgoing" is a keyword
1179 1179 l = getargs(x, 0, 1, _("outgoing takes one or no arguments"))
1180 1180 # i18n: "outgoing" is a keyword
1181 1181 dest = l and getstring(l[0], _("outgoing requires a repository path")) or ''
1182 1182 dest = repo.ui.expandpath(dest or 'default-push', dest or 'default')
1183 1183 dest, branches = hg.parseurl(dest)
1184 1184 revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
1185 1185 if revs:
1186 1186 revs = [repo.lookup(rev) for rev in revs]
1187 1187 other = hg.peer(repo, {}, dest)
1188 1188 repo.ui.pushbuffer()
1189 1189 outgoing = discovery.findcommonoutgoing(repo, other, onlyheads=revs)
1190 1190 repo.ui.popbuffer()
1191 1191 cl = repo.changelog
1192 1192 o = set([cl.rev(r) for r in outgoing.missing])
1193 1193 return subset.filter(o.__contains__)
1194 1194
1195 1195 def p1(repo, subset, x):
1196 1196 """``p1([set])``
1197 1197 First parent of changesets in set, or the working directory.
1198 1198 """
1199 1199 if x is None:
1200 1200 p = repo[x].p1().rev()
1201 1201 return subset.filter(lambda r: r == p)
1202 1202
1203 1203 ps = set()
1204 1204 cl = repo.changelog
1205 1205 for r in getset(repo, spanset(repo), x):
1206 1206 ps.add(cl.parentrevs(r)[0])
1207 1207 ps -= set([node.nullrev])
1208 1208 return subset & ps
1209 1209
1210 1210 def p2(repo, subset, x):
1211 1211 """``p2([set])``
1212 1212 Second parent of changesets in set, or the working directory.
1213 1213 """
1214 1214 if x is None:
1215 1215 ps = repo[x].parents()
1216 1216 try:
1217 1217 p = ps[1].rev()
1218 1218 return subset.filter(lambda r: r == p)
1219 1219 except IndexError:
1220 1220 return baseset([])
1221 1221
1222 1222 ps = set()
1223 1223 cl = repo.changelog
1224 1224 for r in getset(repo, spanset(repo), x):
1225 1225 ps.add(cl.parentrevs(r)[1])
1226 1226 ps -= set([node.nullrev])
1227 1227 return subset & ps
1228 1228
1229 1229 def parents(repo, subset, x):
1230 1230 """``parents([set])``
1231 1231 The set of all parents for all changesets in set, or the working directory.
1232 1232 """
1233 1233 if x is None:
1234 ps = tuple(p.rev() for p in repo[x].parents())
1235 return subset & ps
1236
1237 ps = set()
1238 cl = repo.changelog
1239 for r in getset(repo, spanset(repo), x):
1240 ps.update(cl.parentrevs(r))
1234 ps = set(p.rev() for p in repo[x].parents())
1235 else:
1236 ps = set()
1237 cl = repo.changelog
1238 for r in getset(repo, spanset(repo), x):
1239 ps.update(cl.parentrevs(r))
1241 1240 return baseset(ps) & subset
1242 1241
1243 1242 def parentspec(repo, subset, x, n):
1244 1243 """``set^0``
1245 1244 The set.
1246 1245 ``set^1`` (or ``set^``), ``set^2``
1247 1246 First or second parent, respectively, of all changesets in set.
1248 1247 """
1249 1248 try:
1250 1249 n = int(n[1])
1251 1250 if n not in (0, 1, 2):
1252 1251 raise ValueError
1253 1252 except (TypeError, ValueError):
1254 1253 raise error.ParseError(_("^ expects a number 0, 1, or 2"))
1255 1254 ps = set()
1256 1255 cl = repo.changelog
1257 1256 for r in getset(repo, baseset(cl), x):
1258 1257 if n == 0:
1259 1258 ps.add(r)
1260 1259 elif n == 1:
1261 1260 ps.add(cl.parentrevs(r)[0])
1262 1261 elif n == 2:
1263 1262 parents = cl.parentrevs(r)
1264 1263 if len(parents) > 1:
1265 1264 ps.add(parents[1])
1266 1265 return subset & ps
1267 1266
1268 1267 def present(repo, subset, x):
1269 1268 """``present(set)``
1270 1269 An empty set, if any revision in set isn't found; otherwise,
1271 1270 all revisions in set.
1272 1271
1273 1272 If any of specified revisions is not present in the local repository,
1274 1273 the query is normally aborted. But this predicate allows the query
1275 1274 to continue even in such cases.
1276 1275 """
1277 1276 try:
1278 1277 return getset(repo, subset, x)
1279 1278 except error.RepoLookupError:
1280 1279 return baseset([])
1281 1280
1282 1281 def public(repo, subset, x):
1283 1282 """``public()``
1284 1283 Changeset in public phase."""
1285 1284 # i18n: "public" is a keyword
1286 1285 getargs(x, 0, 0, _("public takes no arguments"))
1287 1286 pc = repo._phasecache
1288 1287 return subset.filter(lambda r: pc.phase(repo, r) == phases.public)
1289 1288
1290 1289 def remote(repo, subset, x):
1291 1290 """``remote([id [,path]])``
1292 1291 Local revision that corresponds to the given identifier in a
1293 1292 remote repository, if present. Here, the '.' identifier is a
1294 1293 synonym for the current local branch.
1295 1294 """
1296 1295
1297 1296 import hg # avoid start-up nasties
1298 1297 # i18n: "remote" is a keyword
1299 1298 l = getargs(x, 0, 2, _("remote takes one, two or no arguments"))
1300 1299
1301 1300 q = '.'
1302 1301 if len(l) > 0:
1303 1302 # i18n: "remote" is a keyword
1304 1303 q = getstring(l[0], _("remote requires a string id"))
1305 1304 if q == '.':
1306 1305 q = repo['.'].branch()
1307 1306
1308 1307 dest = ''
1309 1308 if len(l) > 1:
1310 1309 # i18n: "remote" is a keyword
1311 1310 dest = getstring(l[1], _("remote requires a repository path"))
1312 1311 dest = repo.ui.expandpath(dest or 'default')
1313 1312 dest, branches = hg.parseurl(dest)
1314 1313 revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
1315 1314 if revs:
1316 1315 revs = [repo.lookup(rev) for rev in revs]
1317 1316 other = hg.peer(repo, {}, dest)
1318 1317 n = other.lookup(q)
1319 1318 if n in repo:
1320 1319 r = repo[n].rev()
1321 1320 if r in subset:
1322 1321 return baseset([r])
1323 1322 return baseset([])
1324 1323
1325 1324 def removes(repo, subset, x):
1326 1325 """``removes(pattern)``
1327 1326 Changesets which remove files matching pattern.
1328 1327
1329 1328 The pattern without explicit kind like ``glob:`` is expected to be
1330 1329 relative to the current directory and match against a file or a
1331 1330 directory.
1332 1331 """
1333 1332 # i18n: "removes" is a keyword
1334 1333 pat = getstring(x, _("removes requires a pattern"))
1335 1334 return checkstatus(repo, subset, pat, 2)
1336 1335
1337 1336 def rev(repo, subset, x):
1338 1337 """``rev(number)``
1339 1338 Revision with the given numeric identifier.
1340 1339 """
1341 1340 # i18n: "rev" is a keyword
1342 1341 l = getargs(x, 1, 1, _("rev requires one argument"))
1343 1342 try:
1344 1343 # i18n: "rev" is a keyword
1345 1344 l = int(getstring(l[0], _("rev requires a number")))
1346 1345 except (TypeError, ValueError):
1347 1346 # i18n: "rev" is a keyword
1348 1347 raise error.ParseError(_("rev expects a number"))
1349 1348 return subset.filter(lambda r: r == l)
1350 1349
1351 1350 def matching(repo, subset, x):
1352 1351 """``matching(revision [, field])``
1353 1352 Changesets in which a given set of fields match the set of fields in the
1354 1353 selected revision or set.
1355 1354
1356 1355 To match more than one field pass the list of fields to match separated
1357 1356 by spaces (e.g. ``author description``).
1358 1357
1359 1358 Valid fields are most regular revision fields and some special fields.
1360 1359
1361 1360 Regular revision fields are ``description``, ``author``, ``branch``,
1362 1361 ``date``, ``files``, ``phase``, ``parents``, ``substate``, ``user``
1363 1362 and ``diff``.
1364 1363 Note that ``author`` and ``user`` are synonyms. ``diff`` refers to the
1365 1364 contents of the revision. Two revisions matching their ``diff`` will
1366 1365 also match their ``files``.
1367 1366
1368 1367 Special fields are ``summary`` and ``metadata``:
1369 1368 ``summary`` matches the first line of the description.
1370 1369 ``metadata`` is equivalent to matching ``description user date``
1371 1370 (i.e. it matches the main metadata fields).
1372 1371
1373 1372 ``metadata`` is the default field which is used when no fields are
1374 1373 specified. You can match more than one field at a time.
1375 1374 """
1376 1375 # i18n: "matching" is a keyword
1377 1376 l = getargs(x, 1, 2, _("matching takes 1 or 2 arguments"))
1378 1377
1379 1378 revs = getset(repo, baseset(repo.changelog), l[0])
1380 1379
1381 1380 fieldlist = ['metadata']
1382 1381 if len(l) > 1:
1383 1382 fieldlist = getstring(l[1],
1384 1383 # i18n: "matching" is a keyword
1385 1384 _("matching requires a string "
1386 1385 "as its second argument")).split()
1387 1386
1388 1387 # Make sure that there are no repeated fields,
1389 1388 # expand the 'special' 'metadata' field type
1390 1389 # and check the 'files' whenever we check the 'diff'
1391 1390 fields = []
1392 1391 for field in fieldlist:
1393 1392 if field == 'metadata':
1394 1393 fields += ['user', 'description', 'date']
1395 1394 elif field == 'diff':
1396 1395 # a revision matching the diff must also match the files
1397 1396 # since matching the diff is very costly, make sure to
1398 1397 # also match the files first
1399 1398 fields += ['files', 'diff']
1400 1399 else:
1401 1400 if field == 'author':
1402 1401 field = 'user'
1403 1402 fields.append(field)
1404 1403 fields = set(fields)
1405 1404 if 'summary' in fields and 'description' in fields:
1406 1405 # If a revision matches its description it also matches its summary
1407 1406 fields.discard('summary')
1408 1407
1409 1408 # We may want to match more than one field
1410 1409 # Not all fields take the same amount of time to be matched
1411 1410 # Sort the selected fields in order of increasing matching cost
1412 1411 fieldorder = ['phase', 'parents', 'user', 'date', 'branch', 'summary',
1413 1412 'files', 'description', 'substate', 'diff']
1414 1413 def fieldkeyfunc(f):
1415 1414 try:
1416 1415 return fieldorder.index(f)
1417 1416 except ValueError:
1418 1417 # assume an unknown field is very costly
1419 1418 return len(fieldorder)
1420 1419 fields = list(fields)
1421 1420 fields.sort(key=fieldkeyfunc)
1422 1421
1423 1422 # Each field will be matched with its own "getfield" function
1424 1423 # which will be added to the getfieldfuncs array of functions
1425 1424 getfieldfuncs = []
1426 1425 _funcs = {
1427 1426 'user': lambda r: repo[r].user(),
1428 1427 'branch': lambda r: repo[r].branch(),
1429 1428 'date': lambda r: repo[r].date(),
1430 1429 'description': lambda r: repo[r].description(),
1431 1430 'files': lambda r: repo[r].files(),
1432 1431 'parents': lambda r: repo[r].parents(),
1433 1432 'phase': lambda r: repo[r].phase(),
1434 1433 'substate': lambda r: repo[r].substate,
1435 1434 'summary': lambda r: repo[r].description().splitlines()[0],
1436 1435 'diff': lambda r: list(repo[r].diff(git=True),)
1437 1436 }
1438 1437 for info in fields:
1439 1438 getfield = _funcs.get(info, None)
1440 1439 if getfield is None:
1441 1440 raise error.ParseError(
1442 1441 # i18n: "matching" is a keyword
1443 1442 _("unexpected field name passed to matching: %s") % info)
1444 1443 getfieldfuncs.append(getfield)
1445 1444 # convert the getfield array of functions into a "getinfo" function
1446 1445 # which returns an array of field values (or a single value if there
1447 1446 # is only one field to match)
1448 1447 getinfo = lambda r: [f(r) for f in getfieldfuncs]
1449 1448
1450 1449 def matches(x):
1451 1450 for rev in revs:
1452 1451 target = getinfo(rev)
1453 1452 match = True
1454 1453 for n, f in enumerate(getfieldfuncs):
1455 1454 if target[n] != f(x):
1456 1455 match = False
1457 1456 if match:
1458 1457 return True
1459 1458 return False
1460 1459
1461 1460 return subset.filter(matches)
1462 1461
1463 1462 def reverse(repo, subset, x):
1464 1463 """``reverse(set)``
1465 1464 Reverse order of set.
1466 1465 """
1467 1466 l = getset(repo, subset, x)
1468 1467 l.reverse()
1469 1468 return l
1470 1469
1471 1470 def roots(repo, subset, x):
1472 1471 """``roots(set)``
1473 1472 Changesets in set with no parent changeset in set.
1474 1473 """
1475 1474 s = getset(repo, spanset(repo), x).set()
1476 1475 subset = baseset([r for r in s if r in subset.set()])
1477 1476 cs = _children(repo, subset, s)
1478 1477 return subset - cs
1479 1478
1480 1479 def secret(repo, subset, x):
1481 1480 """``secret()``
1482 1481 Changeset in secret phase."""
1483 1482 # i18n: "secret" is a keyword
1484 1483 getargs(x, 0, 0, _("secret takes no arguments"))
1485 1484 pc = repo._phasecache
1486 1485 return subset.filter(lambda x: pc.phase(repo, x) == phases.secret)
1487 1486
1488 1487 def sort(repo, subset, x):
1489 1488 """``sort(set[, [-]key...])``
1490 1489 Sort set by keys. The default sort order is ascending, specify a key
1491 1490 as ``-key`` to sort in descending order.
1492 1491
1493 1492 The keys can be:
1494 1493
1495 1494 - ``rev`` for the revision number,
1496 1495 - ``branch`` for the branch name,
1497 1496 - ``desc`` for the commit message (description),
1498 1497 - ``user`` for user name (``author`` can be used as an alias),
1499 1498 - ``date`` for the commit date
1500 1499 """
1501 1500 # i18n: "sort" is a keyword
1502 1501 l = getargs(x, 1, 2, _("sort requires one or two arguments"))
1503 1502 keys = "rev"
1504 1503 if len(l) == 2:
1505 1504 # i18n: "sort" is a keyword
1506 1505 keys = getstring(l[1], _("sort spec must be a string"))
1507 1506
1508 1507 s = l[0]
1509 1508 keys = keys.split()
1510 1509 l = []
1511 1510 def invert(s):
1512 1511 return "".join(chr(255 - ord(c)) for c in s)
1513 1512 revs = getset(repo, subset, s)
1514 1513 if keys == ["rev"]:
1515 1514 revs.sort()
1516 1515 return revs
1517 1516 elif keys == ["-rev"]:
1518 1517 revs.sort(reverse=True)
1519 1518 return revs
1520 1519 for r in revs:
1521 1520 c = repo[r]
1522 1521 e = []
1523 1522 for k in keys:
1524 1523 if k == 'rev':
1525 1524 e.append(r)
1526 1525 elif k == '-rev':
1527 1526 e.append(-r)
1528 1527 elif k == 'branch':
1529 1528 e.append(c.branch())
1530 1529 elif k == '-branch':
1531 1530 e.append(invert(c.branch()))
1532 1531 elif k == 'desc':
1533 1532 e.append(c.description())
1534 1533 elif k == '-desc':
1535 1534 e.append(invert(c.description()))
1536 1535 elif k in 'user author':
1537 1536 e.append(c.user())
1538 1537 elif k in '-user -author':
1539 1538 e.append(invert(c.user()))
1540 1539 elif k == 'date':
1541 1540 e.append(c.date()[0])
1542 1541 elif k == '-date':
1543 1542 e.append(-c.date()[0])
1544 1543 else:
1545 1544 raise error.ParseError(_("unknown sort key %r") % k)
1546 1545 e.append(r)
1547 1546 l.append(e)
1548 1547 l.sort()
1549 1548 return baseset([e[-1] for e in l])
1550 1549
1551 1550 def _stringmatcher(pattern):
1552 1551 """
1553 1552 accepts a string, possibly starting with 're:' or 'literal:' prefix.
1554 1553 returns the matcher name, pattern, and matcher function.
1555 1554 missing or unknown prefixes are treated as literal matches.
1556 1555
1557 1556 helper for tests:
1558 1557 >>> def test(pattern, *tests):
1559 1558 ... kind, pattern, matcher = _stringmatcher(pattern)
1560 1559 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
1561 1560
1562 1561 exact matching (no prefix):
1563 1562 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
1564 1563 ('literal', 'abcdefg', [False, False, True])
1565 1564
1566 1565 regex matching ('re:' prefix)
1567 1566 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
1568 1567 ('re', 'a.+b', [False, False, True])
1569 1568
1570 1569 force exact matches ('literal:' prefix)
1571 1570 >>> test('literal:re:foobar', 'foobar', 're:foobar')
1572 1571 ('literal', 're:foobar', [False, True])
1573 1572
1574 1573 unknown prefixes are ignored and treated as literals
1575 1574 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
1576 1575 ('literal', 'foo:bar', [False, False, True])
1577 1576 """
1578 1577 if pattern.startswith('re:'):
1579 1578 pattern = pattern[3:]
1580 1579 try:
1581 1580 regex = re.compile(pattern)
1582 1581 except re.error, e:
1583 1582 raise error.ParseError(_('invalid regular expression: %s')
1584 1583 % e)
1585 1584 return 're', pattern, regex.search
1586 1585 elif pattern.startswith('literal:'):
1587 1586 pattern = pattern[8:]
1588 1587 return 'literal', pattern, pattern.__eq__
1589 1588
1590 1589 def _substringmatcher(pattern):
1591 1590 kind, pattern, matcher = _stringmatcher(pattern)
1592 1591 if kind == 'literal':
1593 1592 matcher = lambda s: pattern in s
1594 1593 return kind, pattern, matcher
1595 1594
1596 1595 def tag(repo, subset, x):
1597 1596 """``tag([name])``
1598 1597 The specified tag by name, or all tagged revisions if no name is given.
1599 1598
1600 1599 If `name` starts with `re:`, the remainder of the name is treated as
1601 1600 a regular expression. To match a tag that actually starts with `re:`,
1602 1601 use the prefix `literal:`.
1603 1602 """
1604 1603 # i18n: "tag" is a keyword
1605 1604 args = getargs(x, 0, 1, _("tag takes one or no arguments"))
1606 1605 cl = repo.changelog
1607 1606 if args:
1608 1607 pattern = getstring(args[0],
1609 1608 # i18n: "tag" is a keyword
1610 1609 _('the argument to tag must be a string'))
1611 1610 kind, pattern, matcher = _stringmatcher(pattern)
1612 1611 if kind == 'literal':
1613 1612 # avoid resolving all tags
1614 1613 tn = repo._tagscache.tags.get(pattern, None)
1615 1614 if tn is None:
1616 1615 raise util.Abort(_("tag '%s' does not exist") % pattern)
1617 1616 s = set([repo[tn].rev()])
1618 1617 else:
1619 1618 s = set([cl.rev(n) for t, n in repo.tagslist() if matcher(t)])
1620 1619 else:
1621 1620 s = set([cl.rev(n) for t, n in repo.tagslist() if t != 'tip'])
1622 1621 return subset & s
1623 1622
1624 1623 def tagged(repo, subset, x):
1625 1624 return tag(repo, subset, x)
1626 1625
1627 1626 def unstable(repo, subset, x):
1628 1627 """``unstable()``
1629 1628 Non-obsolete changesets with obsolete ancestors.
1630 1629 """
1631 1630 # i18n: "unstable" is a keyword
1632 1631 getargs(x, 0, 0, _("unstable takes no arguments"))
1633 1632 unstables = obsmod.getrevs(repo, 'unstable')
1634 1633 return subset & unstables
1635 1634
1636 1635
1637 1636 def user(repo, subset, x):
1638 1637 """``user(string)``
1639 1638 User name contains string. The match is case-insensitive.
1640 1639
1641 1640 If `string` starts with `re:`, the remainder of the string is treated as
1642 1641 a regular expression. To match a user that actually contains `re:`, use
1643 1642 the prefix `literal:`.
1644 1643 """
1645 1644 return author(repo, subset, x)
1646 1645
1647 1646 # for internal use
1648 1647 def _list(repo, subset, x):
1649 1648 s = getstring(x, "internal error")
1650 1649 if not s:
1651 1650 return baseset([])
1652 1651 ls = [repo[r].rev() for r in s.split('\0')]
1653 1652 s = subset.set()
1654 1653 return baseset([r for r in ls if r in s])
1655 1654
1656 1655 # for internal use
1657 1656 def _intlist(repo, subset, x):
1658 1657 s = getstring(x, "internal error")
1659 1658 if not s:
1660 1659 return baseset([])
1661 1660 ls = [int(r) for r in s.split('\0')]
1662 1661 s = subset.set()
1663 1662 return baseset([r for r in ls if r in s])
1664 1663
1665 1664 # for internal use
1666 1665 def _hexlist(repo, subset, x):
1667 1666 s = getstring(x, "internal error")
1668 1667 if not s:
1669 1668 return baseset([])
1670 1669 cl = repo.changelog
1671 1670 ls = [cl.rev(node.bin(r)) for r in s.split('\0')]
1672 1671 s = subset.set()
1673 1672 return baseset([r for r in ls if r in s])
1674 1673
1675 1674 symbols = {
1676 1675 "adds": adds,
1677 1676 "all": getall,
1678 1677 "ancestor": ancestor,
1679 1678 "ancestors": ancestors,
1680 1679 "_firstancestors": _firstancestors,
1681 1680 "author": author,
1682 1681 "only": only,
1683 1682 "bisect": bisect,
1684 1683 "bisected": bisected,
1685 1684 "bookmark": bookmark,
1686 1685 "branch": branch,
1687 1686 "branchpoint": branchpoint,
1688 1687 "bumped": bumped,
1689 1688 "bundle": bundle,
1690 1689 "children": children,
1691 1690 "closed": closed,
1692 1691 "contains": contains,
1693 1692 "converted": converted,
1694 1693 "date": date,
1695 1694 "desc": desc,
1696 1695 "descendants": descendants,
1697 1696 "_firstdescendants": _firstdescendants,
1698 1697 "destination": destination,
1699 1698 "divergent": divergent,
1700 1699 "draft": draft,
1701 1700 "extinct": extinct,
1702 1701 "extra": extra,
1703 1702 "file": hasfile,
1704 1703 "filelog": filelog,
1705 1704 "first": first,
1706 1705 "follow": follow,
1707 1706 "_followfirst": _followfirst,
1708 1707 "grep": grep,
1709 1708 "head": head,
1710 1709 "heads": heads,
1711 1710 "hidden": hidden,
1712 1711 "id": node_,
1713 1712 "keyword": keyword,
1714 1713 "last": last,
1715 1714 "limit": limit,
1716 1715 "_matchfiles": _matchfiles,
1717 1716 "max": maxrev,
1718 1717 "merge": merge,
1719 1718 "min": minrev,
1720 1719 "modifies": modifies,
1721 1720 "obsolete": obsolete,
1722 1721 "origin": origin,
1723 1722 "outgoing": outgoing,
1724 1723 "p1": p1,
1725 1724 "p2": p2,
1726 1725 "parents": parents,
1727 1726 "present": present,
1728 1727 "public": public,
1729 1728 "remote": remote,
1730 1729 "removes": removes,
1731 1730 "rev": rev,
1732 1731 "reverse": reverse,
1733 1732 "roots": roots,
1734 1733 "sort": sort,
1735 1734 "secret": secret,
1736 1735 "matching": matching,
1737 1736 "tag": tag,
1738 1737 "tagged": tagged,
1739 1738 "user": user,
1740 1739 "unstable": unstable,
1741 1740 "_list": _list,
1742 1741 "_intlist": _intlist,
1743 1742 "_hexlist": _hexlist,
1744 1743 }
1745 1744
1746 1745 # symbols which can't be used for a DoS attack for any given input
1747 1746 # (e.g. those which accept regexes as plain strings shouldn't be included)
1748 1747 # functions that just return a lot of changesets (like all) don't count here
1749 1748 safesymbols = set([
1750 1749 "adds",
1751 1750 "all",
1752 1751 "ancestor",
1753 1752 "ancestors",
1754 1753 "_firstancestors",
1755 1754 "author",
1756 1755 "bisect",
1757 1756 "bisected",
1758 1757 "bookmark",
1759 1758 "branch",
1760 1759 "branchpoint",
1761 1760 "bumped",
1762 1761 "bundle",
1763 1762 "children",
1764 1763 "closed",
1765 1764 "converted",
1766 1765 "date",
1767 1766 "desc",
1768 1767 "descendants",
1769 1768 "_firstdescendants",
1770 1769 "destination",
1771 1770 "divergent",
1772 1771 "draft",
1773 1772 "extinct",
1774 1773 "extra",
1775 1774 "file",
1776 1775 "filelog",
1777 1776 "first",
1778 1777 "follow",
1779 1778 "_followfirst",
1780 1779 "head",
1781 1780 "heads",
1782 1781 "hidden",
1783 1782 "id",
1784 1783 "keyword",
1785 1784 "last",
1786 1785 "limit",
1787 1786 "_matchfiles",
1788 1787 "max",
1789 1788 "merge",
1790 1789 "min",
1791 1790 "modifies",
1792 1791 "obsolete",
1793 1792 "origin",
1794 1793 "outgoing",
1795 1794 "p1",
1796 1795 "p2",
1797 1796 "parents",
1798 1797 "present",
1799 1798 "public",
1800 1799 "remote",
1801 1800 "removes",
1802 1801 "rev",
1803 1802 "reverse",
1804 1803 "roots",
1805 1804 "sort",
1806 1805 "secret",
1807 1806 "matching",
1808 1807 "tag",
1809 1808 "tagged",
1810 1809 "user",
1811 1810 "unstable",
1812 1811 "_list",
1813 1812 "_intlist",
1814 1813 "_hexlist",
1815 1814 ])
1816 1815
1817 1816 methods = {
1818 1817 "range": rangeset,
1819 1818 "dagrange": dagrange,
1820 1819 "string": stringset,
1821 1820 "symbol": symbolset,
1822 1821 "and": andset,
1823 1822 "or": orset,
1824 1823 "not": notset,
1825 1824 "list": listset,
1826 1825 "func": func,
1827 1826 "ancestor": ancestorspec,
1828 1827 "parent": parentspec,
1829 1828 "parentpost": p1,
1830 1829 }
1831 1830
1832 1831 def optimize(x, small):
1833 1832 if x is None:
1834 1833 return 0, x
1835 1834
1836 1835 smallbonus = 1
1837 1836 if small:
1838 1837 smallbonus = .5
1839 1838
1840 1839 op = x[0]
1841 1840 if op == 'minus':
1842 1841 return optimize(('and', x[1], ('not', x[2])), small)
1843 1842 elif op == 'dagrangepre':
1844 1843 return optimize(('func', ('symbol', 'ancestors'), x[1]), small)
1845 1844 elif op == 'dagrangepost':
1846 1845 return optimize(('func', ('symbol', 'descendants'), x[1]), small)
1847 1846 elif op == 'rangepre':
1848 1847 return optimize(('range', ('string', '0'), x[1]), small)
1849 1848 elif op == 'rangepost':
1850 1849 return optimize(('range', x[1], ('string', 'tip')), small)
1851 1850 elif op == 'negate':
1852 1851 return optimize(('string',
1853 1852 '-' + getstring(x[1], _("can't negate that"))), small)
1854 1853 elif op in 'string symbol negate':
1855 1854 return smallbonus, x # single revisions are small
1856 1855 elif op == 'and':
1857 1856 wa, ta = optimize(x[1], True)
1858 1857 wb, tb = optimize(x[2], True)
1859 1858
1860 1859 # (::x and not ::y)/(not ::y and ::x) have a fast path
1861 1860 def isonly(revs, bases):
1862 1861 return (
1863 1862 revs[0] == 'func'
1864 1863 and getstring(revs[1], _('not a symbol')) == 'ancestors'
1865 1864 and bases[0] == 'not'
1866 1865 and bases[1][0] == 'func'
1867 1866 and getstring(bases[1][1], _('not a symbol')) == 'ancestors')
1868 1867
1869 1868 w = min(wa, wb)
1870 1869 if isonly(ta, tb):
1871 1870 return w, ('func', ('symbol', 'only'), ('list', ta[2], tb[1][2]))
1872 1871 if isonly(tb, ta):
1873 1872 return w, ('func', ('symbol', 'only'), ('list', tb[2], ta[1][2]))
1874 1873
1875 1874 if wa > wb:
1876 1875 return w, (op, tb, ta)
1877 1876 return w, (op, ta, tb)
1878 1877 elif op == 'or':
1879 1878 wa, ta = optimize(x[1], False)
1880 1879 wb, tb = optimize(x[2], False)
1881 1880 if wb < wa:
1882 1881 wb, wa = wa, wb
1883 1882 return max(wa, wb), (op, ta, tb)
1884 1883 elif op == 'not':
1885 1884 o = optimize(x[1], not small)
1886 1885 return o[0], (op, o[1])
1887 1886 elif op == 'parentpost':
1888 1887 o = optimize(x[1], small)
1889 1888 return o[0], (op, o[1])
1890 1889 elif op == 'group':
1891 1890 return optimize(x[1], small)
1892 1891 elif op in 'dagrange range list parent ancestorspec':
1893 1892 if op == 'parent':
1894 1893 # x^:y means (x^) : y, not x ^ (:y)
1895 1894 post = ('parentpost', x[1])
1896 1895 if x[2][0] == 'dagrangepre':
1897 1896 return optimize(('dagrange', post, x[2][1]), small)
1898 1897 elif x[2][0] == 'rangepre':
1899 1898 return optimize(('range', post, x[2][1]), small)
1900 1899
1901 1900 wa, ta = optimize(x[1], small)
1902 1901 wb, tb = optimize(x[2], small)
1903 1902 return wa + wb, (op, ta, tb)
1904 1903 elif op == 'func':
1905 1904 f = getstring(x[1], _("not a symbol"))
1906 1905 wa, ta = optimize(x[2], small)
1907 1906 if f in ("author branch closed date desc file grep keyword "
1908 1907 "outgoing user"):
1909 1908 w = 10 # slow
1910 1909 elif f in "modifies adds removes":
1911 1910 w = 30 # slower
1912 1911 elif f == "contains":
1913 1912 w = 100 # very slow
1914 1913 elif f == "ancestor":
1915 1914 w = 1 * smallbonus
1916 1915 elif f in "reverse limit first _intlist":
1917 1916 w = 0
1918 1917 elif f in "sort":
1919 1918 w = 10 # assume most sorts look at changelog
1920 1919 else:
1921 1920 w = 1
1922 1921 return w + wa, (op, x[1], ta)
1923 1922 return 1, x
1924 1923
1925 1924 _aliasarg = ('func', ('symbol', '_aliasarg'))
1926 1925 def _getaliasarg(tree):
1927 1926 """If tree matches ('func', ('symbol', '_aliasarg'), ('string', X))
1928 1927 return X, None otherwise.
1929 1928 """
1930 1929 if (len(tree) == 3 and tree[:2] == _aliasarg
1931 1930 and tree[2][0] == 'string'):
1932 1931 return tree[2][1]
1933 1932 return None
1934 1933
1935 1934 def _checkaliasarg(tree, known=None):
1936 1935 """Check tree contains no _aliasarg construct or only ones which
1937 1936 value is in known. Used to avoid alias placeholders injection.
1938 1937 """
1939 1938 if isinstance(tree, tuple):
1940 1939 arg = _getaliasarg(tree)
1941 1940 if arg is not None and (not known or arg not in known):
1942 1941 raise error.ParseError(_("not a function: %s") % '_aliasarg')
1943 1942 for t in tree:
1944 1943 _checkaliasarg(t, known)
1945 1944
1946 1945 class revsetalias(object):
1947 1946 funcre = re.compile('^([^(]+)\(([^)]+)\)$')
1948 1947 args = None
1949 1948
1950 1949 def __init__(self, name, value):
1951 1950 '''Aliases like:
1952 1951
1953 1952 h = heads(default)
1954 1953 b($1) = ancestors($1) - ancestors(default)
1955 1954 '''
1956 1955 m = self.funcre.search(name)
1957 1956 if m:
1958 1957 self.name = m.group(1)
1959 1958 self.tree = ('func', ('symbol', m.group(1)))
1960 1959 self.args = [x.strip() for x in m.group(2).split(',')]
1961 1960 for arg in self.args:
1962 1961 # _aliasarg() is an unknown symbol only used separate
1963 1962 # alias argument placeholders from regular strings.
1964 1963 value = value.replace(arg, '_aliasarg(%r)' % (arg,))
1965 1964 else:
1966 1965 self.name = name
1967 1966 self.tree = ('symbol', name)
1968 1967
1969 1968 self.replacement, pos = parse(value)
1970 1969 if pos != len(value):
1971 1970 raise error.ParseError(_('invalid token'), pos)
1972 1971 # Check for placeholder injection
1973 1972 _checkaliasarg(self.replacement, self.args)
1974 1973
1975 1974 def _getalias(aliases, tree):
1976 1975 """If tree looks like an unexpanded alias, return it. Return None
1977 1976 otherwise.
1978 1977 """
1979 1978 if isinstance(tree, tuple) and tree:
1980 1979 if tree[0] == 'symbol' and len(tree) == 2:
1981 1980 name = tree[1]
1982 1981 alias = aliases.get(name)
1983 1982 if alias and alias.args is None and alias.tree == tree:
1984 1983 return alias
1985 1984 if tree[0] == 'func' and len(tree) > 1:
1986 1985 if tree[1][0] == 'symbol' and len(tree[1]) == 2:
1987 1986 name = tree[1][1]
1988 1987 alias = aliases.get(name)
1989 1988 if alias and alias.args is not None and alias.tree == tree[:2]:
1990 1989 return alias
1991 1990 return None
1992 1991
1993 1992 def _expandargs(tree, args):
1994 1993 """Replace _aliasarg instances with the substitution value of the
1995 1994 same name in args, recursively.
1996 1995 """
1997 1996 if not tree or not isinstance(tree, tuple):
1998 1997 return tree
1999 1998 arg = _getaliasarg(tree)
2000 1999 if arg is not None:
2001 2000 return args[arg]
2002 2001 return tuple(_expandargs(t, args) for t in tree)
2003 2002
2004 2003 def _expandaliases(aliases, tree, expanding, cache):
2005 2004 """Expand aliases in tree, recursively.
2006 2005
2007 2006 'aliases' is a dictionary mapping user defined aliases to
2008 2007 revsetalias objects.
2009 2008 """
2010 2009 if not isinstance(tree, tuple):
2011 2010 # Do not expand raw strings
2012 2011 return tree
2013 2012 alias = _getalias(aliases, tree)
2014 2013 if alias is not None:
2015 2014 if alias in expanding:
2016 2015 raise error.ParseError(_('infinite expansion of revset alias "%s" '
2017 2016 'detected') % alias.name)
2018 2017 expanding.append(alias)
2019 2018 if alias.name not in cache:
2020 2019 cache[alias.name] = _expandaliases(aliases, alias.replacement,
2021 2020 expanding, cache)
2022 2021 result = cache[alias.name]
2023 2022 expanding.pop()
2024 2023 if alias.args is not None:
2025 2024 l = getlist(tree[2])
2026 2025 if len(l) != len(alias.args):
2027 2026 raise error.ParseError(
2028 2027 _('invalid number of arguments: %s') % len(l))
2029 2028 l = [_expandaliases(aliases, a, [], cache) for a in l]
2030 2029 result = _expandargs(result, dict(zip(alias.args, l)))
2031 2030 else:
2032 2031 result = tuple(_expandaliases(aliases, t, expanding, cache)
2033 2032 for t in tree)
2034 2033 return result
2035 2034
2036 2035 def findaliases(ui, tree):
2037 2036 _checkaliasarg(tree)
2038 2037 aliases = {}
2039 2038 for k, v in ui.configitems('revsetalias'):
2040 2039 alias = revsetalias(k, v)
2041 2040 aliases[alias.name] = alias
2042 2041 return _expandaliases(aliases, tree, [], {})
2043 2042
2044 2043 def parse(spec, lookup=None):
2045 2044 p = parser.parser(tokenize, elements)
2046 2045 return p.parse(spec, lookup=lookup)
2047 2046
2048 2047 def match(ui, spec, repo=None):
2049 2048 if not spec:
2050 2049 raise error.ParseError(_("empty query"))
2051 2050 lookup = None
2052 2051 if repo:
2053 2052 lookup = repo.__contains__
2054 2053 tree, pos = parse(spec, lookup)
2055 2054 if (pos != len(spec)):
2056 2055 raise error.ParseError(_("invalid token"), pos)
2057 2056 if ui:
2058 2057 tree = findaliases(ui, tree)
2059 2058 weight, tree = optimize(tree, True)
2060 2059 def mfunc(repo, subset):
2061 2060 if util.safehasattr(subset, 'set'):
2062 2061 return getset(repo, subset, tree)
2063 2062 return getset(repo, baseset(subset), tree)
2064 2063 return mfunc
2065 2064
2066 2065 def formatspec(expr, *args):
2067 2066 '''
2068 2067 This is a convenience function for using revsets internally, and
2069 2068 escapes arguments appropriately. Aliases are intentionally ignored
2070 2069 so that intended expression behavior isn't accidentally subverted.
2071 2070
2072 2071 Supported arguments:
2073 2072
2074 2073 %r = revset expression, parenthesized
2075 2074 %d = int(arg), no quoting
2076 2075 %s = string(arg), escaped and single-quoted
2077 2076 %b = arg.branch(), escaped and single-quoted
2078 2077 %n = hex(arg), single-quoted
2079 2078 %% = a literal '%'
2080 2079
2081 2080 Prefixing the type with 'l' specifies a parenthesized list of that type.
2082 2081
2083 2082 >>> formatspec('%r:: and %lr', '10 or 11', ("this()", "that()"))
2084 2083 '(10 or 11):: and ((this()) or (that()))'
2085 2084 >>> formatspec('%d:: and not %d::', 10, 20)
2086 2085 '10:: and not 20::'
2087 2086 >>> formatspec('%ld or %ld', [], [1])
2088 2087 "_list('') or 1"
2089 2088 >>> formatspec('keyword(%s)', 'foo\\xe9')
2090 2089 "keyword('foo\\\\xe9')"
2091 2090 >>> b = lambda: 'default'
2092 2091 >>> b.branch = b
2093 2092 >>> formatspec('branch(%b)', b)
2094 2093 "branch('default')"
2095 2094 >>> formatspec('root(%ls)', ['a', 'b', 'c', 'd'])
2096 2095 "root(_list('a\\x00b\\x00c\\x00d'))"
2097 2096 '''
2098 2097
2099 2098 def quote(s):
2100 2099 return repr(str(s))
2101 2100
2102 2101 def argtype(c, arg):
2103 2102 if c == 'd':
2104 2103 return str(int(arg))
2105 2104 elif c == 's':
2106 2105 return quote(arg)
2107 2106 elif c == 'r':
2108 2107 parse(arg) # make sure syntax errors are confined
2109 2108 return '(%s)' % arg
2110 2109 elif c == 'n':
2111 2110 return quote(node.hex(arg))
2112 2111 elif c == 'b':
2113 2112 return quote(arg.branch())
2114 2113
2115 2114 def listexp(s, t):
2116 2115 l = len(s)
2117 2116 if l == 0:
2118 2117 return "_list('')"
2119 2118 elif l == 1:
2120 2119 return argtype(t, s[0])
2121 2120 elif t == 'd':
2122 2121 return "_intlist('%s')" % "\0".join(str(int(a)) for a in s)
2123 2122 elif t == 's':
2124 2123 return "_list('%s')" % "\0".join(s)
2125 2124 elif t == 'n':
2126 2125 return "_hexlist('%s')" % "\0".join(node.hex(a) for a in s)
2127 2126 elif t == 'b':
2128 2127 return "_list('%s')" % "\0".join(a.branch() for a in s)
2129 2128
2130 2129 m = l // 2
2131 2130 return '(%s or %s)' % (listexp(s[:m], t), listexp(s[m:], t))
2132 2131
2133 2132 ret = ''
2134 2133 pos = 0
2135 2134 arg = 0
2136 2135 while pos < len(expr):
2137 2136 c = expr[pos]
2138 2137 if c == '%':
2139 2138 pos += 1
2140 2139 d = expr[pos]
2141 2140 if d == '%':
2142 2141 ret += d
2143 2142 elif d in 'dsnbr':
2144 2143 ret += argtype(d, args[arg])
2145 2144 arg += 1
2146 2145 elif d == 'l':
2147 2146 # a list of some type
2148 2147 pos += 1
2149 2148 d = expr[pos]
2150 2149 ret += listexp(list(args[arg]), d)
2151 2150 arg += 1
2152 2151 else:
2153 2152 raise util.Abort('unexpected revspec format character %s' % d)
2154 2153 else:
2155 2154 ret += c
2156 2155 pos += 1
2157 2156
2158 2157 return ret
2159 2158
2160 2159 def prettyformat(tree):
2161 2160 def _prettyformat(tree, level, lines):
2162 2161 if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
2163 2162 lines.append((level, str(tree)))
2164 2163 else:
2165 2164 lines.append((level, '(%s' % tree[0]))
2166 2165 for s in tree[1:]:
2167 2166 _prettyformat(s, level + 1, lines)
2168 2167 lines[-1:] = [(lines[-1][0], lines[-1][1] + ')')]
2169 2168
2170 2169 lines = []
2171 2170 _prettyformat(tree, 0, lines)
2172 2171 output = '\n'.join((' '*l + s) for l, s in lines)
2173 2172 return output
2174 2173
2175 2174 def depth(tree):
2176 2175 if isinstance(tree, tuple):
2177 2176 return max(map(depth, tree)) + 1
2178 2177 else:
2179 2178 return 0
2180 2179
2181 2180 def funcsused(tree):
2182 2181 if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
2183 2182 return set()
2184 2183 else:
2185 2184 funcs = set()
2186 2185 for s in tree[1:]:
2187 2186 funcs |= funcsused(s)
2188 2187 if tree[0] == 'func':
2189 2188 funcs.add(tree[1][1])
2190 2189 return funcs
2191 2190
2192 2191 class baseset(list):
2193 2192 """Basic data structure that represents a revset and contains the basic
2194 2193 operation that it should be able to perform.
2195 2194
2196 2195 Every method in this class should be implemented by any smartset class.
2197 2196 """
2198 2197 def __init__(self, data=()):
2199 2198 super(baseset, self).__init__(data)
2200 2199 self._set = None
2201 2200
2202 2201 def ascending(self):
2203 2202 """Sorts the set in ascending order (in place).
2204 2203
2205 2204 This is part of the mandatory API for smartset."""
2206 2205 self.sort()
2207 2206
2208 2207 def descending(self):
2209 2208 """Sorts the set in descending order (in place).
2210 2209
2211 2210 This is part of the mandatory API for smartset."""
2212 2211 self.sort(reverse=True)
2213 2212
2214 2213 def min(self):
2215 2214 return min(self)
2216 2215
2217 2216 def max(self):
2218 2217 return max(self)
2219 2218
2220 2219 def set(self):
2221 2220 """Returns a set or a smartset containing all the elements.
2222 2221
2223 2222 The returned structure should be the fastest option for membership
2224 2223 testing.
2225 2224
2226 2225 This is part of the mandatory API for smartset."""
2227 2226 if not self._set:
2228 2227 self._set = set(self)
2229 2228 return self._set
2230 2229
2231 2230 def __sub__(self, other):
2232 2231 """Returns a new object with the substraction of the two collections.
2233 2232
2234 2233 This is part of the mandatory API for smartset."""
2235 2234 # If we are operating on 2 baseset, do the computation now since all
2236 2235 # data is available. The alternative is to involve a lazyset, which
2237 2236 # may be slow.
2238 2237 if isinstance(other, baseset):
2239 2238 other = other.set()
2240 2239 return baseset([x for x in self if x not in other])
2241 2240
2242 2241 return self.filter(lambda x: x not in other)
2243 2242
2244 2243 def __and__(self, other):
2245 2244 """Returns a new object with the intersection of the two collections.
2246 2245
2247 2246 This is part of the mandatory API for smartset."""
2248 2247 if isinstance(other, baseset):
2249 2248 other = other.set()
2250 2249 return baseset([y for y in self if y in other])
2251 2250
2252 2251 def __add__(self, other):
2253 2252 """Returns a new object with the union of the two collections.
2254 2253
2255 2254 This is part of the mandatory API for smartset."""
2256 2255 s = self.set()
2257 2256 l = [r for r in other if r not in s]
2258 2257 return baseset(list(self) + l)
2259 2258
2260 2259 def isascending(self):
2261 2260 """Returns True if the collection is ascending order, False if not.
2262 2261
2263 2262 This is part of the mandatory API for smartset."""
2264 2263 return False
2265 2264
2266 2265 def isdescending(self):
2267 2266 """Returns True if the collection is descending order, False if not.
2268 2267
2269 2268 This is part of the mandatory API for smartset."""
2270 2269 return False
2271 2270
2272 2271 def filter(self, condition):
2273 2272 """Returns this smartset filtered by condition as a new smartset.
2274 2273
2275 2274 `condition` is a callable which takes a revision number and returns a
2276 2275 boolean.
2277 2276
2278 2277 This is part of the mandatory API for smartset."""
2279 2278 return lazyset(self, condition)
2280 2279
2281 2280 class _orderedsetmixin(object):
2282 2281 """Mixin class with utility methods for smartsets
2283 2282
2284 2283 This should be extended by smartsets which have the isascending(),
2285 2284 isdescending() and reverse() methods"""
2286 2285
2287 2286 def _first(self):
2288 2287 """return the first revision in the set"""
2289 2288 for r in self:
2290 2289 return r
2291 2290 raise ValueError('arg is an empty sequence')
2292 2291
2293 2292 def _last(self):
2294 2293 """return the last revision in the set"""
2295 2294 self.reverse()
2296 2295 m = self._first()
2297 2296 self.reverse()
2298 2297 return m
2299 2298
2300 2299 def min(self):
2301 2300 """return the smallest element in the set"""
2302 2301 if self.isascending():
2303 2302 return self._first()
2304 2303 return self._last()
2305 2304
2306 2305 def max(self):
2307 2306 """return the largest element in the set"""
2308 2307 if self.isascending():
2309 2308 return self._last()
2310 2309 return self._first()
2311 2310
2312 2311 class lazyset(object):
2313 2312 """Duck type for baseset class which iterates lazily over the revisions in
2314 2313 the subset and contains a function which tests for membership in the
2315 2314 revset
2316 2315 """
2317 2316 def __init__(self, subset, condition=lambda x: True):
2318 2317 """
2319 2318 condition: a function that decide whether a revision in the subset
2320 2319 belongs to the revset or not.
2321 2320 """
2322 2321 self._subset = subset
2323 2322 self._condition = condition
2324 2323 self._cache = {}
2325 2324
2326 2325 def ascending(self):
2327 2326 self._subset.sort()
2328 2327
2329 2328 def descending(self):
2330 2329 self._subset.sort(reverse=True)
2331 2330
2332 2331 def min(self):
2333 2332 return min(self)
2334 2333
2335 2334 def max(self):
2336 2335 return max(self)
2337 2336
2338 2337 def __contains__(self, x):
2339 2338 c = self._cache
2340 2339 if x not in c:
2341 2340 c[x] = x in self._subset and self._condition(x)
2342 2341 return c[x]
2343 2342
2344 2343 def __iter__(self):
2345 2344 cond = self._condition
2346 2345 for x in self._subset:
2347 2346 if cond(x):
2348 2347 yield x
2349 2348
2350 2349 def __and__(self, x):
2351 2350 return lazyset(self, x.__contains__)
2352 2351
2353 2352 def __sub__(self, x):
2354 2353 return lazyset(self, lambda r: r not in x)
2355 2354
2356 2355 def __add__(self, x):
2357 2356 return _addset(self, x)
2358 2357
2359 2358 def __nonzero__(self):
2360 2359 for r in self:
2361 2360 return True
2362 2361 return False
2363 2362
2364 2363 def __len__(self):
2365 2364 # Basic implementation to be changed in future patches.
2366 2365 l = baseset([r for r in self])
2367 2366 return len(l)
2368 2367
2369 2368 def __getitem__(self, x):
2370 2369 # Basic implementation to be changed in future patches.
2371 2370 l = baseset([r for r in self])
2372 2371 return l[x]
2373 2372
2374 2373 def sort(self, reverse=False):
2375 2374 if not util.safehasattr(self._subset, 'sort'):
2376 2375 self._subset = baseset(self._subset)
2377 2376 self._subset.sort(reverse=reverse)
2378 2377
2379 2378 def reverse(self):
2380 2379 self._subset.reverse()
2381 2380
2382 2381 def set(self):
2383 2382 return set([r for r in self])
2384 2383
2385 2384 def isascending(self):
2386 2385 return False
2387 2386
2388 2387 def isdescending(self):
2389 2388 return False
2390 2389
2391 2390 def filter(self, l):
2392 2391 return lazyset(self, l)
2393 2392
2394 2393 class orderedlazyset(_orderedsetmixin, lazyset):
2395 2394 """Subclass of lazyset which subset can be ordered either ascending or
2396 2395 descendingly
2397 2396 """
2398 2397 def __init__(self, subset, condition, ascending=True):
2399 2398 super(orderedlazyset, self).__init__(subset, condition)
2400 2399 self._ascending = ascending
2401 2400
2402 2401 def filter(self, l):
2403 2402 return orderedlazyset(self, l, ascending=self._ascending)
2404 2403
2405 2404 def ascending(self):
2406 2405 if not self._ascending:
2407 2406 self.reverse()
2408 2407
2409 2408 def descending(self):
2410 2409 if self._ascending:
2411 2410 self.reverse()
2412 2411
2413 2412 def __and__(self, x):
2414 2413 return orderedlazyset(self, x.__contains__,
2415 2414 ascending=self._ascending)
2416 2415
2417 2416 def __sub__(self, x):
2418 2417 return orderedlazyset(self, lambda r: r not in x,
2419 2418 ascending=self._ascending)
2420 2419
2421 2420 def __add__(self, x):
2422 2421 kwargs = {}
2423 2422 if self.isascending() and x.isascending():
2424 2423 kwargs['ascending'] = True
2425 2424 if self.isdescending() and x.isdescending():
2426 2425 kwargs['ascending'] = False
2427 2426 return _addset(self, x, **kwargs)
2428 2427
2429 2428 def sort(self, reverse=False):
2430 2429 if reverse:
2431 2430 if self._ascending:
2432 2431 self._subset.sort(reverse=reverse)
2433 2432 else:
2434 2433 if not self._ascending:
2435 2434 self._subset.sort(reverse=reverse)
2436 2435 self._ascending = not reverse
2437 2436
2438 2437 def isascending(self):
2439 2438 return self._ascending
2440 2439
2441 2440 def isdescending(self):
2442 2441 return not self._ascending
2443 2442
2444 2443 def reverse(self):
2445 2444 self._subset.reverse()
2446 2445 self._ascending = not self._ascending
2447 2446
2448 2447 class _addset(_orderedsetmixin):
2449 2448 """Represent the addition of two sets
2450 2449
2451 2450 Wrapper structure for lazily adding two structures without losing much
2452 2451 performance on the __contains__ method
2453 2452
2454 2453 If the ascending attribute is set, that means the two structures are
2455 2454 ordered in either an ascending or descending way. Therefore, we can add
2456 2455 them maintaining the order by iterating over both at the same time
2457 2456
2458 2457 This class does not duck-type baseset and it's only supposed to be used
2459 2458 internally
2460 2459 """
2461 2460 def __init__(self, revs1, revs2, ascending=None):
2462 2461 self._r1 = revs1
2463 2462 self._r2 = revs2
2464 2463 self._iter = None
2465 2464 self._ascending = ascending
2466 2465 self._genlist = None
2467 2466
2468 2467 def __len__(self):
2469 2468 return len(self._list)
2470 2469
2471 2470 @util.propertycache
2472 2471 def _list(self):
2473 2472 if not self._genlist:
2474 2473 self._genlist = baseset(self._iterator())
2475 2474 return self._genlist
2476 2475
2477 2476 def filter(self, condition):
2478 2477 if self._ascending is not None:
2479 2478 return orderedlazyset(self, condition, ascending=self._ascending)
2480 2479 return lazyset(self, condition)
2481 2480
2482 2481 def ascending(self):
2483 2482 if self._ascending is None:
2484 2483 self.sort()
2485 2484 self._ascending = True
2486 2485 else:
2487 2486 if not self._ascending:
2488 2487 self.reverse()
2489 2488
2490 2489 def descending(self):
2491 2490 if self._ascending is None:
2492 2491 self.sort(reverse=True)
2493 2492 self._ascending = False
2494 2493 else:
2495 2494 if self._ascending:
2496 2495 self.reverse()
2497 2496
2498 2497 def __and__(self, other):
2499 2498 filterfunc = other.__contains__
2500 2499 if self._ascending is not None:
2501 2500 return orderedlazyset(self, filterfunc, ascending=self._ascending)
2502 2501 return lazyset(self, filterfunc)
2503 2502
2504 2503 def __sub__(self, other):
2505 2504 filterfunc = lambda r: r not in other
2506 2505 if self._ascending is not None:
2507 2506 return orderedlazyset(self, filterfunc, ascending=self._ascending)
2508 2507 return lazyset(self, filterfunc)
2509 2508
2510 2509 def __add__(self, other):
2511 2510 """When both collections are ascending or descending, preserve the order
2512 2511 """
2513 2512 kwargs = {}
2514 2513 if self._ascending is not None:
2515 2514 if self.isascending() and other.isascending():
2516 2515 kwargs['ascending'] = True
2517 2516 if self.isdescending() and other.isdescending():
2518 2517 kwargs['ascending'] = False
2519 2518 return _addset(self, other, **kwargs)
2520 2519
2521 2520 def _iterator(self):
2522 2521 """Iterate over both collections without repeating elements
2523 2522
2524 2523 If the ascending attribute is not set, iterate over the first one and
2525 2524 then over the second one checking for membership on the first one so we
2526 2525 dont yield any duplicates.
2527 2526
2528 2527 If the ascending attribute is set, iterate over both collections at the
2529 2528 same time, yielding only one value at a time in the given order.
2530 2529 """
2531 2530 if not self._iter:
2532 2531 def gen():
2533 2532 if self._ascending is None:
2534 2533 for r in self._r1:
2535 2534 yield r
2536 2535 s = self._r1.set()
2537 2536 for r in self._r2:
2538 2537 if r not in s:
2539 2538 yield r
2540 2539 else:
2541 2540 iter1 = iter(self._r1)
2542 2541 iter2 = iter(self._r2)
2543 2542
2544 2543 val1 = None
2545 2544 val2 = None
2546 2545
2547 2546 choice = max
2548 2547 if self._ascending:
2549 2548 choice = min
2550 2549 try:
2551 2550 # Consume both iterators in an ordered way until one is
2552 2551 # empty
2553 2552 while True:
2554 2553 if val1 is None:
2555 2554 val1 = iter1.next()
2556 2555 if val2 is None:
2557 2556 val2 = iter2.next()
2558 2557 next = choice(val1, val2)
2559 2558 yield next
2560 2559 if val1 == next:
2561 2560 val1 = None
2562 2561 if val2 == next:
2563 2562 val2 = None
2564 2563 except StopIteration:
2565 2564 # Flush any remaining values and consume the other one
2566 2565 it = iter2
2567 2566 if val1 is not None:
2568 2567 yield val1
2569 2568 it = iter1
2570 2569 elif val2 is not None:
2571 2570 # might have been equality and both are empty
2572 2571 yield val2
2573 2572 for val in it:
2574 2573 yield val
2575 2574
2576 2575 self._iter = _generatorset(gen())
2577 2576
2578 2577 return self._iter
2579 2578
2580 2579 def __iter__(self):
2581 2580 if self._genlist:
2582 2581 return iter(self._genlist)
2583 2582 return iter(self._iterator())
2584 2583
2585 2584 def __contains__(self, x):
2586 2585 return x in self._r1 or x in self._r2
2587 2586
2588 2587 def set(self):
2589 2588 return self
2590 2589
2591 2590 def sort(self, reverse=False):
2592 2591 """Sort the added set
2593 2592
2594 2593 For this we use the cached list with all the generated values and if we
2595 2594 know they are ascending or descending we can sort them in a smart way.
2596 2595 """
2597 2596 if self._ascending is None:
2598 2597 self._list.sort(reverse=reverse)
2599 2598 self._ascending = not reverse
2600 2599 else:
2601 2600 if bool(self._ascending) == bool(reverse):
2602 2601 self.reverse()
2603 2602
2604 2603 def isascending(self):
2605 2604 return self._ascending is not None and self._ascending
2606 2605
2607 2606 def isdescending(self):
2608 2607 return self._ascending is not None and not self._ascending
2609 2608
2610 2609 def reverse(self):
2611 2610 self._list.reverse()
2612 2611 if self._ascending is not None:
2613 2612 self._ascending = not self._ascending
2614 2613
2615 2614 class _generatorset(object):
2616 2615 """Wrap a generator for lazy iteration
2617 2616
2618 2617 Wrapper structure for generators that provides lazy membership and can
2619 2618 be iterated more than once.
2620 2619 When asked for membership it generates values until either it finds the
2621 2620 requested one or has gone through all the elements in the generator
2622 2621
2623 2622 This class does not duck-type baseset and it's only supposed to be used
2624 2623 internally
2625 2624 """
2626 2625 def __init__(self, gen):
2627 2626 """
2628 2627 gen: a generator producing the values for the generatorset.
2629 2628 """
2630 2629 self._gen = gen
2631 2630 self._cache = {}
2632 2631 self._genlist = baseset([])
2633 2632 self._finished = False
2634 2633
2635 2634 def __contains__(self, x):
2636 2635 if x in self._cache:
2637 2636 return self._cache[x]
2638 2637
2639 2638 # Use new values only, as existing values would be cached.
2640 2639 for l in self._consumegen():
2641 2640 if l == x:
2642 2641 return True
2643 2642
2644 2643 self._cache[x] = False
2645 2644 return False
2646 2645
2647 2646 def __iter__(self):
2648 2647 if self._finished:
2649 2648 for x in self._genlist:
2650 2649 yield x
2651 2650 return
2652 2651
2653 2652 # We have to use this complex iteration strategy to allow multiple
2654 2653 # iterations at the same time. We need to be able to catch revision
2655 2654 # removed from `consumegen` and added to genlist in another instance.
2656 2655 #
2657 2656 # Getting rid of it would provide an about 15% speed up on this
2658 2657 # iteration.
2659 2658 i = 0
2660 2659 genlist = self._genlist
2661 2660 consume = self._consumegen()
2662 2661 while True:
2663 2662 if i < len(genlist):
2664 2663 yield genlist[i]
2665 2664 else:
2666 2665 yield consume.next()
2667 2666 i += 1
2668 2667
2669 2668 def _consumegen(self):
2670 2669 for item in self._gen:
2671 2670 self._cache[item] = True
2672 2671 self._genlist.append(item)
2673 2672 yield item
2674 2673 self._finished = True
2675 2674
2676 2675 def set(self):
2677 2676 return self
2678 2677
2679 2678 def sort(self, reverse=False):
2680 2679 if not self._finished:
2681 2680 for i in self:
2682 2681 continue
2683 2682 self._genlist.sort(reverse=reverse)
2684 2683
2685 2684 class _ascgeneratorset(_generatorset):
2686 2685 """Wrap a generator of ascending elements for lazy iteration
2687 2686
2688 2687 Same structure as _generatorset but stops iterating after it goes past
2689 2688 the value when asked for membership and the element is not contained
2690 2689
2691 2690 This class does not duck-type baseset and it's only supposed to be used
2692 2691 internally
2693 2692 """
2694 2693 def __contains__(self, x):
2695 2694 if x in self._cache:
2696 2695 return self._cache[x]
2697 2696
2698 2697 # Use new values only, as existing values would be cached.
2699 2698 for l in self._consumegen():
2700 2699 if l == x:
2701 2700 return True
2702 2701 if l > x:
2703 2702 break
2704 2703
2705 2704 self._cache[x] = False
2706 2705 return False
2707 2706
2708 2707 class _descgeneratorset(_generatorset):
2709 2708 """Wrap a generator of descending elements for lazy iteration
2710 2709
2711 2710 Same structure as _generatorset but stops iterating after it goes past
2712 2711 the value when asked for membership and the element is not contained
2713 2712
2714 2713 This class does not duck-type baseset and it's only supposed to be used
2715 2714 internally
2716 2715 """
2717 2716 def __contains__(self, x):
2718 2717 if x in self._cache:
2719 2718 return self._cache[x]
2720 2719
2721 2720 # Use new values only, as existing values would be cached.
2722 2721 for l in self._consumegen():
2723 2722 if l == x:
2724 2723 return True
2725 2724 if l < x:
2726 2725 break
2727 2726
2728 2727 self._cache[x] = False
2729 2728 return False
2730 2729
2731 2730 class spanset(_orderedsetmixin):
2732 2731 """Duck type for baseset class which represents a range of revisions and
2733 2732 can work lazily and without having all the range in memory
2734 2733
2735 2734 Note that spanset(x, y) behave almost like xrange(x, y) except for two
2736 2735 notable points:
2737 2736 - when x < y it will be automatically descending,
2738 2737 - revision filtered with this repoview will be skipped.
2739 2738
2740 2739 """
2741 2740 def __init__(self, repo, start=0, end=None):
2742 2741 """
2743 2742 start: first revision included the set
2744 2743 (default to 0)
2745 2744 end: first revision excluded (last+1)
2746 2745 (default to len(repo)
2747 2746
2748 2747 Spanset will be descending if `end` < `start`.
2749 2748 """
2750 2749 self._start = start
2751 2750 if end is not None:
2752 2751 self._end = end
2753 2752 else:
2754 2753 self._end = len(repo)
2755 2754 self._hiddenrevs = repo.changelog.filteredrevs
2756 2755
2757 2756 def ascending(self):
2758 2757 if not self.isascending():
2759 2758 self.reverse()
2760 2759
2761 2760 def descending(self):
2762 2761 if not self.isdescending():
2763 2762 self.reverse()
2764 2763
2765 2764 def __iter__(self):
2766 2765 if self.isascending():
2767 2766 iterrange = xrange(self._start, self._end)
2768 2767 else:
2769 2768 iterrange = xrange(self._start, self._end, -1)
2770 2769
2771 2770 if self._hiddenrevs:
2772 2771 s = self._hiddenrevs
2773 2772 for r in iterrange:
2774 2773 if r not in s:
2775 2774 yield r
2776 2775 else:
2777 2776 for r in iterrange:
2778 2777 yield r
2779 2778
2780 2779 def __contains__(self, rev):
2781 2780 return (((self._end < rev <= self._start)
2782 2781 or (self._start <= rev < self._end))
2783 2782 and not (self._hiddenrevs and rev in self._hiddenrevs))
2784 2783
2785 2784 def __nonzero__(self):
2786 2785 for r in self:
2787 2786 return True
2788 2787 return False
2789 2788
2790 2789 def __and__(self, x):
2791 2790 if isinstance(x, baseset):
2792 2791 x = x.set()
2793 2792 return orderedlazyset(self, x.__contains__,
2794 2793 ascending=self.isascending())
2795 2794
2796 2795 def __sub__(self, x):
2797 2796 if isinstance(x, baseset):
2798 2797 x = x.set()
2799 2798 return orderedlazyset(self, lambda r: r not in x,
2800 2799 ascending=self.isascending())
2801 2800
2802 2801 def __add__(self, x):
2803 2802 kwargs = {}
2804 2803 if self.isascending() and x.isascending():
2805 2804 kwargs['ascending'] = True
2806 2805 if self.isdescending() and x.isdescending():
2807 2806 kwargs['ascending'] = False
2808 2807 return _addset(self, x, **kwargs)
2809 2808
2810 2809 def __len__(self):
2811 2810 if not self._hiddenrevs:
2812 2811 return abs(self._end - self._start)
2813 2812 else:
2814 2813 count = 0
2815 2814 start = self._start
2816 2815 end = self._end
2817 2816 for rev in self._hiddenrevs:
2818 2817 if (end < rev <= start) or (start <= rev < end):
2819 2818 count += 1
2820 2819 return abs(self._end - self._start) - count
2821 2820
2822 2821 def __getitem__(self, x):
2823 2822 # Basic implementation to be changed in future patches.
2824 2823 l = baseset([r for r in self])
2825 2824 return l[x]
2826 2825
2827 2826 def sort(self, reverse=False):
2828 2827 if bool(reverse) != (self._start > self._end):
2829 2828 self.reverse()
2830 2829
2831 2830 def reverse(self):
2832 2831 # Just switch the _start and _end parameters
2833 2832 if self.isascending():
2834 2833 self._start, self._end = self._end - 1, self._start - 1
2835 2834 else:
2836 2835 self._start, self._end = self._end + 1, self._start + 1
2837 2836
2838 2837 def set(self):
2839 2838 return self
2840 2839
2841 2840 def isascending(self):
2842 2841 return self._start <= self._end
2843 2842
2844 2843 def isdescending(self):
2845 2844 return self._start >= self._end
2846 2845
2847 2846 def filter(self, l):
2848 2847 return orderedlazyset(self, l, ascending=self.isascending())
2849 2848
2850 2849 # tell hggettext to extract docstrings from these functions:
2851 2850 i18nfunctions = symbols.values()
General Comments 0
You need to be logged in to leave comments. Login now