##// END OF EJS Templates
revset: optimize baseset.__sub__ (issue4313)...
Gregory Szorc -
r21939:f486001f stable
parent child Browse files
Show More
@@ -1,23 +1,24 b''
1 1 all()
2 2 draft()
3 3 ::tip
4 4 draft() and ::tip
5 5 ::tip and draft()
6 6 0::tip
7 7 roots(0::tip)
8 8 author(lmoscovicz)
9 9 author(mpm)
10 10 author(lmoscovicz) or author(mpm)
11 11 author(mpm) or author(lmoscovicz)
12 12 tip:0
13 13 max(tip:0)
14 14 min(0:tip)
15 15 0::
16 16 min(0::)
17 17 roots((tip~100::) - (tip~100::tip))
18 18 ::p1(p1(tip))::
19 19 public()
20 20 :10000 and public()
21 21 draft()
22 22 :10000 and draft()
23 23 max(::(tip~20) - obsolete())
24 roots((0:tip)::)
@@ -1,2845 +1,2852 b''
1 1 # revset.py - revision set queries for mercurial
2 2 #
3 3 # Copyright 2010 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 import re
9 9 import parser, util, error, discovery, hbisect, phases
10 10 import node
11 11 import heapq
12 12 import match as matchmod
13 13 import ancestor as ancestormod
14 14 from i18n import _
15 15 import encoding
16 16 import obsolete as obsmod
17 17 import pathutil
18 18 import repoview
19 19
20 20 def _revancestors(repo, revs, followfirst):
21 21 """Like revlog.ancestors(), but supports followfirst."""
22 22 cut = followfirst and 1 or None
23 23 cl = repo.changelog
24 24
25 25 def iterate():
26 26 revqueue, revsnode = None, None
27 27 h = []
28 28
29 29 revs.descending()
30 30 revqueue = util.deque(revs)
31 31 if revqueue:
32 32 revsnode = revqueue.popleft()
33 33 heapq.heappush(h, -revsnode)
34 34
35 35 seen = set([node.nullrev])
36 36 while h:
37 37 current = -heapq.heappop(h)
38 38 if current not in seen:
39 39 if revsnode and current == revsnode:
40 40 if revqueue:
41 41 revsnode = revqueue.popleft()
42 42 heapq.heappush(h, -revsnode)
43 43 seen.add(current)
44 44 yield current
45 45 for parent in cl.parentrevs(current)[:cut]:
46 46 if parent != node.nullrev:
47 47 heapq.heappush(h, -parent)
48 48
49 49 return _descgeneratorset(iterate())
50 50
51 51 def _revdescendants(repo, revs, followfirst):
52 52 """Like revlog.descendants() but supports followfirst."""
53 53 cut = followfirst and 1 or None
54 54
55 55 def iterate():
56 56 cl = repo.changelog
57 57 first = min(revs)
58 58 nullrev = node.nullrev
59 59 if first == nullrev:
60 60 # Are there nodes with a null first parent and a non-null
61 61 # second one? Maybe. Do we care? Probably not.
62 62 for i in cl:
63 63 yield i
64 64 else:
65 65 seen = set(revs)
66 66 for i in cl.revs(first + 1):
67 67 for x in cl.parentrevs(i)[:cut]:
68 68 if x != nullrev and x in seen:
69 69 seen.add(i)
70 70 yield i
71 71 break
72 72
73 73 return _ascgeneratorset(iterate())
74 74
75 75 def _revsbetween(repo, roots, heads):
76 76 """Return all paths between roots and heads, inclusive of both endpoint
77 77 sets."""
78 78 if not roots:
79 79 return baseset([])
80 80 parentrevs = repo.changelog.parentrevs
81 81 visit = baseset(heads)
82 82 reachable = set()
83 83 seen = {}
84 84 minroot = min(roots)
85 85 roots = set(roots)
86 86 # open-code the post-order traversal due to the tiny size of
87 87 # sys.getrecursionlimit()
88 88 while visit:
89 89 rev = visit.pop()
90 90 if rev in roots:
91 91 reachable.add(rev)
92 92 parents = parentrevs(rev)
93 93 seen[rev] = parents
94 94 for parent in parents:
95 95 if parent >= minroot and parent not in seen:
96 96 visit.append(parent)
97 97 if not reachable:
98 98 return baseset([])
99 99 for rev in sorted(seen):
100 100 for parent in seen[rev]:
101 101 if parent in reachable:
102 102 reachable.add(rev)
103 103 return baseset(sorted(reachable))
104 104
105 105 elements = {
106 106 "(": (20, ("group", 1, ")"), ("func", 1, ")")),
107 107 "~": (18, None, ("ancestor", 18)),
108 108 "^": (18, None, ("parent", 18), ("parentpost", 18)),
109 109 "-": (5, ("negate", 19), ("minus", 5)),
110 110 "::": (17, ("dagrangepre", 17), ("dagrange", 17),
111 111 ("dagrangepost", 17)),
112 112 "..": (17, ("dagrangepre", 17), ("dagrange", 17),
113 113 ("dagrangepost", 17)),
114 114 ":": (15, ("rangepre", 15), ("range", 15), ("rangepost", 15)),
115 115 "not": (10, ("not", 10)),
116 116 "!": (10, ("not", 10)),
117 117 "and": (5, None, ("and", 5)),
118 118 "&": (5, None, ("and", 5)),
119 119 "or": (4, None, ("or", 4)),
120 120 "|": (4, None, ("or", 4)),
121 121 "+": (4, None, ("or", 4)),
122 122 ",": (2, None, ("list", 2)),
123 123 ")": (0, None, None),
124 124 "symbol": (0, ("symbol",), None),
125 125 "string": (0, ("string",), None),
126 126 "end": (0, None, None),
127 127 }
128 128
129 129 keywords = set(['and', 'or', 'not'])
130 130
131 131 def tokenize(program, lookup=None):
132 132 '''
133 133 Parse a revset statement into a stream of tokens
134 134
135 135 Check that @ is a valid unquoted token character (issue3686):
136 136 >>> list(tokenize("@::"))
137 137 [('symbol', '@', 0), ('::', None, 1), ('end', None, 3)]
138 138
139 139 '''
140 140
141 141 pos, l = 0, len(program)
142 142 while pos < l:
143 143 c = program[pos]
144 144 if c.isspace(): # skip inter-token whitespace
145 145 pass
146 146 elif c == ':' and program[pos:pos + 2] == '::': # look ahead carefully
147 147 yield ('::', None, pos)
148 148 pos += 1 # skip ahead
149 149 elif c == '.' and program[pos:pos + 2] == '..': # look ahead carefully
150 150 yield ('..', None, pos)
151 151 pos += 1 # skip ahead
152 152 elif c in "():,-|&+!~^": # handle simple operators
153 153 yield (c, None, pos)
154 154 elif (c in '"\'' or c == 'r' and
155 155 program[pos:pos + 2] in ("r'", 'r"')): # handle quoted strings
156 156 if c == 'r':
157 157 pos += 1
158 158 c = program[pos]
159 159 decode = lambda x: x
160 160 else:
161 161 decode = lambda x: x.decode('string-escape')
162 162 pos += 1
163 163 s = pos
164 164 while pos < l: # find closing quote
165 165 d = program[pos]
166 166 if d == '\\': # skip over escaped characters
167 167 pos += 2
168 168 continue
169 169 if d == c:
170 170 yield ('string', decode(program[s:pos]), s)
171 171 break
172 172 pos += 1
173 173 else:
174 174 raise error.ParseError(_("unterminated string"), s)
175 175 # gather up a symbol/keyword
176 176 elif c.isalnum() or c in '._@' or ord(c) > 127:
177 177 s = pos
178 178 pos += 1
179 179 while pos < l: # find end of symbol
180 180 d = program[pos]
181 181 if not (d.isalnum() or d in "-._/@" or ord(d) > 127):
182 182 break
183 183 if d == '.' and program[pos - 1] == '.': # special case for ..
184 184 pos -= 1
185 185 break
186 186 pos += 1
187 187 sym = program[s:pos]
188 188 if sym in keywords: # operator keywords
189 189 yield (sym, None, s)
190 190 elif '-' in sym:
191 191 # some jerk gave us foo-bar-baz, try to check if it's a symbol
192 192 if lookup and lookup(sym):
193 193 # looks like a real symbol
194 194 yield ('symbol', sym, s)
195 195 else:
196 196 # looks like an expression
197 197 parts = sym.split('-')
198 198 for p in parts[:-1]:
199 199 if p: # possible consecutive -
200 200 yield ('symbol', p, s)
201 201 s += len(p)
202 202 yield ('-', None, pos)
203 203 s += 1
204 204 if parts[-1]: # possible trailing -
205 205 yield ('symbol', parts[-1], s)
206 206 else:
207 207 yield ('symbol', sym, s)
208 208 pos -= 1
209 209 else:
210 210 raise error.ParseError(_("syntax error"), pos)
211 211 pos += 1
212 212 yield ('end', None, pos)
213 213
214 214 # helpers
215 215
216 216 def getstring(x, err):
217 217 if x and (x[0] == 'string' or x[0] == 'symbol'):
218 218 return x[1]
219 219 raise error.ParseError(err)
220 220
221 221 def getlist(x):
222 222 if not x:
223 223 return []
224 224 if x[0] == 'list':
225 225 return getlist(x[1]) + [x[2]]
226 226 return [x]
227 227
228 228 def getargs(x, min, max, err):
229 229 l = getlist(x)
230 230 if len(l) < min or (max >= 0 and len(l) > max):
231 231 raise error.ParseError(err)
232 232 return l
233 233
234 234 def getset(repo, subset, x):
235 235 if not x:
236 236 raise error.ParseError(_("missing argument"))
237 237 s = methods[x[0]](repo, subset, *x[1:])
238 238 if util.safehasattr(s, 'set'):
239 239 return s
240 240 return baseset(s)
241 241
242 242 def _getrevsource(repo, r):
243 243 extra = repo[r].extra()
244 244 for label in ('source', 'transplant_source', 'rebase_source'):
245 245 if label in extra:
246 246 try:
247 247 return repo[extra[label]].rev()
248 248 except error.RepoLookupError:
249 249 pass
250 250 return None
251 251
252 252 # operator methods
253 253
254 254 def stringset(repo, subset, x):
255 255 x = repo[x].rev()
256 256 if x == -1 and len(subset) == len(repo):
257 257 return baseset([-1])
258 258 if len(subset) == len(repo) or x in subset:
259 259 return baseset([x])
260 260 return baseset([])
261 261
262 262 def symbolset(repo, subset, x):
263 263 if x in symbols:
264 264 raise error.ParseError(_("can't use %s here") % x)
265 265 return stringset(repo, subset, x)
266 266
267 267 def rangeset(repo, subset, x, y):
268 268 cl = baseset(repo.changelog)
269 269 m = getset(repo, cl, x)
270 270 n = getset(repo, cl, y)
271 271
272 272 if not m or not n:
273 273 return baseset([])
274 274 m, n = m[0], n[-1]
275 275
276 276 if m < n:
277 277 r = spanset(repo, m, n + 1)
278 278 else:
279 279 r = spanset(repo, m, n - 1)
280 280 return r & subset
281 281
282 282 def dagrange(repo, subset, x, y):
283 283 r = spanset(repo)
284 284 xs = _revsbetween(repo, getset(repo, r, x), getset(repo, r, y))
285 285 s = subset.set()
286 286 return xs.filter(s.__contains__)
287 287
288 288 def andset(repo, subset, x, y):
289 289 return getset(repo, getset(repo, subset, x), y)
290 290
291 291 def orset(repo, subset, x, y):
292 292 xl = getset(repo, subset, x)
293 293 yl = getset(repo, subset - xl, y)
294 294 return xl + yl
295 295
296 296 def notset(repo, subset, x):
297 297 return subset - getset(repo, subset, x)
298 298
299 299 def listset(repo, subset, a, b):
300 300 raise error.ParseError(_("can't use a list in this context"))
301 301
302 302 def func(repo, subset, a, b):
303 303 if a[0] == 'symbol' and a[1] in symbols:
304 304 return symbols[a[1]](repo, subset, b)
305 305 raise error.ParseError(_("not a function: %s") % a[1])
306 306
307 307 # functions
308 308
309 309 def adds(repo, subset, x):
310 310 """``adds(pattern)``
311 311 Changesets that add a file matching pattern.
312 312
313 313 The pattern without explicit kind like ``glob:`` is expected to be
314 314 relative to the current directory and match against a file or a
315 315 directory.
316 316 """
317 317 # i18n: "adds" is a keyword
318 318 pat = getstring(x, _("adds requires a pattern"))
319 319 return checkstatus(repo, subset, pat, 1)
320 320
321 321 def ancestor(repo, subset, x):
322 322 """``ancestor(*changeset)``
323 323 A greatest common ancestor of the changesets.
324 324
325 325 Accepts 0 or more changesets.
326 326 Will return empty list when passed no args.
327 327 Greatest common ancestor of a single changeset is that changeset.
328 328 """
329 329 # i18n: "ancestor" is a keyword
330 330 l = getlist(x)
331 331 rl = spanset(repo)
332 332 anc = None
333 333
334 334 # (getset(repo, rl, i) for i in l) generates a list of lists
335 335 for revs in (getset(repo, rl, i) for i in l):
336 336 for r in revs:
337 337 if anc is None:
338 338 anc = repo[r]
339 339 else:
340 340 anc = anc.ancestor(repo[r])
341 341
342 342 if anc is not None and anc.rev() in subset:
343 343 return baseset([anc.rev()])
344 344 return baseset([])
345 345
346 346 def _ancestors(repo, subset, x, followfirst=False):
347 347 args = getset(repo, spanset(repo), x)
348 348 if not args:
349 349 return baseset([])
350 350 s = _revancestors(repo, args, followfirst)
351 351 return subset.filter(s.__contains__)
352 352
353 353 def ancestors(repo, subset, x):
354 354 """``ancestors(set)``
355 355 Changesets that are ancestors of a changeset in set.
356 356 """
357 357 return _ancestors(repo, subset, x)
358 358
359 359 def _firstancestors(repo, subset, x):
360 360 # ``_firstancestors(set)``
361 361 # Like ``ancestors(set)`` but follows only the first parents.
362 362 return _ancestors(repo, subset, x, followfirst=True)
363 363
364 364 def ancestorspec(repo, subset, x, n):
365 365 """``set~n``
366 366 Changesets that are the Nth ancestor (first parents only) of a changeset
367 367 in set.
368 368 """
369 369 try:
370 370 n = int(n[1])
371 371 except (TypeError, ValueError):
372 372 raise error.ParseError(_("~ expects a number"))
373 373 ps = set()
374 374 cl = repo.changelog
375 375 for r in getset(repo, baseset(cl), x):
376 376 for i in range(n):
377 377 r = cl.parentrevs(r)[0]
378 378 ps.add(r)
379 379 return subset.filter(ps.__contains__)
380 380
381 381 def author(repo, subset, x):
382 382 """``author(string)``
383 383 Alias for ``user(string)``.
384 384 """
385 385 # i18n: "author" is a keyword
386 386 n = encoding.lower(getstring(x, _("author requires a string")))
387 387 kind, pattern, matcher = _substringmatcher(n)
388 388 return subset.filter(lambda x: matcher(encoding.lower(repo[x].user())))
389 389
390 390 def only(repo, subset, x):
391 391 """``only(set, [set])``
392 392 Changesets that are ancestors of the first set that are not ancestors
393 393 of any other head in the repo. If a second set is specified, the result
394 394 is ancestors of the first set that are not ancestors of the second set
395 395 (i.e. ::<set1> - ::<set2>).
396 396 """
397 397 cl = repo.changelog
398 398 # i18n: "only" is a keyword
399 399 args = getargs(x, 1, 2, _('only takes one or two arguments'))
400 400 include = getset(repo, spanset(repo), args[0]).set()
401 401 if len(args) == 1:
402 402 if len(include) == 0:
403 403 return baseset([])
404 404
405 405 descendants = set(_revdescendants(repo, include, False))
406 406 exclude = [rev for rev in cl.headrevs()
407 407 if not rev in descendants and not rev in include]
408 408 else:
409 409 exclude = getset(repo, spanset(repo), args[1])
410 410
411 411 results = set(ancestormod.missingancestors(include, exclude, cl.parentrevs))
412 412 return lazyset(subset, results.__contains__)
413 413
414 414 def bisect(repo, subset, x):
415 415 """``bisect(string)``
416 416 Changesets marked in the specified bisect status:
417 417
418 418 - ``good``, ``bad``, ``skip``: csets explicitly marked as good/bad/skip
419 419 - ``goods``, ``bads`` : csets topologically good/bad
420 420 - ``range`` : csets taking part in the bisection
421 421 - ``pruned`` : csets that are goods, bads or skipped
422 422 - ``untested`` : csets whose fate is yet unknown
423 423 - ``ignored`` : csets ignored due to DAG topology
424 424 - ``current`` : the cset currently being bisected
425 425 """
426 426 # i18n: "bisect" is a keyword
427 427 status = getstring(x, _("bisect requires a string")).lower()
428 428 state = set(hbisect.get(repo, status))
429 429 return subset.filter(state.__contains__)
430 430
431 431 # Backward-compatibility
432 432 # - no help entry so that we do not advertise it any more
433 433 def bisected(repo, subset, x):
434 434 return bisect(repo, subset, x)
435 435
436 436 def bookmark(repo, subset, x):
437 437 """``bookmark([name])``
438 438 The named bookmark or all bookmarks.
439 439
440 440 If `name` starts with `re:`, the remainder of the name is treated as
441 441 a regular expression. To match a bookmark that actually starts with `re:`,
442 442 use the prefix `literal:`.
443 443 """
444 444 # i18n: "bookmark" is a keyword
445 445 args = getargs(x, 0, 1, _('bookmark takes one or no arguments'))
446 446 if args:
447 447 bm = getstring(args[0],
448 448 # i18n: "bookmark" is a keyword
449 449 _('the argument to bookmark must be a string'))
450 450 kind, pattern, matcher = _stringmatcher(bm)
451 451 if kind == 'literal':
452 452 bmrev = repo._bookmarks.get(bm, None)
453 453 if not bmrev:
454 454 raise util.Abort(_("bookmark '%s' does not exist") % bm)
455 455 bmrev = repo[bmrev].rev()
456 456 return subset.filter(lambda r: r == bmrev)
457 457 else:
458 458 matchrevs = set()
459 459 for name, bmrev in repo._bookmarks.iteritems():
460 460 if matcher(name):
461 461 matchrevs.add(bmrev)
462 462 if not matchrevs:
463 463 raise util.Abort(_("no bookmarks exist that match '%s'")
464 464 % pattern)
465 465 bmrevs = set()
466 466 for bmrev in matchrevs:
467 467 bmrevs.add(repo[bmrev].rev())
468 468 return subset & bmrevs
469 469
470 470 bms = set([repo[r].rev()
471 471 for r in repo._bookmarks.values()])
472 472 return subset.filter(bms.__contains__)
473 473
474 474 def branch(repo, subset, x):
475 475 """``branch(string or set)``
476 476 All changesets belonging to the given branch or the branches of the given
477 477 changesets.
478 478
479 479 If `string` starts with `re:`, the remainder of the name is treated as
480 480 a regular expression. To match a branch that actually starts with `re:`,
481 481 use the prefix `literal:`.
482 482 """
483 483 try:
484 484 b = getstring(x, '')
485 485 except error.ParseError:
486 486 # not a string, but another revspec, e.g. tip()
487 487 pass
488 488 else:
489 489 kind, pattern, matcher = _stringmatcher(b)
490 490 if kind == 'literal':
491 491 # note: falls through to the revspec case if no branch with
492 492 # this name exists
493 493 if pattern in repo.branchmap():
494 494 return subset.filter(lambda r: matcher(repo[r].branch()))
495 495 else:
496 496 return subset.filter(lambda r: matcher(repo[r].branch()))
497 497
498 498 s = getset(repo, spanset(repo), x)
499 499 b = set()
500 500 for r in s:
501 501 b.add(repo[r].branch())
502 502 s = s.set()
503 503 return subset.filter(lambda r: r in s or repo[r].branch() in b)
504 504
505 505 def bumped(repo, subset, x):
506 506 """``bumped()``
507 507 Mutable changesets marked as successors of public changesets.
508 508
509 509 Only non-public and non-obsolete changesets can be `bumped`.
510 510 """
511 511 # i18n: "bumped" is a keyword
512 512 getargs(x, 0, 0, _("bumped takes no arguments"))
513 513 bumped = obsmod.getrevs(repo, 'bumped')
514 514 return subset & bumped
515 515
516 516 def bundle(repo, subset, x):
517 517 """``bundle()``
518 518 Changesets in the bundle.
519 519
520 520 Bundle must be specified by the -R option."""
521 521
522 522 try:
523 523 bundlerevs = repo.changelog.bundlerevs
524 524 except AttributeError:
525 525 raise util.Abort(_("no bundle provided - specify with -R"))
526 526 return subset & bundlerevs
527 527
528 528 def checkstatus(repo, subset, pat, field):
529 529 hasset = matchmod.patkind(pat) == 'set'
530 530
531 531 def matches(x):
532 532 m = None
533 533 fname = None
534 534 c = repo[x]
535 535 if not m or hasset:
536 536 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
537 537 if not m.anypats() and len(m.files()) == 1:
538 538 fname = m.files()[0]
539 539 if fname is not None:
540 540 if fname not in c.files():
541 541 return False
542 542 else:
543 543 for f in c.files():
544 544 if m(f):
545 545 break
546 546 else:
547 547 return False
548 548 files = repo.status(c.p1().node(), c.node())[field]
549 549 if fname is not None:
550 550 if fname in files:
551 551 return True
552 552 else:
553 553 for f in files:
554 554 if m(f):
555 555 return True
556 556
557 557 return subset.filter(matches)
558 558
559 559 def _children(repo, narrow, parentset):
560 560 cs = set()
561 561 if not parentset:
562 562 return baseset(cs)
563 563 pr = repo.changelog.parentrevs
564 564 minrev = min(parentset)
565 565 for r in narrow:
566 566 if r <= minrev:
567 567 continue
568 568 for p in pr(r):
569 569 if p in parentset:
570 570 cs.add(r)
571 571 return baseset(cs)
572 572
573 573 def children(repo, subset, x):
574 574 """``children(set)``
575 575 Child changesets of changesets in set.
576 576 """
577 577 s = getset(repo, baseset(repo), x).set()
578 578 cs = _children(repo, subset, s)
579 579 return subset & cs
580 580
581 581 def closed(repo, subset, x):
582 582 """``closed()``
583 583 Changeset is closed.
584 584 """
585 585 # i18n: "closed" is a keyword
586 586 getargs(x, 0, 0, _("closed takes no arguments"))
587 587 return subset.filter(lambda r: repo[r].closesbranch())
588 588
589 589 def contains(repo, subset, x):
590 590 """``contains(pattern)``
591 591 The revision's manifest contains a file matching pattern (but might not
592 592 modify it). See :hg:`help patterns` for information about file patterns.
593 593
594 594 The pattern without explicit kind like ``glob:`` is expected to be
595 595 relative to the current directory and match against a file exactly
596 596 for efficiency.
597 597 """
598 598 # i18n: "contains" is a keyword
599 599 pat = getstring(x, _("contains requires a pattern"))
600 600
601 601 def matches(x):
602 602 if not matchmod.patkind(pat):
603 603 pats = pathutil.canonpath(repo.root, repo.getcwd(), pat)
604 604 if pats in repo[x]:
605 605 return True
606 606 else:
607 607 c = repo[x]
608 608 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
609 609 for f in c.manifest():
610 610 if m(f):
611 611 return True
612 612 return False
613 613
614 614 return subset.filter(matches)
615 615
616 616 def converted(repo, subset, x):
617 617 """``converted([id])``
618 618 Changesets converted from the given identifier in the old repository if
619 619 present, or all converted changesets if no identifier is specified.
620 620 """
621 621
622 622 # There is exactly no chance of resolving the revision, so do a simple
623 623 # string compare and hope for the best
624 624
625 625 rev = None
626 626 # i18n: "converted" is a keyword
627 627 l = getargs(x, 0, 1, _('converted takes one or no arguments'))
628 628 if l:
629 629 # i18n: "converted" is a keyword
630 630 rev = getstring(l[0], _('converted requires a revision'))
631 631
632 632 def _matchvalue(r):
633 633 source = repo[r].extra().get('convert_revision', None)
634 634 return source is not None and (rev is None or source.startswith(rev))
635 635
636 636 return subset.filter(lambda r: _matchvalue(r))
637 637
638 638 def date(repo, subset, x):
639 639 """``date(interval)``
640 640 Changesets within the interval, see :hg:`help dates`.
641 641 """
642 642 # i18n: "date" is a keyword
643 643 ds = getstring(x, _("date requires a string"))
644 644 dm = util.matchdate(ds)
645 645 return subset.filter(lambda x: dm(repo[x].date()[0]))
646 646
647 647 def desc(repo, subset, x):
648 648 """``desc(string)``
649 649 Search commit message for string. The match is case-insensitive.
650 650 """
651 651 # i18n: "desc" is a keyword
652 652 ds = encoding.lower(getstring(x, _("desc requires a string")))
653 653
654 654 def matches(x):
655 655 c = repo[x]
656 656 return ds in encoding.lower(c.description())
657 657
658 658 return subset.filter(matches)
659 659
660 660 def _descendants(repo, subset, x, followfirst=False):
661 661 args = getset(repo, spanset(repo), x)
662 662 if not args:
663 663 return baseset([])
664 664 s = _revdescendants(repo, args, followfirst)
665 665
666 666 # Both sets need to be ascending in order to lazily return the union
667 667 # in the correct order.
668 668 args.ascending()
669 669
670 670 subsetset = subset.set()
671 671 result = (orderedlazyset(s, subsetset.__contains__, ascending=True) +
672 672 orderedlazyset(args, subsetset.__contains__, ascending=True))
673 673
674 674 # Wrap result in a lazyset since it's an _addset, which doesn't implement
675 675 # all the necessary functions to be consumed by callers.
676 676 return orderedlazyset(result, lambda r: True, ascending=True)
677 677
678 678 def descendants(repo, subset, x):
679 679 """``descendants(set)``
680 680 Changesets which are descendants of changesets in set.
681 681 """
682 682 return _descendants(repo, subset, x)
683 683
684 684 def _firstdescendants(repo, subset, x):
685 685 # ``_firstdescendants(set)``
686 686 # Like ``descendants(set)`` but follows only the first parents.
687 687 return _descendants(repo, subset, x, followfirst=True)
688 688
689 689 def destination(repo, subset, x):
690 690 """``destination([set])``
691 691 Changesets that were created by a graft, transplant or rebase operation,
692 692 with the given revisions specified as the source. Omitting the optional set
693 693 is the same as passing all().
694 694 """
695 695 if x is not None:
696 696 args = getset(repo, spanset(repo), x).set()
697 697 else:
698 698 args = getall(repo, spanset(repo), x).set()
699 699
700 700 dests = set()
701 701
702 702 # subset contains all of the possible destinations that can be returned, so
703 703 # iterate over them and see if their source(s) were provided in the args.
704 704 # Even if the immediate src of r is not in the args, src's source (or
705 705 # further back) may be. Scanning back further than the immediate src allows
706 706 # transitive transplants and rebases to yield the same results as transitive
707 707 # grafts.
708 708 for r in subset:
709 709 src = _getrevsource(repo, r)
710 710 lineage = None
711 711
712 712 while src is not None:
713 713 if lineage is None:
714 714 lineage = list()
715 715
716 716 lineage.append(r)
717 717
718 718 # The visited lineage is a match if the current source is in the arg
719 719 # set. Since every candidate dest is visited by way of iterating
720 720 # subset, any dests further back in the lineage will be tested by a
721 721 # different iteration over subset. Likewise, if the src was already
722 722 # selected, the current lineage can be selected without going back
723 723 # further.
724 724 if src in args or src in dests:
725 725 dests.update(lineage)
726 726 break
727 727
728 728 r = src
729 729 src = _getrevsource(repo, r)
730 730
731 731 return subset.filter(dests.__contains__)
732 732
733 733 def divergent(repo, subset, x):
734 734 """``divergent()``
735 735 Final successors of changesets with an alternative set of final successors.
736 736 """
737 737 # i18n: "divergent" is a keyword
738 738 getargs(x, 0, 0, _("divergent takes no arguments"))
739 739 divergent = obsmod.getrevs(repo, 'divergent')
740 740 return subset.filter(divergent.__contains__)
741 741
742 742 def draft(repo, subset, x):
743 743 """``draft()``
744 744 Changeset in draft phase."""
745 745 # i18n: "draft" is a keyword
746 746 getargs(x, 0, 0, _("draft takes no arguments"))
747 747 pc = repo._phasecache
748 748 return subset.filter(lambda r: pc.phase(repo, r) == phases.draft)
749 749
750 750 def extinct(repo, subset, x):
751 751 """``extinct()``
752 752 Obsolete changesets with obsolete descendants only.
753 753 """
754 754 # i18n: "extinct" is a keyword
755 755 getargs(x, 0, 0, _("extinct takes no arguments"))
756 756 extincts = obsmod.getrevs(repo, 'extinct')
757 757 return subset & extincts
758 758
759 759 def extra(repo, subset, x):
760 760 """``extra(label, [value])``
761 761 Changesets with the given label in the extra metadata, with the given
762 762 optional value.
763 763
764 764 If `value` starts with `re:`, the remainder of the value is treated as
765 765 a regular expression. To match a value that actually starts with `re:`,
766 766 use the prefix `literal:`.
767 767 """
768 768
769 769 # i18n: "extra" is a keyword
770 770 l = getargs(x, 1, 2, _('extra takes at least 1 and at most 2 arguments'))
771 771 # i18n: "extra" is a keyword
772 772 label = getstring(l[0], _('first argument to extra must be a string'))
773 773 value = None
774 774
775 775 if len(l) > 1:
776 776 # i18n: "extra" is a keyword
777 777 value = getstring(l[1], _('second argument to extra must be a string'))
778 778 kind, value, matcher = _stringmatcher(value)
779 779
780 780 def _matchvalue(r):
781 781 extra = repo[r].extra()
782 782 return label in extra and (value is None or matcher(extra[label]))
783 783
784 784 return subset.filter(lambda r: _matchvalue(r))
785 785
786 786 def filelog(repo, subset, x):
787 787 """``filelog(pattern)``
788 788 Changesets connected to the specified filelog.
789 789
790 790 For performance reasons, visits only revisions mentioned in the file-level
791 791 filelog, rather than filtering through all changesets (much faster, but
792 792 doesn't include deletes or duplicate changes). For a slower, more accurate
793 793 result, use ``file()``.
794 794
795 795 The pattern without explicit kind like ``glob:`` is expected to be
796 796 relative to the current directory and match against a file exactly
797 797 for efficiency.
798 798 """
799 799
800 800 # i18n: "filelog" is a keyword
801 801 pat = getstring(x, _("filelog requires a pattern"))
802 802 s = set()
803 803
804 804 if not matchmod.patkind(pat):
805 805 f = pathutil.canonpath(repo.root, repo.getcwd(), pat)
806 806 fl = repo.file(f)
807 807 for fr in fl:
808 808 s.add(fl.linkrev(fr))
809 809 else:
810 810 m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=repo[None])
811 811 for f in repo[None]:
812 812 if m(f):
813 813 fl = repo.file(f)
814 814 for fr in fl:
815 815 s.add(fl.linkrev(fr))
816 816
817 817 return subset.filter(s.__contains__)
818 818
819 819 def first(repo, subset, x):
820 820 """``first(set, [n])``
821 821 An alias for limit().
822 822 """
823 823 return limit(repo, subset, x)
824 824
825 825 def _follow(repo, subset, x, name, followfirst=False):
826 826 l = getargs(x, 0, 1, _("%s takes no arguments or a filename") % name)
827 827 c = repo['.']
828 828 if l:
829 829 x = getstring(l[0], _("%s expected a filename") % name)
830 830 if x in c:
831 831 cx = c[x]
832 832 s = set(ctx.rev() for ctx in cx.ancestors(followfirst=followfirst))
833 833 # include the revision responsible for the most recent version
834 834 s.add(cx.linkrev())
835 835 else:
836 836 return baseset([])
837 837 else:
838 838 s = _revancestors(repo, baseset([c.rev()]), followfirst)
839 839
840 840 return subset.filter(s.__contains__)
841 841
842 842 def follow(repo, subset, x):
843 843 """``follow([file])``
844 844 An alias for ``::.`` (ancestors of the working copy's first parent).
845 845 If a filename is specified, the history of the given file is followed,
846 846 including copies.
847 847 """
848 848 return _follow(repo, subset, x, 'follow')
849 849
850 850 def _followfirst(repo, subset, x):
851 851 # ``followfirst([file])``
852 852 # Like ``follow([file])`` but follows only the first parent of
853 853 # every revision or file revision.
854 854 return _follow(repo, subset, x, '_followfirst', followfirst=True)
855 855
856 856 def getall(repo, subset, x):
857 857 """``all()``
858 858 All changesets, the same as ``0:tip``.
859 859 """
860 860 # i18n: "all" is a keyword
861 861 getargs(x, 0, 0, _("all takes no arguments"))
862 862 return subset
863 863
864 864 def grep(repo, subset, x):
865 865 """``grep(regex)``
866 866 Like ``keyword(string)`` but accepts a regex. Use ``grep(r'...')``
867 867 to ensure special escape characters are handled correctly. Unlike
868 868 ``keyword(string)``, the match is case-sensitive.
869 869 """
870 870 try:
871 871 # i18n: "grep" is a keyword
872 872 gr = re.compile(getstring(x, _("grep requires a string")))
873 873 except re.error, e:
874 874 raise error.ParseError(_('invalid match pattern: %s') % e)
875 875
876 876 def matches(x):
877 877 c = repo[x]
878 878 for e in c.files() + [c.user(), c.description()]:
879 879 if gr.search(e):
880 880 return True
881 881 return False
882 882
883 883 return subset.filter(matches)
884 884
885 885 def _matchfiles(repo, subset, x):
886 886 # _matchfiles takes a revset list of prefixed arguments:
887 887 #
888 888 # [p:foo, i:bar, x:baz]
889 889 #
890 890 # builds a match object from them and filters subset. Allowed
891 891 # prefixes are 'p:' for regular patterns, 'i:' for include
892 892 # patterns and 'x:' for exclude patterns. Use 'r:' prefix to pass
893 893 # a revision identifier, or the empty string to reference the
894 894 # working directory, from which the match object is
895 895 # initialized. Use 'd:' to set the default matching mode, default
896 896 # to 'glob'. At most one 'r:' and 'd:' argument can be passed.
897 897
898 898 # i18n: "_matchfiles" is a keyword
899 899 l = getargs(x, 1, -1, _("_matchfiles requires at least one argument"))
900 900 pats, inc, exc = [], [], []
901 901 hasset = False
902 902 rev, default = None, None
903 903 for arg in l:
904 904 # i18n: "_matchfiles" is a keyword
905 905 s = getstring(arg, _("_matchfiles requires string arguments"))
906 906 prefix, value = s[:2], s[2:]
907 907 if prefix == 'p:':
908 908 pats.append(value)
909 909 elif prefix == 'i:':
910 910 inc.append(value)
911 911 elif prefix == 'x:':
912 912 exc.append(value)
913 913 elif prefix == 'r:':
914 914 if rev is not None:
915 915 # i18n: "_matchfiles" is a keyword
916 916 raise error.ParseError(_('_matchfiles expected at most one '
917 917 'revision'))
918 918 rev = value
919 919 elif prefix == 'd:':
920 920 if default is not None:
921 921 # i18n: "_matchfiles" is a keyword
922 922 raise error.ParseError(_('_matchfiles expected at most one '
923 923 'default mode'))
924 924 default = value
925 925 else:
926 926 # i18n: "_matchfiles" is a keyword
927 927 raise error.ParseError(_('invalid _matchfiles prefix: %s') % prefix)
928 928 if not hasset and matchmod.patkind(value) == 'set':
929 929 hasset = True
930 930 if not default:
931 931 default = 'glob'
932 932
933 933 def matches(x):
934 934 m = None
935 935 c = repo[x]
936 936 if not m or (hasset and rev is None):
937 937 ctx = c
938 938 if rev is not None:
939 939 ctx = repo[rev or None]
940 940 m = matchmod.match(repo.root, repo.getcwd(), pats, include=inc,
941 941 exclude=exc, ctx=ctx, default=default)
942 942 for f in c.files():
943 943 if m(f):
944 944 return True
945 945 return False
946 946
947 947 return subset.filter(matches)
948 948
949 949 def hasfile(repo, subset, x):
950 950 """``file(pattern)``
951 951 Changesets affecting files matched by pattern.
952 952
953 953 For a faster but less accurate result, consider using ``filelog()``
954 954 instead.
955 955
956 956 This predicate uses ``glob:`` as the default kind of pattern.
957 957 """
958 958 # i18n: "file" is a keyword
959 959 pat = getstring(x, _("file requires a pattern"))
960 960 return _matchfiles(repo, subset, ('string', 'p:' + pat))
961 961
962 962 def head(repo, subset, x):
963 963 """``head()``
964 964 Changeset is a named branch head.
965 965 """
966 966 # i18n: "head" is a keyword
967 967 getargs(x, 0, 0, _("head takes no arguments"))
968 968 hs = set()
969 969 for b, ls in repo.branchmap().iteritems():
970 970 hs.update(repo[h].rev() for h in ls)
971 971 return baseset(hs).filter(subset.__contains__)
972 972
973 973 def heads(repo, subset, x):
974 974 """``heads(set)``
975 975 Members of set with no children in set.
976 976 """
977 977 s = getset(repo, subset, x)
978 978 ps = parents(repo, subset, x)
979 979 return s - ps
980 980
981 981 def hidden(repo, subset, x):
982 982 """``hidden()``
983 983 Hidden changesets.
984 984 """
985 985 # i18n: "hidden" is a keyword
986 986 getargs(x, 0, 0, _("hidden takes no arguments"))
987 987 hiddenrevs = repoview.filterrevs(repo, 'visible')
988 988 return subset & hiddenrevs
989 989
990 990 def keyword(repo, subset, x):
991 991 """``keyword(string)``
992 992 Search commit message, user name, and names of changed files for
993 993 string. The match is case-insensitive.
994 994 """
995 995 # i18n: "keyword" is a keyword
996 996 kw = encoding.lower(getstring(x, _("keyword requires a string")))
997 997
998 998 def matches(r):
999 999 c = repo[r]
1000 1000 return util.any(kw in encoding.lower(t) for t in c.files() + [c.user(),
1001 1001 c.description()])
1002 1002
1003 1003 return subset.filter(matches)
1004 1004
1005 1005 def limit(repo, subset, x):
1006 1006 """``limit(set, [n])``
1007 1007 First n members of set, defaulting to 1.
1008 1008 """
1009 1009 # i18n: "limit" is a keyword
1010 1010 l = getargs(x, 1, 2, _("limit requires one or two arguments"))
1011 1011 try:
1012 1012 lim = 1
1013 1013 if len(l) == 2:
1014 1014 # i18n: "limit" is a keyword
1015 1015 lim = int(getstring(l[1], _("limit requires a number")))
1016 1016 except (TypeError, ValueError):
1017 1017 # i18n: "limit" is a keyword
1018 1018 raise error.ParseError(_("limit expects a number"))
1019 1019 ss = subset.set()
1020 1020 os = getset(repo, spanset(repo), l[0])
1021 1021 bs = baseset([])
1022 1022 it = iter(os)
1023 1023 for x in xrange(lim):
1024 1024 try:
1025 1025 y = it.next()
1026 1026 if y in ss:
1027 1027 bs.append(y)
1028 1028 except (StopIteration):
1029 1029 break
1030 1030 return bs
1031 1031
1032 1032 def last(repo, subset, x):
1033 1033 """``last(set, [n])``
1034 1034 Last n members of set, defaulting to 1.
1035 1035 """
1036 1036 # i18n: "last" is a keyword
1037 1037 l = getargs(x, 1, 2, _("last requires one or two arguments"))
1038 1038 try:
1039 1039 lim = 1
1040 1040 if len(l) == 2:
1041 1041 # i18n: "last" is a keyword
1042 1042 lim = int(getstring(l[1], _("last requires a number")))
1043 1043 except (TypeError, ValueError):
1044 1044 # i18n: "last" is a keyword
1045 1045 raise error.ParseError(_("last expects a number"))
1046 1046 ss = subset.set()
1047 1047 os = getset(repo, spanset(repo), l[0])
1048 1048 os.reverse()
1049 1049 bs = baseset([])
1050 1050 it = iter(os)
1051 1051 for x in xrange(lim):
1052 1052 try:
1053 1053 y = it.next()
1054 1054 if y in ss:
1055 1055 bs.append(y)
1056 1056 except (StopIteration):
1057 1057 break
1058 1058 return bs
1059 1059
1060 1060 def maxrev(repo, subset, x):
1061 1061 """``max(set)``
1062 1062 Changeset with highest revision number in set.
1063 1063 """
1064 1064 os = getset(repo, spanset(repo), x)
1065 1065 if os:
1066 1066 m = os.max()
1067 1067 if m in subset:
1068 1068 return baseset([m])
1069 1069 return baseset([])
1070 1070
1071 1071 def merge(repo, subset, x):
1072 1072 """``merge()``
1073 1073 Changeset is a merge changeset.
1074 1074 """
1075 1075 # i18n: "merge" is a keyword
1076 1076 getargs(x, 0, 0, _("merge takes no arguments"))
1077 1077 cl = repo.changelog
1078 1078 return subset.filter(lambda r: cl.parentrevs(r)[1] != -1)
1079 1079
1080 1080 def branchpoint(repo, subset, x):
1081 1081 """``branchpoint()``
1082 1082 Changesets with more than one child.
1083 1083 """
1084 1084 # i18n: "branchpoint" is a keyword
1085 1085 getargs(x, 0, 0, _("branchpoint takes no arguments"))
1086 1086 cl = repo.changelog
1087 1087 if not subset:
1088 1088 return baseset([])
1089 1089 baserev = min(subset)
1090 1090 parentscount = [0]*(len(repo) - baserev)
1091 1091 for r in cl.revs(start=baserev + 1):
1092 1092 for p in cl.parentrevs(r):
1093 1093 if p >= baserev:
1094 1094 parentscount[p - baserev] += 1
1095 1095 return subset.filter(lambda r: parentscount[r - baserev] > 1)
1096 1096
1097 1097 def minrev(repo, subset, x):
1098 1098 """``min(set)``
1099 1099 Changeset with lowest revision number in set.
1100 1100 """
1101 1101 os = getset(repo, spanset(repo), x)
1102 1102 if os:
1103 1103 m = os.min()
1104 1104 if m in subset:
1105 1105 return baseset([m])
1106 1106 return baseset([])
1107 1107
1108 1108 def modifies(repo, subset, x):
1109 1109 """``modifies(pattern)``
1110 1110 Changesets modifying files matched by pattern.
1111 1111
1112 1112 The pattern without explicit kind like ``glob:`` is expected to be
1113 1113 relative to the current directory and match against a file or a
1114 1114 directory.
1115 1115 """
1116 1116 # i18n: "modifies" is a keyword
1117 1117 pat = getstring(x, _("modifies requires a pattern"))
1118 1118 return checkstatus(repo, subset, pat, 0)
1119 1119
1120 1120 def node_(repo, subset, x):
1121 1121 """``id(string)``
1122 1122 Revision non-ambiguously specified by the given hex string prefix.
1123 1123 """
1124 1124 # i18n: "id" is a keyword
1125 1125 l = getargs(x, 1, 1, _("id requires one argument"))
1126 1126 # i18n: "id" is a keyword
1127 1127 n = getstring(l[0], _("id requires a string"))
1128 1128 if len(n) == 40:
1129 1129 rn = repo[n].rev()
1130 1130 else:
1131 1131 rn = None
1132 1132 pm = repo.changelog._partialmatch(n)
1133 1133 if pm is not None:
1134 1134 rn = repo.changelog.rev(pm)
1135 1135
1136 1136 return subset.filter(lambda r: r == rn)
1137 1137
1138 1138 def obsolete(repo, subset, x):
1139 1139 """``obsolete()``
1140 1140 Mutable changeset with a newer version."""
1141 1141 # i18n: "obsolete" is a keyword
1142 1142 getargs(x, 0, 0, _("obsolete takes no arguments"))
1143 1143 obsoletes = obsmod.getrevs(repo, 'obsolete')
1144 1144 return subset & obsoletes
1145 1145
1146 1146 def origin(repo, subset, x):
1147 1147 """``origin([set])``
1148 1148 Changesets that were specified as a source for the grafts, transplants or
1149 1149 rebases that created the given revisions. Omitting the optional set is the
1150 1150 same as passing all(). If a changeset created by these operations is itself
1151 1151 specified as a source for one of these operations, only the source changeset
1152 1152 for the first operation is selected.
1153 1153 """
1154 1154 if x is not None:
1155 1155 args = getset(repo, spanset(repo), x).set()
1156 1156 else:
1157 1157 args = getall(repo, spanset(repo), x).set()
1158 1158
1159 1159 def _firstsrc(rev):
1160 1160 src = _getrevsource(repo, rev)
1161 1161 if src is None:
1162 1162 return None
1163 1163
1164 1164 while True:
1165 1165 prev = _getrevsource(repo, src)
1166 1166
1167 1167 if prev is None:
1168 1168 return src
1169 1169 src = prev
1170 1170
1171 1171 o = set([_firstsrc(r) for r in args])
1172 1172 return subset.filter(o.__contains__)
1173 1173
1174 1174 def outgoing(repo, subset, x):
1175 1175 """``outgoing([path])``
1176 1176 Changesets not found in the specified destination repository, or the
1177 1177 default push location.
1178 1178 """
1179 1179 import hg # avoid start-up nasties
1180 1180 # i18n: "outgoing" is a keyword
1181 1181 l = getargs(x, 0, 1, _("outgoing takes one or no arguments"))
1182 1182 # i18n: "outgoing" is a keyword
1183 1183 dest = l and getstring(l[0], _("outgoing requires a repository path")) or ''
1184 1184 dest = repo.ui.expandpath(dest or 'default-push', dest or 'default')
1185 1185 dest, branches = hg.parseurl(dest)
1186 1186 revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
1187 1187 if revs:
1188 1188 revs = [repo.lookup(rev) for rev in revs]
1189 1189 other = hg.peer(repo, {}, dest)
1190 1190 repo.ui.pushbuffer()
1191 1191 outgoing = discovery.findcommonoutgoing(repo, other, onlyheads=revs)
1192 1192 repo.ui.popbuffer()
1193 1193 cl = repo.changelog
1194 1194 o = set([cl.rev(r) for r in outgoing.missing])
1195 1195 return subset.filter(o.__contains__)
1196 1196
1197 1197 def p1(repo, subset, x):
1198 1198 """``p1([set])``
1199 1199 First parent of changesets in set, or the working directory.
1200 1200 """
1201 1201 if x is None:
1202 1202 p = repo[x].p1().rev()
1203 1203 return subset.filter(lambda r: r == p)
1204 1204
1205 1205 ps = set()
1206 1206 cl = repo.changelog
1207 1207 for r in getset(repo, spanset(repo), x):
1208 1208 ps.add(cl.parentrevs(r)[0])
1209 1209 return subset & ps
1210 1210
1211 1211 def p2(repo, subset, x):
1212 1212 """``p2([set])``
1213 1213 Second parent of changesets in set, or the working directory.
1214 1214 """
1215 1215 if x is None:
1216 1216 ps = repo[x].parents()
1217 1217 try:
1218 1218 p = ps[1].rev()
1219 1219 return subset.filter(lambda r: r == p)
1220 1220 except IndexError:
1221 1221 return baseset([])
1222 1222
1223 1223 ps = set()
1224 1224 cl = repo.changelog
1225 1225 for r in getset(repo, spanset(repo), x):
1226 1226 ps.add(cl.parentrevs(r)[1])
1227 1227 return subset & ps
1228 1228
1229 1229 def parents(repo, subset, x):
1230 1230 """``parents([set])``
1231 1231 The set of all parents for all changesets in set, or the working directory.
1232 1232 """
1233 1233 if x is None:
1234 1234 ps = tuple(p.rev() for p in repo[x].parents())
1235 1235 return subset & ps
1236 1236
1237 1237 ps = set()
1238 1238 cl = repo.changelog
1239 1239 for r in getset(repo, spanset(repo), x):
1240 1240 ps.update(cl.parentrevs(r))
1241 1241 return subset & ps
1242 1242
1243 1243 def parentspec(repo, subset, x, n):
1244 1244 """``set^0``
1245 1245 The set.
1246 1246 ``set^1`` (or ``set^``), ``set^2``
1247 1247 First or second parent, respectively, of all changesets in set.
1248 1248 """
1249 1249 try:
1250 1250 n = int(n[1])
1251 1251 if n not in (0, 1, 2):
1252 1252 raise ValueError
1253 1253 except (TypeError, ValueError):
1254 1254 raise error.ParseError(_("^ expects a number 0, 1, or 2"))
1255 1255 ps = set()
1256 1256 cl = repo.changelog
1257 1257 for r in getset(repo, baseset(cl), x):
1258 1258 if n == 0:
1259 1259 ps.add(r)
1260 1260 elif n == 1:
1261 1261 ps.add(cl.parentrevs(r)[0])
1262 1262 elif n == 2:
1263 1263 parents = cl.parentrevs(r)
1264 1264 if len(parents) > 1:
1265 1265 ps.add(parents[1])
1266 1266 return subset & ps
1267 1267
1268 1268 def present(repo, subset, x):
1269 1269 """``present(set)``
1270 1270 An empty set, if any revision in set isn't found; otherwise,
1271 1271 all revisions in set.
1272 1272
1273 1273 If any of specified revisions is not present in the local repository,
1274 1274 the query is normally aborted. But this predicate allows the query
1275 1275 to continue even in such cases.
1276 1276 """
1277 1277 try:
1278 1278 return getset(repo, subset, x)
1279 1279 except error.RepoLookupError:
1280 1280 return baseset([])
1281 1281
1282 1282 def public(repo, subset, x):
1283 1283 """``public()``
1284 1284 Changeset in public phase."""
1285 1285 # i18n: "public" is a keyword
1286 1286 getargs(x, 0, 0, _("public takes no arguments"))
1287 1287 pc = repo._phasecache
1288 1288 return subset.filter(lambda r: pc.phase(repo, r) == phases.public)
1289 1289
1290 1290 def remote(repo, subset, x):
1291 1291 """``remote([id [,path]])``
1292 1292 Local revision that corresponds to the given identifier in a
1293 1293 remote repository, if present. Here, the '.' identifier is a
1294 1294 synonym for the current local branch.
1295 1295 """
1296 1296
1297 1297 import hg # avoid start-up nasties
1298 1298 # i18n: "remote" is a keyword
1299 1299 l = getargs(x, 0, 2, _("remote takes one, two or no arguments"))
1300 1300
1301 1301 q = '.'
1302 1302 if len(l) > 0:
1303 1303 # i18n: "remote" is a keyword
1304 1304 q = getstring(l[0], _("remote requires a string id"))
1305 1305 if q == '.':
1306 1306 q = repo['.'].branch()
1307 1307
1308 1308 dest = ''
1309 1309 if len(l) > 1:
1310 1310 # i18n: "remote" is a keyword
1311 1311 dest = getstring(l[1], _("remote requires a repository path"))
1312 1312 dest = repo.ui.expandpath(dest or 'default')
1313 1313 dest, branches = hg.parseurl(dest)
1314 1314 revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
1315 1315 if revs:
1316 1316 revs = [repo.lookup(rev) for rev in revs]
1317 1317 other = hg.peer(repo, {}, dest)
1318 1318 n = other.lookup(q)
1319 1319 if n in repo:
1320 1320 r = repo[n].rev()
1321 1321 if r in subset:
1322 1322 return baseset([r])
1323 1323 return baseset([])
1324 1324
1325 1325 def removes(repo, subset, x):
1326 1326 """``removes(pattern)``
1327 1327 Changesets which remove files matching pattern.
1328 1328
1329 1329 The pattern without explicit kind like ``glob:`` is expected to be
1330 1330 relative to the current directory and match against a file or a
1331 1331 directory.
1332 1332 """
1333 1333 # i18n: "removes" is a keyword
1334 1334 pat = getstring(x, _("removes requires a pattern"))
1335 1335 return checkstatus(repo, subset, pat, 2)
1336 1336
1337 1337 def rev(repo, subset, x):
1338 1338 """``rev(number)``
1339 1339 Revision with the given numeric identifier.
1340 1340 """
1341 1341 # i18n: "rev" is a keyword
1342 1342 l = getargs(x, 1, 1, _("rev requires one argument"))
1343 1343 try:
1344 1344 # i18n: "rev" is a keyword
1345 1345 l = int(getstring(l[0], _("rev requires a number")))
1346 1346 except (TypeError, ValueError):
1347 1347 # i18n: "rev" is a keyword
1348 1348 raise error.ParseError(_("rev expects a number"))
1349 1349 return subset.filter(lambda r: r == l)
1350 1350
1351 1351 def matching(repo, subset, x):
1352 1352 """``matching(revision [, field])``
1353 1353 Changesets in which a given set of fields match the set of fields in the
1354 1354 selected revision or set.
1355 1355
1356 1356 To match more than one field pass the list of fields to match separated
1357 1357 by spaces (e.g. ``author description``).
1358 1358
1359 1359 Valid fields are most regular revision fields and some special fields.
1360 1360
1361 1361 Regular revision fields are ``description``, ``author``, ``branch``,
1362 1362 ``date``, ``files``, ``phase``, ``parents``, ``substate``, ``user``
1363 1363 and ``diff``.
1364 1364 Note that ``author`` and ``user`` are synonyms. ``diff`` refers to the
1365 1365 contents of the revision. Two revisions matching their ``diff`` will
1366 1366 also match their ``files``.
1367 1367
1368 1368 Special fields are ``summary`` and ``metadata``:
1369 1369 ``summary`` matches the first line of the description.
1370 1370 ``metadata`` is equivalent to matching ``description user date``
1371 1371 (i.e. it matches the main metadata fields).
1372 1372
1373 1373 ``metadata`` is the default field which is used when no fields are
1374 1374 specified. You can match more than one field at a time.
1375 1375 """
1376 1376 # i18n: "matching" is a keyword
1377 1377 l = getargs(x, 1, 2, _("matching takes 1 or 2 arguments"))
1378 1378
1379 1379 revs = getset(repo, baseset(repo.changelog), l[0])
1380 1380
1381 1381 fieldlist = ['metadata']
1382 1382 if len(l) > 1:
1383 1383 fieldlist = getstring(l[1],
1384 1384 # i18n: "matching" is a keyword
1385 1385 _("matching requires a string "
1386 1386 "as its second argument")).split()
1387 1387
1388 1388 # Make sure that there are no repeated fields,
1389 1389 # expand the 'special' 'metadata' field type
1390 1390 # and check the 'files' whenever we check the 'diff'
1391 1391 fields = []
1392 1392 for field in fieldlist:
1393 1393 if field == 'metadata':
1394 1394 fields += ['user', 'description', 'date']
1395 1395 elif field == 'diff':
1396 1396 # a revision matching the diff must also match the files
1397 1397 # since matching the diff is very costly, make sure to
1398 1398 # also match the files first
1399 1399 fields += ['files', 'diff']
1400 1400 else:
1401 1401 if field == 'author':
1402 1402 field = 'user'
1403 1403 fields.append(field)
1404 1404 fields = set(fields)
1405 1405 if 'summary' in fields and 'description' in fields:
1406 1406 # If a revision matches its description it also matches its summary
1407 1407 fields.discard('summary')
1408 1408
1409 1409 # We may want to match more than one field
1410 1410 # Not all fields take the same amount of time to be matched
1411 1411 # Sort the selected fields in order of increasing matching cost
1412 1412 fieldorder = ['phase', 'parents', 'user', 'date', 'branch', 'summary',
1413 1413 'files', 'description', 'substate', 'diff']
1414 1414 def fieldkeyfunc(f):
1415 1415 try:
1416 1416 return fieldorder.index(f)
1417 1417 except ValueError:
1418 1418 # assume an unknown field is very costly
1419 1419 return len(fieldorder)
1420 1420 fields = list(fields)
1421 1421 fields.sort(key=fieldkeyfunc)
1422 1422
1423 1423 # Each field will be matched with its own "getfield" function
1424 1424 # which will be added to the getfieldfuncs array of functions
1425 1425 getfieldfuncs = []
1426 1426 _funcs = {
1427 1427 'user': lambda r: repo[r].user(),
1428 1428 'branch': lambda r: repo[r].branch(),
1429 1429 'date': lambda r: repo[r].date(),
1430 1430 'description': lambda r: repo[r].description(),
1431 1431 'files': lambda r: repo[r].files(),
1432 1432 'parents': lambda r: repo[r].parents(),
1433 1433 'phase': lambda r: repo[r].phase(),
1434 1434 'substate': lambda r: repo[r].substate,
1435 1435 'summary': lambda r: repo[r].description().splitlines()[0],
1436 1436 'diff': lambda r: list(repo[r].diff(git=True),)
1437 1437 }
1438 1438 for info in fields:
1439 1439 getfield = _funcs.get(info, None)
1440 1440 if getfield is None:
1441 1441 raise error.ParseError(
1442 1442 # i18n: "matching" is a keyword
1443 1443 _("unexpected field name passed to matching: %s") % info)
1444 1444 getfieldfuncs.append(getfield)
1445 1445 # convert the getfield array of functions into a "getinfo" function
1446 1446 # which returns an array of field values (or a single value if there
1447 1447 # is only one field to match)
1448 1448 getinfo = lambda r: [f(r) for f in getfieldfuncs]
1449 1449
1450 1450 def matches(x):
1451 1451 for rev in revs:
1452 1452 target = getinfo(rev)
1453 1453 match = True
1454 1454 for n, f in enumerate(getfieldfuncs):
1455 1455 if target[n] != f(x):
1456 1456 match = False
1457 1457 if match:
1458 1458 return True
1459 1459 return False
1460 1460
1461 1461 return subset.filter(matches)
1462 1462
1463 1463 def reverse(repo, subset, x):
1464 1464 """``reverse(set)``
1465 1465 Reverse order of set.
1466 1466 """
1467 1467 l = getset(repo, subset, x)
1468 1468 l.reverse()
1469 1469 return l
1470 1470
1471 1471 def roots(repo, subset, x):
1472 1472 """``roots(set)``
1473 1473 Changesets in set with no parent changeset in set.
1474 1474 """
1475 1475 s = getset(repo, spanset(repo), x).set()
1476 1476 subset = baseset([r for r in s if r in subset.set()])
1477 1477 cs = _children(repo, subset, s)
1478 1478 return subset - cs
1479 1479
1480 1480 def secret(repo, subset, x):
1481 1481 """``secret()``
1482 1482 Changeset in secret phase."""
1483 1483 # i18n: "secret" is a keyword
1484 1484 getargs(x, 0, 0, _("secret takes no arguments"))
1485 1485 pc = repo._phasecache
1486 1486 return subset.filter(lambda x: pc.phase(repo, x) == phases.secret)
1487 1487
1488 1488 def sort(repo, subset, x):
1489 1489 """``sort(set[, [-]key...])``
1490 1490 Sort set by keys. The default sort order is ascending, specify a key
1491 1491 as ``-key`` to sort in descending order.
1492 1492
1493 1493 The keys can be:
1494 1494
1495 1495 - ``rev`` for the revision number,
1496 1496 - ``branch`` for the branch name,
1497 1497 - ``desc`` for the commit message (description),
1498 1498 - ``user`` for user name (``author`` can be used as an alias),
1499 1499 - ``date`` for the commit date
1500 1500 """
1501 1501 # i18n: "sort" is a keyword
1502 1502 l = getargs(x, 1, 2, _("sort requires one or two arguments"))
1503 1503 keys = "rev"
1504 1504 if len(l) == 2:
1505 1505 # i18n: "sort" is a keyword
1506 1506 keys = getstring(l[1], _("sort spec must be a string"))
1507 1507
1508 1508 s = l[0]
1509 1509 keys = keys.split()
1510 1510 l = []
1511 1511 def invert(s):
1512 1512 return "".join(chr(255 - ord(c)) for c in s)
1513 1513 revs = getset(repo, subset, s)
1514 1514 if keys == ["rev"]:
1515 1515 revs.sort()
1516 1516 return revs
1517 1517 elif keys == ["-rev"]:
1518 1518 revs.sort(reverse=True)
1519 1519 return revs
1520 1520 for r in revs:
1521 1521 c = repo[r]
1522 1522 e = []
1523 1523 for k in keys:
1524 1524 if k == 'rev':
1525 1525 e.append(r)
1526 1526 elif k == '-rev':
1527 1527 e.append(-r)
1528 1528 elif k == 'branch':
1529 1529 e.append(c.branch())
1530 1530 elif k == '-branch':
1531 1531 e.append(invert(c.branch()))
1532 1532 elif k == 'desc':
1533 1533 e.append(c.description())
1534 1534 elif k == '-desc':
1535 1535 e.append(invert(c.description()))
1536 1536 elif k in 'user author':
1537 1537 e.append(c.user())
1538 1538 elif k in '-user -author':
1539 1539 e.append(invert(c.user()))
1540 1540 elif k == 'date':
1541 1541 e.append(c.date()[0])
1542 1542 elif k == '-date':
1543 1543 e.append(-c.date()[0])
1544 1544 else:
1545 1545 raise error.ParseError(_("unknown sort key %r") % k)
1546 1546 e.append(r)
1547 1547 l.append(e)
1548 1548 l.sort()
1549 1549 return baseset([e[-1] for e in l])
1550 1550
1551 1551 def _stringmatcher(pattern):
1552 1552 """
1553 1553 accepts a string, possibly starting with 're:' or 'literal:' prefix.
1554 1554 returns the matcher name, pattern, and matcher function.
1555 1555 missing or unknown prefixes are treated as literal matches.
1556 1556
1557 1557 helper for tests:
1558 1558 >>> def test(pattern, *tests):
1559 1559 ... kind, pattern, matcher = _stringmatcher(pattern)
1560 1560 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
1561 1561
1562 1562 exact matching (no prefix):
1563 1563 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
1564 1564 ('literal', 'abcdefg', [False, False, True])
1565 1565
1566 1566 regex matching ('re:' prefix)
1567 1567 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
1568 1568 ('re', 'a.+b', [False, False, True])
1569 1569
1570 1570 force exact matches ('literal:' prefix)
1571 1571 >>> test('literal:re:foobar', 'foobar', 're:foobar')
1572 1572 ('literal', 're:foobar', [False, True])
1573 1573
1574 1574 unknown prefixes are ignored and treated as literals
1575 1575 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
1576 1576 ('literal', 'foo:bar', [False, False, True])
1577 1577 """
1578 1578 if pattern.startswith('re:'):
1579 1579 pattern = pattern[3:]
1580 1580 try:
1581 1581 regex = re.compile(pattern)
1582 1582 except re.error, e:
1583 1583 raise error.ParseError(_('invalid regular expression: %s')
1584 1584 % e)
1585 1585 return 're', pattern, regex.search
1586 1586 elif pattern.startswith('literal:'):
1587 1587 pattern = pattern[8:]
1588 1588 return 'literal', pattern, pattern.__eq__
1589 1589
1590 1590 def _substringmatcher(pattern):
1591 1591 kind, pattern, matcher = _stringmatcher(pattern)
1592 1592 if kind == 'literal':
1593 1593 matcher = lambda s: pattern in s
1594 1594 return kind, pattern, matcher
1595 1595
1596 1596 def tag(repo, subset, x):
1597 1597 """``tag([name])``
1598 1598 The specified tag by name, or all tagged revisions if no name is given.
1599 1599
1600 1600 If `name` starts with `re:`, the remainder of the name is treated as
1601 1601 a regular expression. To match a tag that actually starts with `re:`,
1602 1602 use the prefix `literal:`.
1603 1603 """
1604 1604 # i18n: "tag" is a keyword
1605 1605 args = getargs(x, 0, 1, _("tag takes one or no arguments"))
1606 1606 cl = repo.changelog
1607 1607 if args:
1608 1608 pattern = getstring(args[0],
1609 1609 # i18n: "tag" is a keyword
1610 1610 _('the argument to tag must be a string'))
1611 1611 kind, pattern, matcher = _stringmatcher(pattern)
1612 1612 if kind == 'literal':
1613 1613 # avoid resolving all tags
1614 1614 tn = repo._tagscache.tags.get(pattern, None)
1615 1615 if tn is None:
1616 1616 raise util.Abort(_("tag '%s' does not exist") % pattern)
1617 1617 s = set([repo[tn].rev()])
1618 1618 else:
1619 1619 s = set([cl.rev(n) for t, n in repo.tagslist() if matcher(t)])
1620 1620 else:
1621 1621 s = set([cl.rev(n) for t, n in repo.tagslist() if t != 'tip'])
1622 1622 return subset & s
1623 1623
1624 1624 def tagged(repo, subset, x):
1625 1625 return tag(repo, subset, x)
1626 1626
1627 1627 def unstable(repo, subset, x):
1628 1628 """``unstable()``
1629 1629 Non-obsolete changesets with obsolete ancestors.
1630 1630 """
1631 1631 # i18n: "unstable" is a keyword
1632 1632 getargs(x, 0, 0, _("unstable takes no arguments"))
1633 1633 unstables = obsmod.getrevs(repo, 'unstable')
1634 1634 return subset & unstables
1635 1635
1636 1636
1637 1637 def user(repo, subset, x):
1638 1638 """``user(string)``
1639 1639 User name contains string. The match is case-insensitive.
1640 1640
1641 1641 If `string` starts with `re:`, the remainder of the string is treated as
1642 1642 a regular expression. To match a user that actually contains `re:`, use
1643 1643 the prefix `literal:`.
1644 1644 """
1645 1645 return author(repo, subset, x)
1646 1646
1647 1647 # for internal use
1648 1648 def _list(repo, subset, x):
1649 1649 s = getstring(x, "internal error")
1650 1650 if not s:
1651 1651 return baseset([])
1652 1652 ls = [repo[r].rev() for r in s.split('\0')]
1653 1653 s = subset.set()
1654 1654 return baseset([r for r in ls if r in s])
1655 1655
1656 1656 # for internal use
1657 1657 def _intlist(repo, subset, x):
1658 1658 s = getstring(x, "internal error")
1659 1659 if not s:
1660 1660 return baseset([])
1661 1661 ls = [int(r) for r in s.split('\0')]
1662 1662 s = subset.set()
1663 1663 return baseset([r for r in ls if r in s])
1664 1664
1665 1665 # for internal use
1666 1666 def _hexlist(repo, subset, x):
1667 1667 s = getstring(x, "internal error")
1668 1668 if not s:
1669 1669 return baseset([])
1670 1670 cl = repo.changelog
1671 1671 ls = [cl.rev(node.bin(r)) for r in s.split('\0')]
1672 1672 s = subset.set()
1673 1673 return baseset([r for r in ls if r in s])
1674 1674
1675 1675 symbols = {
1676 1676 "adds": adds,
1677 1677 "all": getall,
1678 1678 "ancestor": ancestor,
1679 1679 "ancestors": ancestors,
1680 1680 "_firstancestors": _firstancestors,
1681 1681 "author": author,
1682 1682 "only": only,
1683 1683 "bisect": bisect,
1684 1684 "bisected": bisected,
1685 1685 "bookmark": bookmark,
1686 1686 "branch": branch,
1687 1687 "branchpoint": branchpoint,
1688 1688 "bumped": bumped,
1689 1689 "bundle": bundle,
1690 1690 "children": children,
1691 1691 "closed": closed,
1692 1692 "contains": contains,
1693 1693 "converted": converted,
1694 1694 "date": date,
1695 1695 "desc": desc,
1696 1696 "descendants": descendants,
1697 1697 "_firstdescendants": _firstdescendants,
1698 1698 "destination": destination,
1699 1699 "divergent": divergent,
1700 1700 "draft": draft,
1701 1701 "extinct": extinct,
1702 1702 "extra": extra,
1703 1703 "file": hasfile,
1704 1704 "filelog": filelog,
1705 1705 "first": first,
1706 1706 "follow": follow,
1707 1707 "_followfirst": _followfirst,
1708 1708 "grep": grep,
1709 1709 "head": head,
1710 1710 "heads": heads,
1711 1711 "hidden": hidden,
1712 1712 "id": node_,
1713 1713 "keyword": keyword,
1714 1714 "last": last,
1715 1715 "limit": limit,
1716 1716 "_matchfiles": _matchfiles,
1717 1717 "max": maxrev,
1718 1718 "merge": merge,
1719 1719 "min": minrev,
1720 1720 "modifies": modifies,
1721 1721 "obsolete": obsolete,
1722 1722 "origin": origin,
1723 1723 "outgoing": outgoing,
1724 1724 "p1": p1,
1725 1725 "p2": p2,
1726 1726 "parents": parents,
1727 1727 "present": present,
1728 1728 "public": public,
1729 1729 "remote": remote,
1730 1730 "removes": removes,
1731 1731 "rev": rev,
1732 1732 "reverse": reverse,
1733 1733 "roots": roots,
1734 1734 "sort": sort,
1735 1735 "secret": secret,
1736 1736 "matching": matching,
1737 1737 "tag": tag,
1738 1738 "tagged": tagged,
1739 1739 "user": user,
1740 1740 "unstable": unstable,
1741 1741 "_list": _list,
1742 1742 "_intlist": _intlist,
1743 1743 "_hexlist": _hexlist,
1744 1744 }
1745 1745
1746 1746 # symbols which can't be used for a DoS attack for any given input
1747 1747 # (e.g. those which accept regexes as plain strings shouldn't be included)
1748 1748 # functions that just return a lot of changesets (like all) don't count here
1749 1749 safesymbols = set([
1750 1750 "adds",
1751 1751 "all",
1752 1752 "ancestor",
1753 1753 "ancestors",
1754 1754 "_firstancestors",
1755 1755 "author",
1756 1756 "bisect",
1757 1757 "bisected",
1758 1758 "bookmark",
1759 1759 "branch",
1760 1760 "branchpoint",
1761 1761 "bumped",
1762 1762 "bundle",
1763 1763 "children",
1764 1764 "closed",
1765 1765 "converted",
1766 1766 "date",
1767 1767 "desc",
1768 1768 "descendants",
1769 1769 "_firstdescendants",
1770 1770 "destination",
1771 1771 "divergent",
1772 1772 "draft",
1773 1773 "extinct",
1774 1774 "extra",
1775 1775 "file",
1776 1776 "filelog",
1777 1777 "first",
1778 1778 "follow",
1779 1779 "_followfirst",
1780 1780 "head",
1781 1781 "heads",
1782 1782 "hidden",
1783 1783 "id",
1784 1784 "keyword",
1785 1785 "last",
1786 1786 "limit",
1787 1787 "_matchfiles",
1788 1788 "max",
1789 1789 "merge",
1790 1790 "min",
1791 1791 "modifies",
1792 1792 "obsolete",
1793 1793 "origin",
1794 1794 "outgoing",
1795 1795 "p1",
1796 1796 "p2",
1797 1797 "parents",
1798 1798 "present",
1799 1799 "public",
1800 1800 "remote",
1801 1801 "removes",
1802 1802 "rev",
1803 1803 "reverse",
1804 1804 "roots",
1805 1805 "sort",
1806 1806 "secret",
1807 1807 "matching",
1808 1808 "tag",
1809 1809 "tagged",
1810 1810 "user",
1811 1811 "unstable",
1812 1812 "_list",
1813 1813 "_intlist",
1814 1814 "_hexlist",
1815 1815 ])
1816 1816
1817 1817 methods = {
1818 1818 "range": rangeset,
1819 1819 "dagrange": dagrange,
1820 1820 "string": stringset,
1821 1821 "symbol": symbolset,
1822 1822 "and": andset,
1823 1823 "or": orset,
1824 1824 "not": notset,
1825 1825 "list": listset,
1826 1826 "func": func,
1827 1827 "ancestor": ancestorspec,
1828 1828 "parent": parentspec,
1829 1829 "parentpost": p1,
1830 1830 }
1831 1831
1832 1832 def optimize(x, small):
1833 1833 if x is None:
1834 1834 return 0, x
1835 1835
1836 1836 smallbonus = 1
1837 1837 if small:
1838 1838 smallbonus = .5
1839 1839
1840 1840 op = x[0]
1841 1841 if op == 'minus':
1842 1842 return optimize(('and', x[1], ('not', x[2])), small)
1843 1843 elif op == 'dagrangepre':
1844 1844 return optimize(('func', ('symbol', 'ancestors'), x[1]), small)
1845 1845 elif op == 'dagrangepost':
1846 1846 return optimize(('func', ('symbol', 'descendants'), x[1]), small)
1847 1847 elif op == 'rangepre':
1848 1848 return optimize(('range', ('string', '0'), x[1]), small)
1849 1849 elif op == 'rangepost':
1850 1850 return optimize(('range', x[1], ('string', 'tip')), small)
1851 1851 elif op == 'negate':
1852 1852 return optimize(('string',
1853 1853 '-' + getstring(x[1], _("can't negate that"))), small)
1854 1854 elif op in 'string symbol negate':
1855 1855 return smallbonus, x # single revisions are small
1856 1856 elif op == 'and':
1857 1857 wa, ta = optimize(x[1], True)
1858 1858 wb, tb = optimize(x[2], True)
1859 1859
1860 1860 # (::x and not ::y)/(not ::y and ::x) have a fast path
1861 1861 def isonly(revs, bases):
1862 1862 return (
1863 1863 revs[0] == 'func'
1864 1864 and getstring(revs[1], _('not a symbol')) == 'ancestors'
1865 1865 and bases[0] == 'not'
1866 1866 and bases[1][0] == 'func'
1867 1867 and getstring(bases[1][1], _('not a symbol')) == 'ancestors')
1868 1868
1869 1869 w = min(wa, wb)
1870 1870 if isonly(ta, tb):
1871 1871 return w, ('func', ('symbol', 'only'), ('list', ta[2], tb[1][2]))
1872 1872 if isonly(tb, ta):
1873 1873 return w, ('func', ('symbol', 'only'), ('list', tb[2], ta[1][2]))
1874 1874
1875 1875 if wa > wb:
1876 1876 return w, (op, tb, ta)
1877 1877 return w, (op, ta, tb)
1878 1878 elif op == 'or':
1879 1879 wa, ta = optimize(x[1], False)
1880 1880 wb, tb = optimize(x[2], False)
1881 1881 if wb < wa:
1882 1882 wb, wa = wa, wb
1883 1883 return max(wa, wb), (op, ta, tb)
1884 1884 elif op == 'not':
1885 1885 o = optimize(x[1], not small)
1886 1886 return o[0], (op, o[1])
1887 1887 elif op == 'parentpost':
1888 1888 o = optimize(x[1], small)
1889 1889 return o[0], (op, o[1])
1890 1890 elif op == 'group':
1891 1891 return optimize(x[1], small)
1892 1892 elif op in 'dagrange range list parent ancestorspec':
1893 1893 if op == 'parent':
1894 1894 # x^:y means (x^) : y, not x ^ (:y)
1895 1895 post = ('parentpost', x[1])
1896 1896 if x[2][0] == 'dagrangepre':
1897 1897 return optimize(('dagrange', post, x[2][1]), small)
1898 1898 elif x[2][0] == 'rangepre':
1899 1899 return optimize(('range', post, x[2][1]), small)
1900 1900
1901 1901 wa, ta = optimize(x[1], small)
1902 1902 wb, tb = optimize(x[2], small)
1903 1903 return wa + wb, (op, ta, tb)
1904 1904 elif op == 'func':
1905 1905 f = getstring(x[1], _("not a symbol"))
1906 1906 wa, ta = optimize(x[2], small)
1907 1907 if f in ("author branch closed date desc file grep keyword "
1908 1908 "outgoing user"):
1909 1909 w = 10 # slow
1910 1910 elif f in "modifies adds removes":
1911 1911 w = 30 # slower
1912 1912 elif f == "contains":
1913 1913 w = 100 # very slow
1914 1914 elif f == "ancestor":
1915 1915 w = 1 * smallbonus
1916 1916 elif f in "reverse limit first":
1917 1917 w = 0
1918 1918 elif f in "sort":
1919 1919 w = 10 # assume most sorts look at changelog
1920 1920 else:
1921 1921 w = 1
1922 1922 return w + wa, (op, x[1], ta)
1923 1923 return 1, x
1924 1924
1925 1925 _aliasarg = ('func', ('symbol', '_aliasarg'))
1926 1926 def _getaliasarg(tree):
1927 1927 """If tree matches ('func', ('symbol', '_aliasarg'), ('string', X))
1928 1928 return X, None otherwise.
1929 1929 """
1930 1930 if (len(tree) == 3 and tree[:2] == _aliasarg
1931 1931 and tree[2][0] == 'string'):
1932 1932 return tree[2][1]
1933 1933 return None
1934 1934
1935 1935 def _checkaliasarg(tree, known=None):
1936 1936 """Check tree contains no _aliasarg construct or only ones which
1937 1937 value is in known. Used to avoid alias placeholders injection.
1938 1938 """
1939 1939 if isinstance(tree, tuple):
1940 1940 arg = _getaliasarg(tree)
1941 1941 if arg is not None and (not known or arg not in known):
1942 1942 raise error.ParseError(_("not a function: %s") % '_aliasarg')
1943 1943 for t in tree:
1944 1944 _checkaliasarg(t, known)
1945 1945
1946 1946 class revsetalias(object):
1947 1947 funcre = re.compile('^([^(]+)\(([^)]+)\)$')
1948 1948 args = None
1949 1949
1950 1950 def __init__(self, name, value):
1951 1951 '''Aliases like:
1952 1952
1953 1953 h = heads(default)
1954 1954 b($1) = ancestors($1) - ancestors(default)
1955 1955 '''
1956 1956 m = self.funcre.search(name)
1957 1957 if m:
1958 1958 self.name = m.group(1)
1959 1959 self.tree = ('func', ('symbol', m.group(1)))
1960 1960 self.args = [x.strip() for x in m.group(2).split(',')]
1961 1961 for arg in self.args:
1962 1962 # _aliasarg() is an unknown symbol only used separate
1963 1963 # alias argument placeholders from regular strings.
1964 1964 value = value.replace(arg, '_aliasarg(%r)' % (arg,))
1965 1965 else:
1966 1966 self.name = name
1967 1967 self.tree = ('symbol', name)
1968 1968
1969 1969 self.replacement, pos = parse(value)
1970 1970 if pos != len(value):
1971 1971 raise error.ParseError(_('invalid token'), pos)
1972 1972 # Check for placeholder injection
1973 1973 _checkaliasarg(self.replacement, self.args)
1974 1974
1975 1975 def _getalias(aliases, tree):
1976 1976 """If tree looks like an unexpanded alias, return it. Return None
1977 1977 otherwise.
1978 1978 """
1979 1979 if isinstance(tree, tuple) and tree:
1980 1980 if tree[0] == 'symbol' and len(tree) == 2:
1981 1981 name = tree[1]
1982 1982 alias = aliases.get(name)
1983 1983 if alias and alias.args is None and alias.tree == tree:
1984 1984 return alias
1985 1985 if tree[0] == 'func' and len(tree) > 1:
1986 1986 if tree[1][0] == 'symbol' and len(tree[1]) == 2:
1987 1987 name = tree[1][1]
1988 1988 alias = aliases.get(name)
1989 1989 if alias and alias.args is not None and alias.tree == tree[:2]:
1990 1990 return alias
1991 1991 return None
1992 1992
1993 1993 def _expandargs(tree, args):
1994 1994 """Replace _aliasarg instances with the substitution value of the
1995 1995 same name in args, recursively.
1996 1996 """
1997 1997 if not tree or not isinstance(tree, tuple):
1998 1998 return tree
1999 1999 arg = _getaliasarg(tree)
2000 2000 if arg is not None:
2001 2001 return args[arg]
2002 2002 return tuple(_expandargs(t, args) for t in tree)
2003 2003
2004 2004 def _expandaliases(aliases, tree, expanding, cache):
2005 2005 """Expand aliases in tree, recursively.
2006 2006
2007 2007 'aliases' is a dictionary mapping user defined aliases to
2008 2008 revsetalias objects.
2009 2009 """
2010 2010 if not isinstance(tree, tuple):
2011 2011 # Do not expand raw strings
2012 2012 return tree
2013 2013 alias = _getalias(aliases, tree)
2014 2014 if alias is not None:
2015 2015 if alias in expanding:
2016 2016 raise error.ParseError(_('infinite expansion of revset alias "%s" '
2017 2017 'detected') % alias.name)
2018 2018 expanding.append(alias)
2019 2019 if alias.name not in cache:
2020 2020 cache[alias.name] = _expandaliases(aliases, alias.replacement,
2021 2021 expanding, cache)
2022 2022 result = cache[alias.name]
2023 2023 expanding.pop()
2024 2024 if alias.args is not None:
2025 2025 l = getlist(tree[2])
2026 2026 if len(l) != len(alias.args):
2027 2027 raise error.ParseError(
2028 2028 _('invalid number of arguments: %s') % len(l))
2029 2029 l = [_expandaliases(aliases, a, [], cache) for a in l]
2030 2030 result = _expandargs(result, dict(zip(alias.args, l)))
2031 2031 else:
2032 2032 result = tuple(_expandaliases(aliases, t, expanding, cache)
2033 2033 for t in tree)
2034 2034 return result
2035 2035
2036 2036 def findaliases(ui, tree):
2037 2037 _checkaliasarg(tree)
2038 2038 aliases = {}
2039 2039 for k, v in ui.configitems('revsetalias'):
2040 2040 alias = revsetalias(k, v)
2041 2041 aliases[alias.name] = alias
2042 2042 return _expandaliases(aliases, tree, [], {})
2043 2043
2044 2044 def parse(spec, lookup=None):
2045 2045 p = parser.parser(tokenize, elements)
2046 2046 return p.parse(spec, lookup=lookup)
2047 2047
2048 2048 def match(ui, spec, repo=None):
2049 2049 if not spec:
2050 2050 raise error.ParseError(_("empty query"))
2051 2051 lookup = None
2052 2052 if repo:
2053 2053 lookup = repo.__contains__
2054 2054 tree, pos = parse(spec, lookup)
2055 2055 if (pos != len(spec)):
2056 2056 raise error.ParseError(_("invalid token"), pos)
2057 2057 if ui:
2058 2058 tree = findaliases(ui, tree)
2059 2059 weight, tree = optimize(tree, True)
2060 2060 def mfunc(repo, subset):
2061 2061 if util.safehasattr(subset, 'set'):
2062 2062 return getset(repo, subset, tree)
2063 2063 return getset(repo, baseset(subset), tree)
2064 2064 return mfunc
2065 2065
2066 2066 def formatspec(expr, *args):
2067 2067 '''
2068 2068 This is a convenience function for using revsets internally, and
2069 2069 escapes arguments appropriately. Aliases are intentionally ignored
2070 2070 so that intended expression behavior isn't accidentally subverted.
2071 2071
2072 2072 Supported arguments:
2073 2073
2074 2074 %r = revset expression, parenthesized
2075 2075 %d = int(arg), no quoting
2076 2076 %s = string(arg), escaped and single-quoted
2077 2077 %b = arg.branch(), escaped and single-quoted
2078 2078 %n = hex(arg), single-quoted
2079 2079 %% = a literal '%'
2080 2080
2081 2081 Prefixing the type with 'l' specifies a parenthesized list of that type.
2082 2082
2083 2083 >>> formatspec('%r:: and %lr', '10 or 11', ("this()", "that()"))
2084 2084 '(10 or 11):: and ((this()) or (that()))'
2085 2085 >>> formatspec('%d:: and not %d::', 10, 20)
2086 2086 '10:: and not 20::'
2087 2087 >>> formatspec('%ld or %ld', [], [1])
2088 2088 "_list('') or 1"
2089 2089 >>> formatspec('keyword(%s)', 'foo\\xe9')
2090 2090 "keyword('foo\\\\xe9')"
2091 2091 >>> b = lambda: 'default'
2092 2092 >>> b.branch = b
2093 2093 >>> formatspec('branch(%b)', b)
2094 2094 "branch('default')"
2095 2095 >>> formatspec('root(%ls)', ['a', 'b', 'c', 'd'])
2096 2096 "root(_list('a\\x00b\\x00c\\x00d'))"
2097 2097 '''
2098 2098
2099 2099 def quote(s):
2100 2100 return repr(str(s))
2101 2101
2102 2102 def argtype(c, arg):
2103 2103 if c == 'd':
2104 2104 return str(int(arg))
2105 2105 elif c == 's':
2106 2106 return quote(arg)
2107 2107 elif c == 'r':
2108 2108 parse(arg) # make sure syntax errors are confined
2109 2109 return '(%s)' % arg
2110 2110 elif c == 'n':
2111 2111 return quote(node.hex(arg))
2112 2112 elif c == 'b':
2113 2113 return quote(arg.branch())
2114 2114
2115 2115 def listexp(s, t):
2116 2116 l = len(s)
2117 2117 if l == 0:
2118 2118 return "_list('')"
2119 2119 elif l == 1:
2120 2120 return argtype(t, s[0])
2121 2121 elif t == 'd':
2122 2122 return "_intlist('%s')" % "\0".join(str(int(a)) for a in s)
2123 2123 elif t == 's':
2124 2124 return "_list('%s')" % "\0".join(s)
2125 2125 elif t == 'n':
2126 2126 return "_hexlist('%s')" % "\0".join(node.hex(a) for a in s)
2127 2127 elif t == 'b':
2128 2128 return "_list('%s')" % "\0".join(a.branch() for a in s)
2129 2129
2130 2130 m = l // 2
2131 2131 return '(%s or %s)' % (listexp(s[:m], t), listexp(s[m:], t))
2132 2132
2133 2133 ret = ''
2134 2134 pos = 0
2135 2135 arg = 0
2136 2136 while pos < len(expr):
2137 2137 c = expr[pos]
2138 2138 if c == '%':
2139 2139 pos += 1
2140 2140 d = expr[pos]
2141 2141 if d == '%':
2142 2142 ret += d
2143 2143 elif d in 'dsnbr':
2144 2144 ret += argtype(d, args[arg])
2145 2145 arg += 1
2146 2146 elif d == 'l':
2147 2147 # a list of some type
2148 2148 pos += 1
2149 2149 d = expr[pos]
2150 2150 ret += listexp(list(args[arg]), d)
2151 2151 arg += 1
2152 2152 else:
2153 2153 raise util.Abort('unexpected revspec format character %s' % d)
2154 2154 else:
2155 2155 ret += c
2156 2156 pos += 1
2157 2157
2158 2158 return ret
2159 2159
2160 2160 def prettyformat(tree):
2161 2161 def _prettyformat(tree, level, lines):
2162 2162 if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
2163 2163 lines.append((level, str(tree)))
2164 2164 else:
2165 2165 lines.append((level, '(%s' % tree[0]))
2166 2166 for s in tree[1:]:
2167 2167 _prettyformat(s, level + 1, lines)
2168 2168 lines[-1:] = [(lines[-1][0], lines[-1][1] + ')')]
2169 2169
2170 2170 lines = []
2171 2171 _prettyformat(tree, 0, lines)
2172 2172 output = '\n'.join((' '*l + s) for l, s in lines)
2173 2173 return output
2174 2174
2175 2175 def depth(tree):
2176 2176 if isinstance(tree, tuple):
2177 2177 return max(map(depth, tree)) + 1
2178 2178 else:
2179 2179 return 0
2180 2180
2181 2181 def funcsused(tree):
2182 2182 if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
2183 2183 return set()
2184 2184 else:
2185 2185 funcs = set()
2186 2186 for s in tree[1:]:
2187 2187 funcs |= funcsused(s)
2188 2188 if tree[0] == 'func':
2189 2189 funcs.add(tree[1][1])
2190 2190 return funcs
2191 2191
2192 2192 class baseset(list):
2193 2193 """Basic data structure that represents a revset and contains the basic
2194 2194 operation that it should be able to perform.
2195 2195
2196 2196 Every method in this class should be implemented by any smartset class.
2197 2197 """
2198 2198 def __init__(self, data=()):
2199 2199 super(baseset, self).__init__(data)
2200 2200 self._set = None
2201 2201
2202 2202 def ascending(self):
2203 2203 """Sorts the set in ascending order (in place).
2204 2204
2205 2205 This is part of the mandatory API for smartset."""
2206 2206 self.sort()
2207 2207
2208 2208 def descending(self):
2209 2209 """Sorts the set in descending order (in place).
2210 2210
2211 2211 This is part of the mandatory API for smartset."""
2212 2212 self.sort(reverse=True)
2213 2213
2214 2214 def min(self):
2215 2215 return min(self)
2216 2216
2217 2217 def max(self):
2218 2218 return max(self)
2219 2219
2220 2220 def set(self):
2221 2221 """Returns a set or a smartset containing all the elements.
2222 2222
2223 2223 The returned structure should be the fastest option for membership
2224 2224 testing.
2225 2225
2226 2226 This is part of the mandatory API for smartset."""
2227 2227 if not self._set:
2228 2228 self._set = set(self)
2229 2229 return self._set
2230 2230
2231 2231 def __sub__(self, other):
2232 2232 """Returns a new object with the substraction of the two collections.
2233 2233
2234 2234 This is part of the mandatory API for smartset."""
2235 # If we are operating on 2 baseset, do the computation now since all
2236 # data is available. The alternative is to involve a lazyset, which
2237 # may be slow.
2238 if isinstance(other, baseset):
2239 other = other.set()
2240 return baseset([x for x in self if x not in other])
2241
2235 2242 return self.filter(lambda x: x not in other)
2236 2243
2237 2244 def __and__(self, other):
2238 2245 """Returns a new object with the intersection of the two collections.
2239 2246
2240 2247 This is part of the mandatory API for smartset."""
2241 2248 if isinstance(other, baseset):
2242 2249 other = other.set()
2243 2250 return baseset([y for y in self if y in other])
2244 2251
2245 2252 def __add__(self, other):
2246 2253 """Returns a new object with the union of the two collections.
2247 2254
2248 2255 This is part of the mandatory API for smartset."""
2249 2256 s = self.set()
2250 2257 l = [r for r in other if r not in s]
2251 2258 return baseset(list(self) + l)
2252 2259
2253 2260 def isascending(self):
2254 2261 """Returns True if the collection is ascending order, False if not.
2255 2262
2256 2263 This is part of the mandatory API for smartset."""
2257 2264 return False
2258 2265
2259 2266 def isdescending(self):
2260 2267 """Returns True if the collection is descending order, False if not.
2261 2268
2262 2269 This is part of the mandatory API for smartset."""
2263 2270 return False
2264 2271
2265 2272 def filter(self, condition):
2266 2273 """Returns this smartset filtered by condition as a new smartset.
2267 2274
2268 2275 `condition` is a callable which takes a revision number and returns a
2269 2276 boolean.
2270 2277
2271 2278 This is part of the mandatory API for smartset."""
2272 2279 return lazyset(self, condition)
2273 2280
2274 2281 class _orderedsetmixin(object):
2275 2282 """Mixin class with utility methods for smartsets
2276 2283
2277 2284 This should be extended by smartsets which have the isascending(),
2278 2285 isdescending() and reverse() methods"""
2279 2286
2280 2287 def _first(self):
2281 2288 """return the first revision in the set"""
2282 2289 for r in self:
2283 2290 return r
2284 2291 raise ValueError('arg is an empty sequence')
2285 2292
2286 2293 def _last(self):
2287 2294 """return the last revision in the set"""
2288 2295 self.reverse()
2289 2296 m = self._first()
2290 2297 self.reverse()
2291 2298 return m
2292 2299
2293 2300 def min(self):
2294 2301 """return the smallest element in the set"""
2295 2302 if self.isascending():
2296 2303 return self._first()
2297 2304 return self._last()
2298 2305
2299 2306 def max(self):
2300 2307 """return the largest element in the set"""
2301 2308 if self.isascending():
2302 2309 return self._last()
2303 2310 return self._first()
2304 2311
2305 2312 class lazyset(object):
2306 2313 """Duck type for baseset class which iterates lazily over the revisions in
2307 2314 the subset and contains a function which tests for membership in the
2308 2315 revset
2309 2316 """
2310 2317 def __init__(self, subset, condition=lambda x: True):
2311 2318 """
2312 2319 condition: a function that decide whether a revision in the subset
2313 2320 belongs to the revset or not.
2314 2321 """
2315 2322 self._subset = subset
2316 2323 self._condition = condition
2317 2324 self._cache = {}
2318 2325
2319 2326 def ascending(self):
2320 2327 self._subset.sort()
2321 2328
2322 2329 def descending(self):
2323 2330 self._subset.sort(reverse=True)
2324 2331
2325 2332 def min(self):
2326 2333 return min(self)
2327 2334
2328 2335 def max(self):
2329 2336 return max(self)
2330 2337
2331 2338 def __contains__(self, x):
2332 2339 c = self._cache
2333 2340 if x not in c:
2334 2341 c[x] = x in self._subset and self._condition(x)
2335 2342 return c[x]
2336 2343
2337 2344 def __iter__(self):
2338 2345 cond = self._condition
2339 2346 for x in self._subset:
2340 2347 if cond(x):
2341 2348 yield x
2342 2349
2343 2350 def __and__(self, x):
2344 2351 return lazyset(self, x.__contains__)
2345 2352
2346 2353 def __sub__(self, x):
2347 2354 return lazyset(self, lambda r: r not in x)
2348 2355
2349 2356 def __add__(self, x):
2350 2357 return _addset(self, x)
2351 2358
2352 2359 def __nonzero__(self):
2353 2360 for r in self:
2354 2361 return True
2355 2362 return False
2356 2363
2357 2364 def __len__(self):
2358 2365 # Basic implementation to be changed in future patches.
2359 2366 l = baseset([r for r in self])
2360 2367 return len(l)
2361 2368
2362 2369 def __getitem__(self, x):
2363 2370 # Basic implementation to be changed in future patches.
2364 2371 l = baseset([r for r in self])
2365 2372 return l[x]
2366 2373
2367 2374 def sort(self, reverse=False):
2368 2375 if not util.safehasattr(self._subset, 'sort'):
2369 2376 self._subset = baseset(self._subset)
2370 2377 self._subset.sort(reverse=reverse)
2371 2378
2372 2379 def reverse(self):
2373 2380 self._subset.reverse()
2374 2381
2375 2382 def set(self):
2376 2383 return set([r for r in self])
2377 2384
2378 2385 def isascending(self):
2379 2386 return False
2380 2387
2381 2388 def isdescending(self):
2382 2389 return False
2383 2390
2384 2391 def filter(self, l):
2385 2392 return lazyset(self, l)
2386 2393
2387 2394 class orderedlazyset(_orderedsetmixin, lazyset):
2388 2395 """Subclass of lazyset which subset can be ordered either ascending or
2389 2396 descendingly
2390 2397 """
2391 2398 def __init__(self, subset, condition, ascending=True):
2392 2399 super(orderedlazyset, self).__init__(subset, condition)
2393 2400 self._ascending = ascending
2394 2401
2395 2402 def filter(self, l):
2396 2403 return orderedlazyset(self, l, ascending=self._ascending)
2397 2404
2398 2405 def ascending(self):
2399 2406 if not self._ascending:
2400 2407 self.reverse()
2401 2408
2402 2409 def descending(self):
2403 2410 if self._ascending:
2404 2411 self.reverse()
2405 2412
2406 2413 def __and__(self, x):
2407 2414 return orderedlazyset(self, x.__contains__,
2408 2415 ascending=self._ascending)
2409 2416
2410 2417 def __sub__(self, x):
2411 2418 return orderedlazyset(self, lambda r: r not in x,
2412 2419 ascending=self._ascending)
2413 2420
2414 2421 def __add__(self, x):
2415 2422 kwargs = {}
2416 2423 if self.isascending() and x.isascending():
2417 2424 kwargs['ascending'] = True
2418 2425 if self.isdescending() and x.isdescending():
2419 2426 kwargs['ascending'] = False
2420 2427 return _addset(self, x, **kwargs)
2421 2428
2422 2429 def sort(self, reverse=False):
2423 2430 if reverse:
2424 2431 if self._ascending:
2425 2432 self._subset.sort(reverse=reverse)
2426 2433 else:
2427 2434 if not self._ascending:
2428 2435 self._subset.sort(reverse=reverse)
2429 2436 self._ascending = not reverse
2430 2437
2431 2438 def isascending(self):
2432 2439 return self._ascending
2433 2440
2434 2441 def isdescending(self):
2435 2442 return not self._ascending
2436 2443
2437 2444 def reverse(self):
2438 2445 self._subset.reverse()
2439 2446 self._ascending = not self._ascending
2440 2447
2441 2448 class _addset(_orderedsetmixin):
2442 2449 """Represent the addition of two sets
2443 2450
2444 2451 Wrapper structure for lazily adding two structures without losing much
2445 2452 performance on the __contains__ method
2446 2453
2447 2454 If the ascending attribute is set, that means the two structures are
2448 2455 ordered in either an ascending or descending way. Therefore, we can add
2449 2456 them maintaining the order by iterating over both at the same time
2450 2457
2451 2458 This class does not duck-type baseset and it's only supposed to be used
2452 2459 internally
2453 2460 """
2454 2461 def __init__(self, revs1, revs2, ascending=None):
2455 2462 self._r1 = revs1
2456 2463 self._r2 = revs2
2457 2464 self._iter = None
2458 2465 self._ascending = ascending
2459 2466 self._genlist = None
2460 2467
2461 2468 def __len__(self):
2462 2469 return len(self._list)
2463 2470
2464 2471 @util.propertycache
2465 2472 def _list(self):
2466 2473 if not self._genlist:
2467 2474 self._genlist = baseset(self._iterator())
2468 2475 return self._genlist
2469 2476
2470 2477 def filter(self, condition):
2471 2478 if self._ascending is not None:
2472 2479 return orderedlazyset(self, condition, ascending=self._ascending)
2473 2480 return lazyset(self, condition)
2474 2481
2475 2482 def ascending(self):
2476 2483 if self._ascending is None:
2477 2484 self.sort()
2478 2485 self._ascending = True
2479 2486 else:
2480 2487 if not self._ascending:
2481 2488 self.reverse()
2482 2489
2483 2490 def descending(self):
2484 2491 if self._ascending is None:
2485 2492 self.sort(reverse=True)
2486 2493 self._ascending = False
2487 2494 else:
2488 2495 if self._ascending:
2489 2496 self.reverse()
2490 2497
2491 2498 def __and__(self, other):
2492 2499 filterfunc = other.__contains__
2493 2500 if self._ascending is not None:
2494 2501 return orderedlazyset(self, filterfunc, ascending=self._ascending)
2495 2502 return lazyset(self, filterfunc)
2496 2503
2497 2504 def __sub__(self, other):
2498 2505 filterfunc = lambda r: r not in other
2499 2506 if self._ascending is not None:
2500 2507 return orderedlazyset(self, filterfunc, ascending=self._ascending)
2501 2508 return lazyset(self, filterfunc)
2502 2509
2503 2510 def __add__(self, other):
2504 2511 """When both collections are ascending or descending, preserve the order
2505 2512 """
2506 2513 kwargs = {}
2507 2514 if self._ascending is not None:
2508 2515 if self.isascending() and other.isascending():
2509 2516 kwargs['ascending'] = True
2510 2517 if self.isdescending() and other.isdescending():
2511 2518 kwargs['ascending'] = False
2512 2519 return _addset(self, other, **kwargs)
2513 2520
2514 2521 def _iterator(self):
2515 2522 """Iterate over both collections without repeating elements
2516 2523
2517 2524 If the ascending attribute is not set, iterate over the first one and
2518 2525 then over the second one checking for membership on the first one so we
2519 2526 dont yield any duplicates.
2520 2527
2521 2528 If the ascending attribute is set, iterate over both collections at the
2522 2529 same time, yielding only one value at a time in the given order.
2523 2530 """
2524 2531 if not self._iter:
2525 2532 def gen():
2526 2533 if self._ascending is None:
2527 2534 for r in self._r1:
2528 2535 yield r
2529 2536 s = self._r1.set()
2530 2537 for r in self._r2:
2531 2538 if r not in s:
2532 2539 yield r
2533 2540 else:
2534 2541 iter1 = iter(self._r1)
2535 2542 iter2 = iter(self._r2)
2536 2543
2537 2544 val1 = None
2538 2545 val2 = None
2539 2546
2540 2547 choice = max
2541 2548 if self._ascending:
2542 2549 choice = min
2543 2550 try:
2544 2551 # Consume both iterators in an ordered way until one is
2545 2552 # empty
2546 2553 while True:
2547 2554 if val1 is None:
2548 2555 val1 = iter1.next()
2549 2556 if val2 is None:
2550 2557 val2 = iter2.next()
2551 2558 next = choice(val1, val2)
2552 2559 yield next
2553 2560 if val1 == next:
2554 2561 val1 = None
2555 2562 if val2 == next:
2556 2563 val2 = None
2557 2564 except StopIteration:
2558 2565 # Flush any remaining values and consume the other one
2559 2566 it = iter2
2560 2567 if val1 is not None:
2561 2568 yield val1
2562 2569 it = iter1
2563 2570 elif val2 is not None:
2564 2571 # might have been equality and both are empty
2565 2572 yield val2
2566 2573 for val in it:
2567 2574 yield val
2568 2575
2569 2576 self._iter = _generatorset(gen())
2570 2577
2571 2578 return self._iter
2572 2579
2573 2580 def __iter__(self):
2574 2581 if self._genlist:
2575 2582 return iter(self._genlist)
2576 2583 return iter(self._iterator())
2577 2584
2578 2585 def __contains__(self, x):
2579 2586 return x in self._r1 or x in self._r2
2580 2587
2581 2588 def set(self):
2582 2589 return self
2583 2590
2584 2591 def sort(self, reverse=False):
2585 2592 """Sort the added set
2586 2593
2587 2594 For this we use the cached list with all the generated values and if we
2588 2595 know they are ascending or descending we can sort them in a smart way.
2589 2596 """
2590 2597 if self._ascending is None:
2591 2598 self._list.sort(reverse=reverse)
2592 2599 self._ascending = not reverse
2593 2600 else:
2594 2601 if bool(self._ascending) == bool(reverse):
2595 2602 self.reverse()
2596 2603
2597 2604 def isascending(self):
2598 2605 return self._ascending is not None and self._ascending
2599 2606
2600 2607 def isdescending(self):
2601 2608 return self._ascending is not None and not self._ascending
2602 2609
2603 2610 def reverse(self):
2604 2611 self._list.reverse()
2605 2612 if self._ascending is not None:
2606 2613 self._ascending = not self._ascending
2607 2614
2608 2615 class _generatorset(object):
2609 2616 """Wrap a generator for lazy iteration
2610 2617
2611 2618 Wrapper structure for generators that provides lazy membership and can
2612 2619 be iterated more than once.
2613 2620 When asked for membership it generates values until either it finds the
2614 2621 requested one or has gone through all the elements in the generator
2615 2622
2616 2623 This class does not duck-type baseset and it's only supposed to be used
2617 2624 internally
2618 2625 """
2619 2626 def __init__(self, gen):
2620 2627 """
2621 2628 gen: a generator producing the values for the generatorset.
2622 2629 """
2623 2630 self._gen = gen
2624 2631 self._cache = {}
2625 2632 self._genlist = baseset([])
2626 2633 self._finished = False
2627 2634
2628 2635 def __contains__(self, x):
2629 2636 if x in self._cache:
2630 2637 return self._cache[x]
2631 2638
2632 2639 # Use new values only, as existing values would be cached.
2633 2640 for l in self._consumegen():
2634 2641 if l == x:
2635 2642 return True
2636 2643
2637 2644 self._cache[x] = False
2638 2645 return False
2639 2646
2640 2647 def __iter__(self):
2641 2648 if self._finished:
2642 2649 for x in self._genlist:
2643 2650 yield x
2644 2651 return
2645 2652
2646 2653 i = 0
2647 2654 genlist = self._genlist
2648 2655 consume = self._consumegen()
2649 2656 while True:
2650 2657 if i < len(genlist):
2651 2658 yield genlist[i]
2652 2659 else:
2653 2660 yield consume.next()
2654 2661 i += 1
2655 2662
2656 2663 def _consumegen(self):
2657 2664 for item in self._gen:
2658 2665 self._cache[item] = True
2659 2666 self._genlist.append(item)
2660 2667 yield item
2661 2668 self._finished = True
2662 2669
2663 2670 def set(self):
2664 2671 return self
2665 2672
2666 2673 def sort(self, reverse=False):
2667 2674 if not self._finished:
2668 2675 for i in self:
2669 2676 continue
2670 2677 self._genlist.sort(reverse=reverse)
2671 2678
2672 2679 class _ascgeneratorset(_generatorset):
2673 2680 """Wrap a generator of ascending elements for lazy iteration
2674 2681
2675 2682 Same structure as _generatorset but stops iterating after it goes past
2676 2683 the value when asked for membership and the element is not contained
2677 2684
2678 2685 This class does not duck-type baseset and it's only supposed to be used
2679 2686 internally
2680 2687 """
2681 2688 def __contains__(self, x):
2682 2689 if x in self._cache:
2683 2690 return self._cache[x]
2684 2691
2685 2692 # Use new values only, as existing values would be cached.
2686 2693 for l in self._consumegen():
2687 2694 if l == x:
2688 2695 return True
2689 2696 if l > x:
2690 2697 break
2691 2698
2692 2699 self._cache[x] = False
2693 2700 return False
2694 2701
2695 2702 class _descgeneratorset(_generatorset):
2696 2703 """Wrap a generator of descending elements for lazy iteration
2697 2704
2698 2705 Same structure as _generatorset but stops iterating after it goes past
2699 2706 the value when asked for membership and the element is not contained
2700 2707
2701 2708 This class does not duck-type baseset and it's only supposed to be used
2702 2709 internally
2703 2710 """
2704 2711 def __contains__(self, x):
2705 2712 if x in self._cache:
2706 2713 return self._cache[x]
2707 2714
2708 2715 # Use new values only, as existing values would be cached.
2709 2716 for l in self._consumegen():
2710 2717 if l == x:
2711 2718 return True
2712 2719 if l < x:
2713 2720 break
2714 2721
2715 2722 self._cache[x] = False
2716 2723 return False
2717 2724
2718 2725 class spanset(_orderedsetmixin):
2719 2726 """Duck type for baseset class which represents a range of revisions and
2720 2727 can work lazily and without having all the range in memory
2721 2728
2722 2729 Note that spanset(x, y) behave almost like xrange(x, y) except for two
2723 2730 notable points:
2724 2731 - when x < y it will be automatically descending,
2725 2732 - revision filtered with this repoview will be skipped.
2726 2733
2727 2734 """
2728 2735 def __init__(self, repo, start=0, end=None):
2729 2736 """
2730 2737 start: first revision included the set
2731 2738 (default to 0)
2732 2739 end: first revision excluded (last+1)
2733 2740 (default to len(repo)
2734 2741
2735 2742 Spanset will be descending if `end` < `start`.
2736 2743 """
2737 2744 self._start = start
2738 2745 if end is not None:
2739 2746 self._end = end
2740 2747 else:
2741 2748 self._end = len(repo)
2742 2749 self._hiddenrevs = repo.changelog.filteredrevs
2743 2750
2744 2751 def ascending(self):
2745 2752 if self._start > self._end:
2746 2753 self.reverse()
2747 2754
2748 2755 def descending(self):
2749 2756 if self._start < self._end:
2750 2757 self.reverse()
2751 2758
2752 2759 def __iter__(self):
2753 2760 if self._start <= self._end:
2754 2761 iterrange = xrange(self._start, self._end)
2755 2762 else:
2756 2763 iterrange = xrange(self._start, self._end, -1)
2757 2764
2758 2765 if self._hiddenrevs:
2759 2766 s = self._hiddenrevs
2760 2767 for r in iterrange:
2761 2768 if r not in s:
2762 2769 yield r
2763 2770 else:
2764 2771 for r in iterrange:
2765 2772 yield r
2766 2773
2767 2774 def __contains__(self, rev):
2768 2775 return (((self._end < rev <= self._start)
2769 2776 or (self._start <= rev < self._end))
2770 2777 and not (self._hiddenrevs and rev in self._hiddenrevs))
2771 2778
2772 2779 def __nonzero__(self):
2773 2780 for r in self:
2774 2781 return True
2775 2782 return False
2776 2783
2777 2784 def __and__(self, x):
2778 2785 if isinstance(x, baseset):
2779 2786 x = x.set()
2780 2787 if self._start <= self._end:
2781 2788 return orderedlazyset(self, x.__contains__)
2782 2789 else:
2783 2790 return orderedlazyset(self, x.__contains__, ascending=False)
2784 2791
2785 2792 def __sub__(self, x):
2786 2793 if isinstance(x, baseset):
2787 2794 x = x.set()
2788 2795 if self._start <= self._end:
2789 2796 return orderedlazyset(self, lambda r: r not in x)
2790 2797 else:
2791 2798 return orderedlazyset(self, lambda r: r not in x, ascending=False)
2792 2799
2793 2800 def __add__(self, x):
2794 2801 kwargs = {}
2795 2802 if self.isascending() and x.isascending():
2796 2803 kwargs['ascending'] = True
2797 2804 if self.isdescending() and x.isdescending():
2798 2805 kwargs['ascending'] = False
2799 2806 return _addset(self, x, **kwargs)
2800 2807
2801 2808 def __len__(self):
2802 2809 if not self._hiddenrevs:
2803 2810 return abs(self._end - self._start)
2804 2811 else:
2805 2812 count = 0
2806 2813 start = self._start
2807 2814 end = self._end
2808 2815 for rev in self._hiddenrevs:
2809 2816 if (end < rev <= start) or (start <= rev < end):
2810 2817 count += 1
2811 2818 return abs(self._end - self._start) - count
2812 2819
2813 2820 def __getitem__(self, x):
2814 2821 # Basic implementation to be changed in future patches.
2815 2822 l = baseset([r for r in self])
2816 2823 return l[x]
2817 2824
2818 2825 def sort(self, reverse=False):
2819 2826 if bool(reverse) != (self._start > self._end):
2820 2827 self.reverse()
2821 2828
2822 2829 def reverse(self):
2823 2830 # Just switch the _start and _end parameters
2824 2831 if self._start <= self._end:
2825 2832 self._start, self._end = self._end - 1, self._start - 1
2826 2833 else:
2827 2834 self._start, self._end = self._end + 1, self._start + 1
2828 2835
2829 2836 def set(self):
2830 2837 return self
2831 2838
2832 2839 def isascending(self):
2833 2840 return self._start < self._end
2834 2841
2835 2842 def isdescending(self):
2836 2843 return self._start > self._end
2837 2844
2838 2845 def filter(self, l):
2839 2846 if self._start <= self._end:
2840 2847 return orderedlazyset(self, l)
2841 2848 else:
2842 2849 return orderedlazyset(self, l, ascending=False)
2843 2850
2844 2851 # tell hggettext to extract docstrings from these functions:
2845 2852 i18nfunctions = symbols.values()
General Comments 0
You need to be logged in to leave comments. Login now