##// END OF EJS Templates
rawdata: forward `rawdata` call on `manifestlog`...
marmoute -
r42949:2128c76c default
parent child Browse files
Show More
@@ -1,2100 +1,2103
1 1 # manifest.py - manifest revision class for mercurial
2 2 #
3 3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 from __future__ import absolute_import
9 9
10 10 import heapq
11 11 import itertools
12 12 import struct
13 13 import weakref
14 14
15 15 from .i18n import _
16 16 from .node import (
17 17 bin,
18 18 hex,
19 19 nullid,
20 20 nullrev,
21 21 )
22 22 from . import (
23 23 error,
24 24 mdiff,
25 25 policy,
26 26 pycompat,
27 27 repository,
28 28 revlog,
29 29 util,
30 30 )
31 31 from .utils import (
32 32 interfaceutil,
33 33 )
34 34
35 35 parsers = policy.importmod(r'parsers')
36 36 propertycache = util.propertycache
37 37
38 38 # Allow tests to more easily test the alternate path in manifestdict.fastdelta()
39 39 FASTDELTA_TEXTDIFF_THRESHOLD = 1000
40 40
41 41 def _parse(data):
42 42 # This method does a little bit of excessive-looking
43 43 # precondition checking. This is so that the behavior of this
44 44 # class exactly matches its C counterpart to try and help
45 45 # prevent surprise breakage for anyone that develops against
46 46 # the pure version.
47 47 if data and data[-1:] != '\n':
48 48 raise ValueError('Manifest did not end in a newline.')
49 49 prev = None
50 50 for l in data.splitlines():
51 51 if prev is not None and prev > l:
52 52 raise ValueError('Manifest lines not in sorted order.')
53 53 prev = l
54 54 f, n = l.split('\0')
55 55 if len(n) > 40:
56 56 yield f, bin(n[:40]), n[40:]
57 57 else:
58 58 yield f, bin(n), ''
59 59
60 60 def _text(it):
61 61 files = []
62 62 lines = []
63 63 for f, n, fl in it:
64 64 files.append(f)
65 65 # if this is changed to support newlines in filenames,
66 66 # be sure to check the templates/ dir again (especially *-raw.tmpl)
67 67 lines.append("%s\0%s%s\n" % (f, hex(n), fl))
68 68
69 69 _checkforbidden(files)
70 70 return ''.join(lines)
71 71
72 72 class lazymanifestiter(object):
73 73 def __init__(self, lm):
74 74 self.pos = 0
75 75 self.lm = lm
76 76
77 77 def __iter__(self):
78 78 return self
79 79
80 80 def next(self):
81 81 try:
82 82 data, pos = self.lm._get(self.pos)
83 83 except IndexError:
84 84 raise StopIteration
85 85 if pos == -1:
86 86 self.pos += 1
87 87 return data[0]
88 88 self.pos += 1
89 89 zeropos = data.find('\x00', pos)
90 90 return data[pos:zeropos]
91 91
92 92 __next__ = next
93 93
94 94 class lazymanifestiterentries(object):
95 95 def __init__(self, lm):
96 96 self.lm = lm
97 97 self.pos = 0
98 98
99 99 def __iter__(self):
100 100 return self
101 101
102 102 def next(self):
103 103 try:
104 104 data, pos = self.lm._get(self.pos)
105 105 except IndexError:
106 106 raise StopIteration
107 107 if pos == -1:
108 108 self.pos += 1
109 109 return data
110 110 zeropos = data.find('\x00', pos)
111 111 hashval = unhexlify(data, self.lm.extrainfo[self.pos],
112 112 zeropos + 1, 40)
113 113 flags = self.lm._getflags(data, self.pos, zeropos)
114 114 self.pos += 1
115 115 return (data[pos:zeropos], hashval, flags)
116 116
117 117 __next__ = next
118 118
119 119 def unhexlify(data, extra, pos, length):
120 120 s = bin(data[pos:pos + length])
121 121 if extra:
122 122 s += chr(extra & 0xff)
123 123 return s
124 124
125 125 def _cmp(a, b):
126 126 return (a > b) - (a < b)
127 127
128 128 class _lazymanifest(object):
129 129 """A pure python manifest backed by a byte string. It is supplimented with
130 130 internal lists as it is modified, until it is compacted back to a pure byte
131 131 string.
132 132
133 133 ``data`` is the initial manifest data.
134 134
135 135 ``positions`` is a list of offsets, one per manifest entry. Positive
136 136 values are offsets into ``data``, negative values are offsets into the
137 137 ``extradata`` list. When an entry is removed, its entry is dropped from
138 138 ``positions``. The values are encoded such that when walking the list and
139 139 indexing into ``data`` or ``extradata`` as appropriate, the entries are
140 140 sorted by filename.
141 141
142 142 ``extradata`` is a list of (key, hash, flags) for entries that were added or
143 143 modified since the manifest was created or compacted.
144 144 """
145 145 def __init__(self, data, positions=None, extrainfo=None, extradata=None,
146 146 hasremovals=False):
147 147 if positions is None:
148 148 self.positions = self.findlines(data)
149 149 self.extrainfo = [0] * len(self.positions)
150 150 self.data = data
151 151 self.extradata = []
152 152 self.hasremovals = False
153 153 else:
154 154 self.positions = positions[:]
155 155 self.extrainfo = extrainfo[:]
156 156 self.extradata = extradata[:]
157 157 self.data = data
158 158 self.hasremovals = hasremovals
159 159
160 160 def findlines(self, data):
161 161 if not data:
162 162 return []
163 163 pos = data.find("\n")
164 164 if pos == -1 or data[-1:] != '\n':
165 165 raise ValueError("Manifest did not end in a newline.")
166 166 positions = [0]
167 167 prev = data[:data.find('\x00')]
168 168 while pos < len(data) - 1 and pos != -1:
169 169 positions.append(pos + 1)
170 170 nexts = data[pos + 1:data.find('\x00', pos + 1)]
171 171 if nexts < prev:
172 172 raise ValueError("Manifest lines not in sorted order.")
173 173 prev = nexts
174 174 pos = data.find("\n", pos + 1)
175 175 return positions
176 176
177 177 def _get(self, index):
178 178 # get the position encoded in pos:
179 179 # positive number is an index in 'data'
180 180 # negative number is in extrapieces
181 181 pos = self.positions[index]
182 182 if pos >= 0:
183 183 return self.data, pos
184 184 return self.extradata[-pos - 1], -1
185 185
186 186 def _getkey(self, pos):
187 187 if pos >= 0:
188 188 return self.data[pos:self.data.find('\x00', pos + 1)]
189 189 return self.extradata[-pos - 1][0]
190 190
191 191 def bsearch(self, key):
192 192 first = 0
193 193 last = len(self.positions) - 1
194 194
195 195 while first <= last:
196 196 midpoint = (first + last)//2
197 197 nextpos = self.positions[midpoint]
198 198 candidate = self._getkey(nextpos)
199 199 r = _cmp(key, candidate)
200 200 if r == 0:
201 201 return midpoint
202 202 else:
203 203 if r < 0:
204 204 last = midpoint - 1
205 205 else:
206 206 first = midpoint + 1
207 207 return -1
208 208
209 209 def bsearch2(self, key):
210 210 # same as the above, but will always return the position
211 211 # done for performance reasons
212 212 first = 0
213 213 last = len(self.positions) - 1
214 214
215 215 while first <= last:
216 216 midpoint = (first + last)//2
217 217 nextpos = self.positions[midpoint]
218 218 candidate = self._getkey(nextpos)
219 219 r = _cmp(key, candidate)
220 220 if r == 0:
221 221 return (midpoint, True)
222 222 else:
223 223 if r < 0:
224 224 last = midpoint - 1
225 225 else:
226 226 first = midpoint + 1
227 227 return (first, False)
228 228
229 229 def __contains__(self, key):
230 230 return self.bsearch(key) != -1
231 231
232 232 def _getflags(self, data, needle, pos):
233 233 start = pos + 41
234 234 end = data.find("\n", start)
235 235 if end == -1:
236 236 end = len(data) - 1
237 237 if start == end:
238 238 return ''
239 239 return self.data[start:end]
240 240
241 241 def __getitem__(self, key):
242 242 if not isinstance(key, bytes):
243 243 raise TypeError("getitem: manifest keys must be a bytes.")
244 244 needle = self.bsearch(key)
245 245 if needle == -1:
246 246 raise KeyError
247 247 data, pos = self._get(needle)
248 248 if pos == -1:
249 249 return (data[1], data[2])
250 250 zeropos = data.find('\x00', pos)
251 251 assert 0 <= needle <= len(self.positions)
252 252 assert len(self.extrainfo) == len(self.positions)
253 253 hashval = unhexlify(data, self.extrainfo[needle], zeropos + 1, 40)
254 254 flags = self._getflags(data, needle, zeropos)
255 255 return (hashval, flags)
256 256
257 257 def __delitem__(self, key):
258 258 needle, found = self.bsearch2(key)
259 259 if not found:
260 260 raise KeyError
261 261 cur = self.positions[needle]
262 262 self.positions = self.positions[:needle] + self.positions[needle + 1:]
263 263 self.extrainfo = self.extrainfo[:needle] + self.extrainfo[needle + 1:]
264 264 if cur >= 0:
265 265 # This does NOT unsort the list as far as the search functions are
266 266 # concerned, as they only examine lines mapped by self.positions.
267 267 self.data = self.data[:cur] + '\x00' + self.data[cur + 1:]
268 268 self.hasremovals = True
269 269
270 270 def __setitem__(self, key, value):
271 271 if not isinstance(key, bytes):
272 272 raise TypeError("setitem: manifest keys must be a byte string.")
273 273 if not isinstance(value, tuple) or len(value) != 2:
274 274 raise TypeError("Manifest values must be a tuple of (node, flags).")
275 275 hashval = value[0]
276 276 if not isinstance(hashval, bytes) or not 20 <= len(hashval) <= 22:
277 277 raise TypeError("node must be a 20-byte byte string")
278 278 flags = value[1]
279 279 if len(hashval) == 22:
280 280 hashval = hashval[:-1]
281 281 if not isinstance(flags, bytes) or len(flags) > 1:
282 282 raise TypeError("flags must a 0 or 1 byte string, got %r", flags)
283 283 needle, found = self.bsearch2(key)
284 284 if found:
285 285 # put the item
286 286 pos = self.positions[needle]
287 287 if pos < 0:
288 288 self.extradata[-pos - 1] = (key, hashval, value[1])
289 289 else:
290 290 # just don't bother
291 291 self.extradata.append((key, hashval, value[1]))
292 292 self.positions[needle] = -len(self.extradata)
293 293 else:
294 294 # not found, put it in with extra positions
295 295 self.extradata.append((key, hashval, value[1]))
296 296 self.positions = (self.positions[:needle] + [-len(self.extradata)]
297 297 + self.positions[needle:])
298 298 self.extrainfo = (self.extrainfo[:needle] + [0] +
299 299 self.extrainfo[needle:])
300 300
301 301 def copy(self):
302 302 # XXX call _compact like in C?
303 303 return _lazymanifest(self.data, self.positions, self.extrainfo,
304 304 self.extradata, self.hasremovals)
305 305
306 306 def _compact(self):
307 307 # hopefully not called TOO often
308 308 if len(self.extradata) == 0 and not self.hasremovals:
309 309 return
310 310 l = []
311 311 i = 0
312 312 offset = 0
313 313 self.extrainfo = [0] * len(self.positions)
314 314 while i < len(self.positions):
315 315 if self.positions[i] >= 0:
316 316 cur = self.positions[i]
317 317 last_cut = cur
318 318
319 319 # Collect all contiguous entries in the buffer at the current
320 320 # offset, breaking out only for added/modified items held in
321 321 # extradata, or a deleted line prior to the next position.
322 322 while True:
323 323 self.positions[i] = offset
324 324 i += 1
325 325 if i == len(self.positions) or self.positions[i] < 0:
326 326 break
327 327
328 328 # A removed file has no positions[] entry, but does have an
329 329 # overwritten first byte. Break out and find the end of the
330 330 # current good entry/entries if there is a removed file
331 331 # before the next position.
332 332 if (self.hasremovals
333 333 and self.data.find('\n\x00', cur,
334 334 self.positions[i]) != -1):
335 335 break
336 336
337 337 offset += self.positions[i] - cur
338 338 cur = self.positions[i]
339 339 end_cut = self.data.find('\n', cur)
340 340 if end_cut != -1:
341 341 end_cut += 1
342 342 offset += end_cut - cur
343 343 l.append(self.data[last_cut:end_cut])
344 344 else:
345 345 while i < len(self.positions) and self.positions[i] < 0:
346 346 cur = self.positions[i]
347 347 t = self.extradata[-cur - 1]
348 348 l.append(self._pack(t))
349 349 self.positions[i] = offset
350 350 if len(t[1]) > 20:
351 351 self.extrainfo[i] = ord(t[1][21])
352 352 offset += len(l[-1])
353 353 i += 1
354 354 self.data = ''.join(l)
355 355 self.hasremovals = False
356 356 self.extradata = []
357 357
358 358 def _pack(self, d):
359 359 return d[0] + '\x00' + hex(d[1][:20]) + d[2] + '\n'
360 360
361 361 def text(self):
362 362 self._compact()
363 363 return self.data
364 364
365 365 def diff(self, m2, clean=False):
366 366 '''Finds changes between the current manifest and m2.'''
367 367 # XXX think whether efficiency matters here
368 368 diff = {}
369 369
370 370 for fn, e1, flags in self.iterentries():
371 371 if fn not in m2:
372 372 diff[fn] = (e1, flags), (None, '')
373 373 else:
374 374 e2 = m2[fn]
375 375 if (e1, flags) != e2:
376 376 diff[fn] = (e1, flags), e2
377 377 elif clean:
378 378 diff[fn] = None
379 379
380 380 for fn, e2, flags in m2.iterentries():
381 381 if fn not in self:
382 382 diff[fn] = (None, ''), (e2, flags)
383 383
384 384 return diff
385 385
386 386 def iterentries(self):
387 387 return lazymanifestiterentries(self)
388 388
389 389 def iterkeys(self):
390 390 return lazymanifestiter(self)
391 391
392 392 def __iter__(self):
393 393 return lazymanifestiter(self)
394 394
395 395 def __len__(self):
396 396 return len(self.positions)
397 397
398 398 def filtercopy(self, filterfn):
399 399 # XXX should be optimized
400 400 c = _lazymanifest('')
401 401 for f, n, fl in self.iterentries():
402 402 if filterfn(f):
403 403 c[f] = n, fl
404 404 return c
405 405
406 406 try:
407 407 _lazymanifest = parsers.lazymanifest
408 408 except AttributeError:
409 409 pass
410 410
411 411 @interfaceutil.implementer(repository.imanifestdict)
412 412 class manifestdict(object):
413 413 def __init__(self, data=''):
414 414 self._lm = _lazymanifest(data)
415 415
416 416 def __getitem__(self, key):
417 417 return self._lm[key][0]
418 418
419 419 def find(self, key):
420 420 return self._lm[key]
421 421
422 422 def __len__(self):
423 423 return len(self._lm)
424 424
425 425 def __nonzero__(self):
426 426 # nonzero is covered by the __len__ function, but implementing it here
427 427 # makes it easier for extensions to override.
428 428 return len(self._lm) != 0
429 429
430 430 __bool__ = __nonzero__
431 431
432 432 def __setitem__(self, key, node):
433 433 self._lm[key] = node, self.flags(key, '')
434 434
435 435 def __contains__(self, key):
436 436 if key is None:
437 437 return False
438 438 return key in self._lm
439 439
440 440 def __delitem__(self, key):
441 441 del self._lm[key]
442 442
443 443 def __iter__(self):
444 444 return self._lm.__iter__()
445 445
446 446 def iterkeys(self):
447 447 return self._lm.iterkeys()
448 448
449 449 def keys(self):
450 450 return list(self.iterkeys())
451 451
452 452 def filesnotin(self, m2, match=None):
453 453 '''Set of files in this manifest that are not in the other'''
454 454 if match:
455 455 m1 = self.matches(match)
456 456 m2 = m2.matches(match)
457 457 return m1.filesnotin(m2)
458 458 diff = self.diff(m2)
459 459 files = set(filepath
460 460 for filepath, hashflags in diff.iteritems()
461 461 if hashflags[1][0] is None)
462 462 return files
463 463
464 464 @propertycache
465 465 def _dirs(self):
466 466 return util.dirs(self)
467 467
468 468 def dirs(self):
469 469 return self._dirs
470 470
471 471 def hasdir(self, dir):
472 472 return dir in self._dirs
473 473
474 474 def _filesfastpath(self, match):
475 475 '''Checks whether we can correctly and quickly iterate over matcher
476 476 files instead of over manifest files.'''
477 477 files = match.files()
478 478 return (len(files) < 100 and (match.isexact() or
479 479 (match.prefix() and all(fn in self for fn in files))))
480 480
481 481 def walk(self, match):
482 482 '''Generates matching file names.
483 483
484 484 Equivalent to manifest.matches(match).iterkeys(), but without creating
485 485 an entirely new manifest.
486 486
487 487 It also reports nonexistent files by marking them bad with match.bad().
488 488 '''
489 489 if match.always():
490 490 for f in iter(self):
491 491 yield f
492 492 return
493 493
494 494 fset = set(match.files())
495 495
496 496 # avoid the entire walk if we're only looking for specific files
497 497 if self._filesfastpath(match):
498 498 for fn in sorted(fset):
499 499 yield fn
500 500 return
501 501
502 502 for fn in self:
503 503 if fn in fset:
504 504 # specified pattern is the exact name
505 505 fset.remove(fn)
506 506 if match(fn):
507 507 yield fn
508 508
509 509 # for dirstate.walk, files=[''] means "walk the whole tree".
510 510 # follow that here, too
511 511 fset.discard('')
512 512
513 513 for fn in sorted(fset):
514 514 if not self.hasdir(fn):
515 515 match.bad(fn, None)
516 516
517 517 def matches(self, match):
518 518 '''generate a new manifest filtered by the match argument'''
519 519 if match.always():
520 520 return self.copy()
521 521
522 522 if self._filesfastpath(match):
523 523 m = manifestdict()
524 524 lm = self._lm
525 525 for fn in match.files():
526 526 if fn in lm:
527 527 m._lm[fn] = lm[fn]
528 528 return m
529 529
530 530 m = manifestdict()
531 531 m._lm = self._lm.filtercopy(match)
532 532 return m
533 533
534 534 def diff(self, m2, match=None, clean=False):
535 535 '''Finds changes between the current manifest and m2.
536 536
537 537 Args:
538 538 m2: the manifest to which this manifest should be compared.
539 539 clean: if true, include files unchanged between these manifests
540 540 with a None value in the returned dictionary.
541 541
542 542 The result is returned as a dict with filename as key and
543 543 values of the form ((n1,fl1),(n2,fl2)), where n1/n2 is the
544 544 nodeid in the current/other manifest and fl1/fl2 is the flag
545 545 in the current/other manifest. Where the file does not exist,
546 546 the nodeid will be None and the flags will be the empty
547 547 string.
548 548 '''
549 549 if match:
550 550 m1 = self.matches(match)
551 551 m2 = m2.matches(match)
552 552 return m1.diff(m2, clean=clean)
553 553 return self._lm.diff(m2._lm, clean)
554 554
555 555 def setflag(self, key, flag):
556 556 self._lm[key] = self[key], flag
557 557
558 558 def get(self, key, default=None):
559 559 try:
560 560 return self._lm[key][0]
561 561 except KeyError:
562 562 return default
563 563
564 564 def flags(self, key, default=''):
565 565 try:
566 566 return self._lm[key][1]
567 567 except KeyError:
568 568 return default
569 569
570 570 def copy(self):
571 571 c = manifestdict()
572 572 c._lm = self._lm.copy()
573 573 return c
574 574
575 575 def items(self):
576 576 return (x[:2] for x in self._lm.iterentries())
577 577
578 578 def iteritems(self):
579 579 return (x[:2] for x in self._lm.iterentries())
580 580
581 581 def iterentries(self):
582 582 return self._lm.iterentries()
583 583
584 584 def text(self):
585 585 # most likely uses native version
586 586 return self._lm.text()
587 587
588 588 def fastdelta(self, base, changes):
589 589 """Given a base manifest text as a bytearray and a list of changes
590 590 relative to that text, compute a delta that can be used by revlog.
591 591 """
592 592 delta = []
593 593 dstart = None
594 594 dend = None
595 595 dline = [""]
596 596 start = 0
597 597 # zero copy representation of base as a buffer
598 598 addbuf = util.buffer(base)
599 599
600 600 changes = list(changes)
601 601 if len(changes) < FASTDELTA_TEXTDIFF_THRESHOLD:
602 602 # start with a readonly loop that finds the offset of
603 603 # each line and creates the deltas
604 604 for f, todelete in changes:
605 605 # bs will either be the index of the item or the insert point
606 606 start, end = _msearch(addbuf, f, start)
607 607 if not todelete:
608 608 h, fl = self._lm[f]
609 609 l = "%s\0%s%s\n" % (f, hex(h), fl)
610 610 else:
611 611 if start == end:
612 612 # item we want to delete was not found, error out
613 613 raise AssertionError(
614 614 _("failed to remove %s from manifest") % f)
615 615 l = ""
616 616 if dstart is not None and dstart <= start and dend >= start:
617 617 if dend < end:
618 618 dend = end
619 619 if l:
620 620 dline.append(l)
621 621 else:
622 622 if dstart is not None:
623 623 delta.append([dstart, dend, "".join(dline)])
624 624 dstart = start
625 625 dend = end
626 626 dline = [l]
627 627
628 628 if dstart is not None:
629 629 delta.append([dstart, dend, "".join(dline)])
630 630 # apply the delta to the base, and get a delta for addrevision
631 631 deltatext, arraytext = _addlistdelta(base, delta)
632 632 else:
633 633 # For large changes, it's much cheaper to just build the text and
634 634 # diff it.
635 635 arraytext = bytearray(self.text())
636 636 deltatext = mdiff.textdiff(
637 637 util.buffer(base), util.buffer(arraytext))
638 638
639 639 return arraytext, deltatext
640 640
641 641 def _msearch(m, s, lo=0, hi=None):
642 642 '''return a tuple (start, end) that says where to find s within m.
643 643
644 644 If the string is found m[start:end] are the line containing
645 645 that string. If start == end the string was not found and
646 646 they indicate the proper sorted insertion point.
647 647
648 648 m should be a buffer, a memoryview or a byte string.
649 649 s is a byte string'''
650 650 def advance(i, c):
651 651 while i < lenm and m[i:i + 1] != c:
652 652 i += 1
653 653 return i
654 654 if not s:
655 655 return (lo, lo)
656 656 lenm = len(m)
657 657 if not hi:
658 658 hi = lenm
659 659 while lo < hi:
660 660 mid = (lo + hi) // 2
661 661 start = mid
662 662 while start > 0 and m[start - 1:start] != '\n':
663 663 start -= 1
664 664 end = advance(start, '\0')
665 665 if bytes(m[start:end]) < s:
666 666 # we know that after the null there are 40 bytes of sha1
667 667 # this translates to the bisect lo = mid + 1
668 668 lo = advance(end + 40, '\n') + 1
669 669 else:
670 670 # this translates to the bisect hi = mid
671 671 hi = start
672 672 end = advance(lo, '\0')
673 673 found = m[lo:end]
674 674 if s == found:
675 675 # we know that after the null there are 40 bytes of sha1
676 676 end = advance(end + 40, '\n')
677 677 return (lo, end + 1)
678 678 else:
679 679 return (lo, lo)
680 680
681 681 def _checkforbidden(l):
682 682 """Check filenames for illegal characters."""
683 683 for f in l:
684 684 if '\n' in f or '\r' in f:
685 685 raise error.StorageError(
686 686 _("'\\n' and '\\r' disallowed in filenames: %r")
687 687 % pycompat.bytestr(f))
688 688
689 689
690 690 # apply the changes collected during the bisect loop to our addlist
691 691 # return a delta suitable for addrevision
692 692 def _addlistdelta(addlist, x):
693 693 # for large addlist arrays, building a new array is cheaper
694 694 # than repeatedly modifying the existing one
695 695 currentposition = 0
696 696 newaddlist = bytearray()
697 697
698 698 for start, end, content in x:
699 699 newaddlist += addlist[currentposition:start]
700 700 if content:
701 701 newaddlist += bytearray(content)
702 702
703 703 currentposition = end
704 704
705 705 newaddlist += addlist[currentposition:]
706 706
707 707 deltatext = "".join(struct.pack(">lll", start, end, len(content))
708 708 + content for start, end, content in x)
709 709 return deltatext, newaddlist
710 710
711 711 def _splittopdir(f):
712 712 if '/' in f:
713 713 dir, subpath = f.split('/', 1)
714 714 return dir + '/', subpath
715 715 else:
716 716 return '', f
717 717
718 718 _noop = lambda s: None
719 719
720 720 class treemanifest(object):
721 721 def __init__(self, dir='', text=''):
722 722 self._dir = dir
723 723 self._node = nullid
724 724 self._loadfunc = _noop
725 725 self._copyfunc = _noop
726 726 self._dirty = False
727 727 self._dirs = {}
728 728 self._lazydirs = {}
729 729 # Using _lazymanifest here is a little slower than plain old dicts
730 730 self._files = {}
731 731 self._flags = {}
732 732 if text:
733 733 def readsubtree(subdir, subm):
734 734 raise AssertionError('treemanifest constructor only accepts '
735 735 'flat manifests')
736 736 self.parse(text, readsubtree)
737 737 self._dirty = True # Mark flat manifest dirty after parsing
738 738
739 739 def _subpath(self, path):
740 740 return self._dir + path
741 741
742 742 def _loadalllazy(self):
743 743 selfdirs = self._dirs
744 744 for d, (path, node, readsubtree, docopy) in self._lazydirs.iteritems():
745 745 if docopy:
746 746 selfdirs[d] = readsubtree(path, node).copy()
747 747 else:
748 748 selfdirs[d] = readsubtree(path, node)
749 749 self._lazydirs = {}
750 750
751 751 def _loadlazy(self, d):
752 752 v = self._lazydirs.get(d)
753 753 if v:
754 754 path, node, readsubtree, docopy = v
755 755 if docopy:
756 756 self._dirs[d] = readsubtree(path, node).copy()
757 757 else:
758 758 self._dirs[d] = readsubtree(path, node)
759 759 del self._lazydirs[d]
760 760
761 761 def _loadchildrensetlazy(self, visit):
762 762 if not visit:
763 763 return None
764 764 if visit == 'all' or visit == 'this':
765 765 self._loadalllazy()
766 766 return None
767 767
768 768 loadlazy = self._loadlazy
769 769 for k in visit:
770 770 loadlazy(k + '/')
771 771 return visit
772 772
773 773 def _loaddifflazy(self, t1, t2):
774 774 """load items in t1 and t2 if they're needed for diffing.
775 775
776 776 The criteria currently is:
777 777 - if it's not present in _lazydirs in either t1 or t2, load it in the
778 778 other (it may already be loaded or it may not exist, doesn't matter)
779 779 - if it's present in _lazydirs in both, compare the nodeid; if it
780 780 differs, load it in both
781 781 """
782 782 toloadlazy = []
783 783 for d, v1 in t1._lazydirs.iteritems():
784 784 v2 = t2._lazydirs.get(d)
785 785 if not v2 or v2[1] != v1[1]:
786 786 toloadlazy.append(d)
787 787 for d, v1 in t2._lazydirs.iteritems():
788 788 if d not in t1._lazydirs:
789 789 toloadlazy.append(d)
790 790
791 791 for d in toloadlazy:
792 792 t1._loadlazy(d)
793 793 t2._loadlazy(d)
794 794
795 795 def __len__(self):
796 796 self._load()
797 797 size = len(self._files)
798 798 self._loadalllazy()
799 799 for m in self._dirs.values():
800 800 size += m.__len__()
801 801 return size
802 802
803 803 def __nonzero__(self):
804 804 # Faster than "__len() != 0" since it avoids loading sub-manifests
805 805 return not self._isempty()
806 806
807 807 __bool__ = __nonzero__
808 808
809 809 def _isempty(self):
810 810 self._load() # for consistency; already loaded by all callers
811 811 # See if we can skip loading everything.
812 812 if self._files or (self._dirs and
813 813 any(not m._isempty() for m in self._dirs.values())):
814 814 return False
815 815 self._loadalllazy()
816 816 return (not self._dirs or
817 817 all(m._isempty() for m in self._dirs.values()))
818 818
819 819 def __repr__(self):
820 820 return ('<treemanifest dir=%s, node=%s, loaded=%s, dirty=%s at 0x%x>' %
821 821 (self._dir, hex(self._node),
822 822 bool(self._loadfunc is _noop),
823 823 self._dirty, id(self)))
824 824
825 825 def dir(self):
826 826 '''The directory that this tree manifest represents, including a
827 827 trailing '/'. Empty string for the repo root directory.'''
828 828 return self._dir
829 829
830 830 def node(self):
831 831 '''This node of this instance. nullid for unsaved instances. Should
832 832 be updated when the instance is read or written from a revlog.
833 833 '''
834 834 assert not self._dirty
835 835 return self._node
836 836
837 837 def setnode(self, node):
838 838 self._node = node
839 839 self._dirty = False
840 840
841 841 def iterentries(self):
842 842 self._load()
843 843 self._loadalllazy()
844 844 for p, n in sorted(itertools.chain(self._dirs.items(),
845 845 self._files.items())):
846 846 if p in self._files:
847 847 yield self._subpath(p), n, self._flags.get(p, '')
848 848 else:
849 849 for x in n.iterentries():
850 850 yield x
851 851
852 852 def items(self):
853 853 self._load()
854 854 self._loadalllazy()
855 855 for p, n in sorted(itertools.chain(self._dirs.items(),
856 856 self._files.items())):
857 857 if p in self._files:
858 858 yield self._subpath(p), n
859 859 else:
860 860 for f, sn in n.iteritems():
861 861 yield f, sn
862 862
863 863 iteritems = items
864 864
865 865 def iterkeys(self):
866 866 self._load()
867 867 self._loadalllazy()
868 868 for p in sorted(itertools.chain(self._dirs, self._files)):
869 869 if p in self._files:
870 870 yield self._subpath(p)
871 871 else:
872 872 for f in self._dirs[p]:
873 873 yield f
874 874
875 875 def keys(self):
876 876 return list(self.iterkeys())
877 877
878 878 def __iter__(self):
879 879 return self.iterkeys()
880 880
881 881 def __contains__(self, f):
882 882 if f is None:
883 883 return False
884 884 self._load()
885 885 dir, subpath = _splittopdir(f)
886 886 if dir:
887 887 self._loadlazy(dir)
888 888
889 889 if dir not in self._dirs:
890 890 return False
891 891
892 892 return self._dirs[dir].__contains__(subpath)
893 893 else:
894 894 return f in self._files
895 895
896 896 def get(self, f, default=None):
897 897 self._load()
898 898 dir, subpath = _splittopdir(f)
899 899 if dir:
900 900 self._loadlazy(dir)
901 901
902 902 if dir not in self._dirs:
903 903 return default
904 904 return self._dirs[dir].get(subpath, default)
905 905 else:
906 906 return self._files.get(f, default)
907 907
908 908 def __getitem__(self, f):
909 909 self._load()
910 910 dir, subpath = _splittopdir(f)
911 911 if dir:
912 912 self._loadlazy(dir)
913 913
914 914 return self._dirs[dir].__getitem__(subpath)
915 915 else:
916 916 return self._files[f]
917 917
918 918 def flags(self, f):
919 919 self._load()
920 920 dir, subpath = _splittopdir(f)
921 921 if dir:
922 922 self._loadlazy(dir)
923 923
924 924 if dir not in self._dirs:
925 925 return ''
926 926 return self._dirs[dir].flags(subpath)
927 927 else:
928 928 if f in self._lazydirs or f in self._dirs:
929 929 return ''
930 930 return self._flags.get(f, '')
931 931
932 932 def find(self, f):
933 933 self._load()
934 934 dir, subpath = _splittopdir(f)
935 935 if dir:
936 936 self._loadlazy(dir)
937 937
938 938 return self._dirs[dir].find(subpath)
939 939 else:
940 940 return self._files[f], self._flags.get(f, '')
941 941
942 942 def __delitem__(self, f):
943 943 self._load()
944 944 dir, subpath = _splittopdir(f)
945 945 if dir:
946 946 self._loadlazy(dir)
947 947
948 948 self._dirs[dir].__delitem__(subpath)
949 949 # If the directory is now empty, remove it
950 950 if self._dirs[dir]._isempty():
951 951 del self._dirs[dir]
952 952 else:
953 953 del self._files[f]
954 954 if f in self._flags:
955 955 del self._flags[f]
956 956 self._dirty = True
957 957
958 958 def __setitem__(self, f, n):
959 959 assert n is not None
960 960 self._load()
961 961 dir, subpath = _splittopdir(f)
962 962 if dir:
963 963 self._loadlazy(dir)
964 964 if dir not in self._dirs:
965 965 self._dirs[dir] = treemanifest(self._subpath(dir))
966 966 self._dirs[dir].__setitem__(subpath, n)
967 967 else:
968 968 self._files[f] = n[:21] # to match manifestdict's behavior
969 969 self._dirty = True
970 970
971 971 def _load(self):
972 972 if self._loadfunc is not _noop:
973 973 lf, self._loadfunc = self._loadfunc, _noop
974 974 lf(self)
975 975 elif self._copyfunc is not _noop:
976 976 cf, self._copyfunc = self._copyfunc, _noop
977 977 cf(self)
978 978
979 979 def setflag(self, f, flags):
980 980 """Set the flags (symlink, executable) for path f."""
981 981 self._load()
982 982 dir, subpath = _splittopdir(f)
983 983 if dir:
984 984 self._loadlazy(dir)
985 985 if dir not in self._dirs:
986 986 self._dirs[dir] = treemanifest(self._subpath(dir))
987 987 self._dirs[dir].setflag(subpath, flags)
988 988 else:
989 989 self._flags[f] = flags
990 990 self._dirty = True
991 991
992 992 def copy(self):
993 993 copy = treemanifest(self._dir)
994 994 copy._node = self._node
995 995 copy._dirty = self._dirty
996 996 if self._copyfunc is _noop:
997 997 def _copyfunc(s):
998 998 self._load()
999 999 s._lazydirs = {d: (p, n, r, True) for
1000 1000 d, (p, n, r, c) in self._lazydirs.iteritems()}
1001 1001 sdirs = s._dirs
1002 1002 for d, v in self._dirs.iteritems():
1003 1003 sdirs[d] = v.copy()
1004 1004 s._files = dict.copy(self._files)
1005 1005 s._flags = dict.copy(self._flags)
1006 1006 if self._loadfunc is _noop:
1007 1007 _copyfunc(copy)
1008 1008 else:
1009 1009 copy._copyfunc = _copyfunc
1010 1010 else:
1011 1011 copy._copyfunc = self._copyfunc
1012 1012 return copy
1013 1013
1014 1014 def filesnotin(self, m2, match=None):
1015 1015 '''Set of files in this manifest that are not in the other'''
1016 1016 if match and not match.always():
1017 1017 m1 = self.matches(match)
1018 1018 m2 = m2.matches(match)
1019 1019 return m1.filesnotin(m2)
1020 1020
1021 1021 files = set()
1022 1022 def _filesnotin(t1, t2):
1023 1023 if t1._node == t2._node and not t1._dirty and not t2._dirty:
1024 1024 return
1025 1025 t1._load()
1026 1026 t2._load()
1027 1027 self._loaddifflazy(t1, t2)
1028 1028 for d, m1 in t1._dirs.iteritems():
1029 1029 if d in t2._dirs:
1030 1030 m2 = t2._dirs[d]
1031 1031 _filesnotin(m1, m2)
1032 1032 else:
1033 1033 files.update(m1.iterkeys())
1034 1034
1035 1035 for fn in t1._files:
1036 1036 if fn not in t2._files:
1037 1037 files.add(t1._subpath(fn))
1038 1038
1039 1039 _filesnotin(self, m2)
1040 1040 return files
1041 1041
1042 1042 @propertycache
1043 1043 def _alldirs(self):
1044 1044 return util.dirs(self)
1045 1045
1046 1046 def dirs(self):
1047 1047 return self._alldirs
1048 1048
1049 1049 def hasdir(self, dir):
1050 1050 self._load()
1051 1051 topdir, subdir = _splittopdir(dir)
1052 1052 if topdir:
1053 1053 self._loadlazy(topdir)
1054 1054 if topdir in self._dirs:
1055 1055 return self._dirs[topdir].hasdir(subdir)
1056 1056 return False
1057 1057 dirslash = dir + '/'
1058 1058 return dirslash in self._dirs or dirslash in self._lazydirs
1059 1059
1060 1060 def walk(self, match):
1061 1061 '''Generates matching file names.
1062 1062
1063 1063 Equivalent to manifest.matches(match).iterkeys(), but without creating
1064 1064 an entirely new manifest.
1065 1065
1066 1066 It also reports nonexistent files by marking them bad with match.bad().
1067 1067 '''
1068 1068 if match.always():
1069 1069 for f in iter(self):
1070 1070 yield f
1071 1071 return
1072 1072
1073 1073 fset = set(match.files())
1074 1074
1075 1075 for fn in self._walk(match):
1076 1076 if fn in fset:
1077 1077 # specified pattern is the exact name
1078 1078 fset.remove(fn)
1079 1079 yield fn
1080 1080
1081 1081 # for dirstate.walk, files=[''] means "walk the whole tree".
1082 1082 # follow that here, too
1083 1083 fset.discard('')
1084 1084
1085 1085 for fn in sorted(fset):
1086 1086 if not self.hasdir(fn):
1087 1087 match.bad(fn, None)
1088 1088
1089 1089 def _walk(self, match):
1090 1090 '''Recursively generates matching file names for walk().'''
1091 1091 visit = match.visitchildrenset(self._dir[:-1])
1092 1092 if not visit:
1093 1093 return
1094 1094
1095 1095 # yield this dir's files and walk its submanifests
1096 1096 self._load()
1097 1097 visit = self._loadchildrensetlazy(visit)
1098 1098 for p in sorted(list(self._dirs) + list(self._files)):
1099 1099 if p in self._files:
1100 1100 fullp = self._subpath(p)
1101 1101 if match(fullp):
1102 1102 yield fullp
1103 1103 else:
1104 1104 if not visit or p[:-1] in visit:
1105 1105 for f in self._dirs[p]._walk(match):
1106 1106 yield f
1107 1107
1108 1108 def matches(self, match):
1109 1109 '''generate a new manifest filtered by the match argument'''
1110 1110 if match.always():
1111 1111 return self.copy()
1112 1112
1113 1113 return self._matches(match)
1114 1114
1115 1115 def _matches(self, match):
1116 1116 '''recursively generate a new manifest filtered by the match argument.
1117 1117 '''
1118 1118
1119 1119 visit = match.visitchildrenset(self._dir[:-1])
1120 1120 if visit == 'all':
1121 1121 return self.copy()
1122 1122 ret = treemanifest(self._dir)
1123 1123 if not visit:
1124 1124 return ret
1125 1125
1126 1126 self._load()
1127 1127 for fn in self._files:
1128 1128 # While visitchildrenset *usually* lists only subdirs, this is
1129 1129 # actually up to the matcher and may have some files in the set().
1130 1130 # If visit == 'this', we should obviously look at the files in this
1131 1131 # directory; if visit is a set, and fn is in it, we should inspect
1132 1132 # fn (but no need to inspect things not in the set).
1133 1133 if visit != 'this' and fn not in visit:
1134 1134 continue
1135 1135 fullp = self._subpath(fn)
1136 1136 # visitchildrenset isn't perfect, we still need to call the regular
1137 1137 # matcher code to further filter results.
1138 1138 if not match(fullp):
1139 1139 continue
1140 1140 ret._files[fn] = self._files[fn]
1141 1141 if fn in self._flags:
1142 1142 ret._flags[fn] = self._flags[fn]
1143 1143
1144 1144 visit = self._loadchildrensetlazy(visit)
1145 1145 for dir, subm in self._dirs.iteritems():
1146 1146 if visit and dir[:-1] not in visit:
1147 1147 continue
1148 1148 m = subm._matches(match)
1149 1149 if not m._isempty():
1150 1150 ret._dirs[dir] = m
1151 1151
1152 1152 if not ret._isempty():
1153 1153 ret._dirty = True
1154 1154 return ret
1155 1155
1156 1156 def diff(self, m2, match=None, clean=False):
1157 1157 '''Finds changes between the current manifest and m2.
1158 1158
1159 1159 Args:
1160 1160 m2: the manifest to which this manifest should be compared.
1161 1161 clean: if true, include files unchanged between these manifests
1162 1162 with a None value in the returned dictionary.
1163 1163
1164 1164 The result is returned as a dict with filename as key and
1165 1165 values of the form ((n1,fl1),(n2,fl2)), where n1/n2 is the
1166 1166 nodeid in the current/other manifest and fl1/fl2 is the flag
1167 1167 in the current/other manifest. Where the file does not exist,
1168 1168 the nodeid will be None and the flags will be the empty
1169 1169 string.
1170 1170 '''
1171 1171 if match and not match.always():
1172 1172 m1 = self.matches(match)
1173 1173 m2 = m2.matches(match)
1174 1174 return m1.diff(m2, clean=clean)
1175 1175 result = {}
1176 1176 emptytree = treemanifest()
1177 1177
1178 1178 def _iterativediff(t1, t2, stack):
1179 1179 """compares two tree manifests and append new tree-manifests which
1180 1180 needs to be compared to stack"""
1181 1181 if t1._node == t2._node and not t1._dirty and not t2._dirty:
1182 1182 return
1183 1183 t1._load()
1184 1184 t2._load()
1185 1185 self._loaddifflazy(t1, t2)
1186 1186
1187 1187 for d, m1 in t1._dirs.iteritems():
1188 1188 m2 = t2._dirs.get(d, emptytree)
1189 1189 stack.append((m1, m2))
1190 1190
1191 1191 for d, m2 in t2._dirs.iteritems():
1192 1192 if d not in t1._dirs:
1193 1193 stack.append((emptytree, m2))
1194 1194
1195 1195 for fn, n1 in t1._files.iteritems():
1196 1196 fl1 = t1._flags.get(fn, '')
1197 1197 n2 = t2._files.get(fn, None)
1198 1198 fl2 = t2._flags.get(fn, '')
1199 1199 if n1 != n2 or fl1 != fl2:
1200 1200 result[t1._subpath(fn)] = ((n1, fl1), (n2, fl2))
1201 1201 elif clean:
1202 1202 result[t1._subpath(fn)] = None
1203 1203
1204 1204 for fn, n2 in t2._files.iteritems():
1205 1205 if fn not in t1._files:
1206 1206 fl2 = t2._flags.get(fn, '')
1207 1207 result[t2._subpath(fn)] = ((None, ''), (n2, fl2))
1208 1208
1209 1209 stackls = []
1210 1210 _iterativediff(self, m2, stackls)
1211 1211 while stackls:
1212 1212 t1, t2 = stackls.pop()
1213 1213 # stackls is populated in the function call
1214 1214 _iterativediff(t1, t2, stackls)
1215 1215 return result
1216 1216
1217 1217 def unmodifiedsince(self, m2):
1218 1218 return not self._dirty and not m2._dirty and self._node == m2._node
1219 1219
1220 1220 def parse(self, text, readsubtree):
1221 1221 selflazy = self._lazydirs
1222 1222 subpath = self._subpath
1223 1223 for f, n, fl in _parse(text):
1224 1224 if fl == 't':
1225 1225 f = f + '/'
1226 1226 # False below means "doesn't need to be copied" and can use the
1227 1227 # cached value from readsubtree directly.
1228 1228 selflazy[f] = (subpath(f), n, readsubtree, False)
1229 1229 elif '/' in f:
1230 1230 # This is a flat manifest, so use __setitem__ and setflag rather
1231 1231 # than assigning directly to _files and _flags, so we can
1232 1232 # assign a path in a subdirectory, and to mark dirty (compared
1233 1233 # to nullid).
1234 1234 self[f] = n
1235 1235 if fl:
1236 1236 self.setflag(f, fl)
1237 1237 else:
1238 1238 # Assigning to _files and _flags avoids marking as dirty,
1239 1239 # and should be a little faster.
1240 1240 self._files[f] = n
1241 1241 if fl:
1242 1242 self._flags[f] = fl
1243 1243
1244 1244 def text(self):
1245 1245 """Get the full data of this manifest as a bytestring."""
1246 1246 self._load()
1247 1247 return _text(self.iterentries())
1248 1248
1249 1249 def dirtext(self):
1250 1250 """Get the full data of this directory as a bytestring. Make sure that
1251 1251 any submanifests have been written first, so their nodeids are correct.
1252 1252 """
1253 1253 self._load()
1254 1254 flags = self.flags
1255 1255 lazydirs = [(d[:-1], v[1], 't') for d, v in self._lazydirs.iteritems()]
1256 1256 dirs = [(d[:-1], self._dirs[d]._node, 't') for d in self._dirs]
1257 1257 files = [(f, self._files[f], flags(f)) for f in self._files]
1258 1258 return _text(sorted(dirs + files + lazydirs))
1259 1259
1260 1260 def read(self, gettext, readsubtree):
1261 1261 def _load_for_read(s):
1262 1262 s.parse(gettext(), readsubtree)
1263 1263 s._dirty = False
1264 1264 self._loadfunc = _load_for_read
1265 1265
1266 1266 def writesubtrees(self, m1, m2, writesubtree, match):
1267 1267 self._load() # for consistency; should never have any effect here
1268 1268 m1._load()
1269 1269 m2._load()
1270 1270 emptytree = treemanifest()
1271 1271 def getnode(m, d):
1272 1272 ld = m._lazydirs.get(d)
1273 1273 if ld:
1274 1274 return ld[1]
1275 1275 return m._dirs.get(d, emptytree)._node
1276 1276
1277 1277 # let's skip investigating things that `match` says we do not need.
1278 1278 visit = match.visitchildrenset(self._dir[:-1])
1279 1279 visit = self._loadchildrensetlazy(visit)
1280 1280 if visit == 'this' or visit == 'all':
1281 1281 visit = None
1282 1282 for d, subm in self._dirs.iteritems():
1283 1283 if visit and d[:-1] not in visit:
1284 1284 continue
1285 1285 subp1 = getnode(m1, d)
1286 1286 subp2 = getnode(m2, d)
1287 1287 if subp1 == nullid:
1288 1288 subp1, subp2 = subp2, subp1
1289 1289 writesubtree(subm, subp1, subp2, match)
1290 1290
1291 1291 def walksubtrees(self, matcher=None):
1292 1292 """Returns an iterator of the subtrees of this manifest, including this
1293 1293 manifest itself.
1294 1294
1295 1295 If `matcher` is provided, it only returns subtrees that match.
1296 1296 """
1297 1297 if matcher and not matcher.visitdir(self._dir[:-1]):
1298 1298 return
1299 1299 if not matcher or matcher(self._dir[:-1]):
1300 1300 yield self
1301 1301
1302 1302 self._load()
1303 1303 # OPT: use visitchildrenset to avoid loading everything.
1304 1304 self._loadalllazy()
1305 1305 for d, subm in self._dirs.iteritems():
1306 1306 for subtree in subm.walksubtrees(matcher=matcher):
1307 1307 yield subtree
1308 1308
1309 1309 class manifestfulltextcache(util.lrucachedict):
1310 1310 """File-backed LRU cache for the manifest cache
1311 1311
1312 1312 File consists of entries, up to EOF:
1313 1313
1314 1314 - 20 bytes node, 4 bytes length, <length> manifest data
1315 1315
1316 1316 These are written in reverse cache order (oldest to newest).
1317 1317
1318 1318 """
1319 1319
1320 1320 _file = 'manifestfulltextcache'
1321 1321
1322 1322 def __init__(self, max):
1323 1323 super(manifestfulltextcache, self).__init__(max)
1324 1324 self._dirty = False
1325 1325 self._read = False
1326 1326 self._opener = None
1327 1327
1328 1328 def read(self):
1329 1329 if self._read or self._opener is None:
1330 1330 return
1331 1331
1332 1332 try:
1333 1333 with self._opener(self._file) as fp:
1334 1334 set = super(manifestfulltextcache, self).__setitem__
1335 1335 # ignore trailing data, this is a cache, corruption is skipped
1336 1336 while True:
1337 1337 node = fp.read(20)
1338 1338 if len(node) < 20:
1339 1339 break
1340 1340 try:
1341 1341 size = struct.unpack('>L', fp.read(4))[0]
1342 1342 except struct.error:
1343 1343 break
1344 1344 value = bytearray(fp.read(size))
1345 1345 if len(value) != size:
1346 1346 break
1347 1347 set(node, value)
1348 1348 except IOError:
1349 1349 # the file is allowed to be missing
1350 1350 pass
1351 1351
1352 1352 self._read = True
1353 1353 self._dirty = False
1354 1354
1355 1355 def write(self):
1356 1356 if not self._dirty or self._opener is None:
1357 1357 return
1358 1358 # rotate backwards to the first used node
1359 1359 with self._opener(self._file, 'w', atomictemp=True, checkambig=True
1360 1360 ) as fp:
1361 1361 node = self._head.prev
1362 1362 while True:
1363 1363 if node.key in self._cache:
1364 1364 fp.write(node.key)
1365 1365 fp.write(struct.pack('>L', len(node.value)))
1366 1366 fp.write(node.value)
1367 1367 if node is self._head:
1368 1368 break
1369 1369 node = node.prev
1370 1370
1371 1371 def __len__(self):
1372 1372 if not self._read:
1373 1373 self.read()
1374 1374 return super(manifestfulltextcache, self).__len__()
1375 1375
1376 1376 def __contains__(self, k):
1377 1377 if not self._read:
1378 1378 self.read()
1379 1379 return super(manifestfulltextcache, self).__contains__(k)
1380 1380
1381 1381 def __iter__(self):
1382 1382 if not self._read:
1383 1383 self.read()
1384 1384 return super(manifestfulltextcache, self).__iter__()
1385 1385
1386 1386 def __getitem__(self, k):
1387 1387 if not self._read:
1388 1388 self.read()
1389 1389 # the cache lru order can change on read
1390 1390 setdirty = self._cache.get(k) is not self._head
1391 1391 value = super(manifestfulltextcache, self).__getitem__(k)
1392 1392 if setdirty:
1393 1393 self._dirty = True
1394 1394 return value
1395 1395
1396 1396 def __setitem__(self, k, v):
1397 1397 if not self._read:
1398 1398 self.read()
1399 1399 super(manifestfulltextcache, self).__setitem__(k, v)
1400 1400 self._dirty = True
1401 1401
1402 1402 def __delitem__(self, k):
1403 1403 if not self._read:
1404 1404 self.read()
1405 1405 super(manifestfulltextcache, self).__delitem__(k)
1406 1406 self._dirty = True
1407 1407
1408 1408 def get(self, k, default=None):
1409 1409 if not self._read:
1410 1410 self.read()
1411 1411 return super(manifestfulltextcache, self).get(k, default=default)
1412 1412
1413 1413 def clear(self, clear_persisted_data=False):
1414 1414 super(manifestfulltextcache, self).clear()
1415 1415 if clear_persisted_data:
1416 1416 self._dirty = True
1417 1417 self.write()
1418 1418 self._read = False
1419 1419
1420 1420 # and upper bound of what we expect from compression
1421 1421 # (real live value seems to be "3")
1422 1422 MAXCOMPRESSION = 3
1423 1423
1424 1424 @interfaceutil.implementer(repository.imanifeststorage)
1425 1425 class manifestrevlog(object):
1426 1426 '''A revlog that stores manifest texts. This is responsible for caching the
1427 1427 full-text manifest contents.
1428 1428 '''
1429 1429 def __init__(self, opener, tree='', dirlogcache=None, indexfile=None,
1430 1430 treemanifest=False):
1431 1431 """Constructs a new manifest revlog
1432 1432
1433 1433 `indexfile` - used by extensions to have two manifests at once, like
1434 1434 when transitioning between flatmanifeset and treemanifests.
1435 1435
1436 1436 `treemanifest` - used to indicate this is a tree manifest revlog. Opener
1437 1437 options can also be used to make this a tree manifest revlog. The opener
1438 1438 option takes precedence, so if it is set to True, we ignore whatever
1439 1439 value is passed in to the constructor.
1440 1440 """
1441 1441 # During normal operations, we expect to deal with not more than four
1442 1442 # revs at a time (such as during commit --amend). When rebasing large
1443 1443 # stacks of commits, the number can go up, hence the config knob below.
1444 1444 cachesize = 4
1445 1445 optiontreemanifest = False
1446 1446 opts = getattr(opener, 'options', None)
1447 1447 if opts is not None:
1448 1448 cachesize = opts.get('manifestcachesize', cachesize)
1449 1449 optiontreemanifest = opts.get('treemanifest', False)
1450 1450
1451 1451 self._treeondisk = optiontreemanifest or treemanifest
1452 1452
1453 1453 self._fulltextcache = manifestfulltextcache(cachesize)
1454 1454
1455 1455 if tree:
1456 1456 assert self._treeondisk, 'opts is %r' % opts
1457 1457
1458 1458 if indexfile is None:
1459 1459 indexfile = '00manifest.i'
1460 1460 if tree:
1461 1461 indexfile = "meta/" + tree + indexfile
1462 1462
1463 1463 self.tree = tree
1464 1464
1465 1465 # The dirlogcache is kept on the root manifest log
1466 1466 if tree:
1467 1467 self._dirlogcache = dirlogcache
1468 1468 else:
1469 1469 self._dirlogcache = {'': self}
1470 1470
1471 1471 self._revlog = revlog.revlog(opener, indexfile,
1472 1472 # only root indexfile is cached
1473 1473 checkambig=not bool(tree),
1474 1474 mmaplargeindex=True,
1475 1475 upperboundcomp=MAXCOMPRESSION)
1476 1476
1477 1477 self.index = self._revlog.index
1478 1478 self.version = self._revlog.version
1479 1479 self._generaldelta = self._revlog._generaldelta
1480 1480
1481 1481 def _setupmanifestcachehooks(self, repo):
1482 1482 """Persist the manifestfulltextcache on lock release"""
1483 1483 if not util.safehasattr(repo, '_wlockref'):
1484 1484 return
1485 1485
1486 1486 self._fulltextcache._opener = repo.wcachevfs
1487 1487 if repo._currentlock(repo._wlockref) is None:
1488 1488 return
1489 1489
1490 1490 reporef = weakref.ref(repo)
1491 1491 manifestrevlogref = weakref.ref(self)
1492 1492
1493 1493 def persistmanifestcache():
1494 1494 repo = reporef()
1495 1495 self = manifestrevlogref()
1496 1496 if repo is None or self is None:
1497 1497 return
1498 1498 if repo.manifestlog.getstorage(b'') is not self:
1499 1499 # there's a different manifest in play now, abort
1500 1500 return
1501 1501 self._fulltextcache.write()
1502 1502
1503 1503 repo._afterlock(persistmanifestcache)
1504 1504
1505 1505 @property
1506 1506 def fulltextcache(self):
1507 1507 return self._fulltextcache
1508 1508
1509 1509 def clearcaches(self, clear_persisted_data=False):
1510 1510 self._revlog.clearcaches()
1511 1511 self._fulltextcache.clear(clear_persisted_data=clear_persisted_data)
1512 1512 self._dirlogcache = {self.tree: self}
1513 1513
1514 1514 def dirlog(self, d):
1515 1515 if d:
1516 1516 assert self._treeondisk
1517 1517 if d not in self._dirlogcache:
1518 1518 mfrevlog = manifestrevlog(self.opener, d,
1519 1519 self._dirlogcache,
1520 1520 treemanifest=self._treeondisk)
1521 1521 self._dirlogcache[d] = mfrevlog
1522 1522 return self._dirlogcache[d]
1523 1523
1524 1524 def add(self, m, transaction, link, p1, p2, added, removed, readtree=None,
1525 1525 match=None):
1526 1526 if p1 in self.fulltextcache and util.safehasattr(m, 'fastdelta'):
1527 1527 # If our first parent is in the manifest cache, we can
1528 1528 # compute a delta here using properties we know about the
1529 1529 # manifest up-front, which may save time later for the
1530 1530 # revlog layer.
1531 1531
1532 1532 _checkforbidden(added)
1533 1533 # combine the changed lists into one sorted iterator
1534 1534 work = heapq.merge([(x, False) for x in sorted(added)],
1535 1535 [(x, True) for x in sorted(removed)])
1536 1536
1537 1537 arraytext, deltatext = m.fastdelta(self.fulltextcache[p1], work)
1538 1538 cachedelta = self._revlog.rev(p1), deltatext
1539 1539 text = util.buffer(arraytext)
1540 1540 n = self._revlog.addrevision(text, transaction, link, p1, p2,
1541 1541 cachedelta)
1542 1542 else:
1543 1543 # The first parent manifest isn't already loaded, so we'll
1544 1544 # just encode a fulltext of the manifest and pass that
1545 1545 # through to the revlog layer, and let it handle the delta
1546 1546 # process.
1547 1547 if self._treeondisk:
1548 1548 assert readtree, "readtree must be set for treemanifest writes"
1549 1549 assert match, "match must be specified for treemanifest writes"
1550 1550 m1 = readtree(self.tree, p1)
1551 1551 m2 = readtree(self.tree, p2)
1552 1552 n = self._addtree(m, transaction, link, m1, m2, readtree,
1553 1553 match=match)
1554 1554 arraytext = None
1555 1555 else:
1556 1556 text = m.text()
1557 1557 n = self._revlog.addrevision(text, transaction, link, p1, p2)
1558 1558 arraytext = bytearray(text)
1559 1559
1560 1560 if arraytext is not None:
1561 1561 self.fulltextcache[n] = arraytext
1562 1562
1563 1563 return n
1564 1564
1565 1565 def _addtree(self, m, transaction, link, m1, m2, readtree, match):
1566 1566 # If the manifest is unchanged compared to one parent,
1567 1567 # don't write a new revision
1568 1568 if self.tree != '' and (m.unmodifiedsince(m1) or m.unmodifiedsince(
1569 1569 m2)):
1570 1570 return m.node()
1571 1571 def writesubtree(subm, subp1, subp2, match):
1572 1572 sublog = self.dirlog(subm.dir())
1573 1573 sublog.add(subm, transaction, link, subp1, subp2, None, None,
1574 1574 readtree=readtree, match=match)
1575 1575 m.writesubtrees(m1, m2, writesubtree, match)
1576 1576 text = m.dirtext()
1577 1577 n = None
1578 1578 if self.tree != '':
1579 1579 # Double-check whether contents are unchanged to one parent
1580 1580 if text == m1.dirtext():
1581 1581 n = m1.node()
1582 1582 elif text == m2.dirtext():
1583 1583 n = m2.node()
1584 1584
1585 1585 if not n:
1586 1586 n = self._revlog.addrevision(text, transaction, link, m1.node(),
1587 1587 m2.node())
1588 1588
1589 1589 # Save nodeid so parent manifest can calculate its nodeid
1590 1590 m.setnode(n)
1591 1591 return n
1592 1592
1593 1593 def __len__(self):
1594 1594 return len(self._revlog)
1595 1595
1596 1596 def __iter__(self):
1597 1597 return self._revlog.__iter__()
1598 1598
1599 1599 def rev(self, node):
1600 1600 return self._revlog.rev(node)
1601 1601
1602 1602 def node(self, rev):
1603 1603 return self._revlog.node(rev)
1604 1604
1605 1605 def lookup(self, value):
1606 1606 return self._revlog.lookup(value)
1607 1607
1608 1608 def parentrevs(self, rev):
1609 1609 return self._revlog.parentrevs(rev)
1610 1610
1611 1611 def parents(self, node):
1612 1612 return self._revlog.parents(node)
1613 1613
1614 1614 def linkrev(self, rev):
1615 1615 return self._revlog.linkrev(rev)
1616 1616
1617 1617 def checksize(self):
1618 1618 return self._revlog.checksize()
1619 1619
1620 1620 def revision(self, node, _df=None, raw=False):
1621 1621 return self._revlog.revision(node, _df=_df, raw=raw)
1622 1622
1623 def rawdata(self, node, _df=None):
1624 return self._revlog.rawdata(node, _df=_df)
1625
1623 1626 def revdiff(self, rev1, rev2):
1624 1627 return self._revlog.revdiff(rev1, rev2)
1625 1628
1626 1629 def cmp(self, node, text):
1627 1630 return self._revlog.cmp(node, text)
1628 1631
1629 1632 def deltaparent(self, rev):
1630 1633 return self._revlog.deltaparent(rev)
1631 1634
1632 1635 def emitrevisions(self, nodes, nodesorder=None,
1633 1636 revisiondata=False, assumehaveparentrevisions=False,
1634 1637 deltamode=repository.CG_DELTAMODE_STD):
1635 1638 return self._revlog.emitrevisions(
1636 1639 nodes, nodesorder=nodesorder, revisiondata=revisiondata,
1637 1640 assumehaveparentrevisions=assumehaveparentrevisions,
1638 1641 deltamode=deltamode)
1639 1642
1640 1643 def addgroup(self, deltas, linkmapper, transaction, addrevisioncb=None):
1641 1644 return self._revlog.addgroup(deltas, linkmapper, transaction,
1642 1645 addrevisioncb=addrevisioncb)
1643 1646
1644 1647 def rawsize(self, rev):
1645 1648 return self._revlog.rawsize(rev)
1646 1649
1647 1650 def getstrippoint(self, minlink):
1648 1651 return self._revlog.getstrippoint(minlink)
1649 1652
1650 1653 def strip(self, minlink, transaction):
1651 1654 return self._revlog.strip(minlink, transaction)
1652 1655
1653 1656 def files(self):
1654 1657 return self._revlog.files()
1655 1658
1656 1659 def clone(self, tr, destrevlog, **kwargs):
1657 1660 if not isinstance(destrevlog, manifestrevlog):
1658 1661 raise error.ProgrammingError('expected manifestrevlog to clone()')
1659 1662
1660 1663 return self._revlog.clone(tr, destrevlog._revlog, **kwargs)
1661 1664
1662 1665 def storageinfo(self, exclusivefiles=False, sharedfiles=False,
1663 1666 revisionscount=False, trackedsize=False,
1664 1667 storedsize=False):
1665 1668 return self._revlog.storageinfo(
1666 1669 exclusivefiles=exclusivefiles, sharedfiles=sharedfiles,
1667 1670 revisionscount=revisionscount, trackedsize=trackedsize,
1668 1671 storedsize=storedsize)
1669 1672
1670 1673 @property
1671 1674 def indexfile(self):
1672 1675 return self._revlog.indexfile
1673 1676
1674 1677 @indexfile.setter
1675 1678 def indexfile(self, value):
1676 1679 self._revlog.indexfile = value
1677 1680
1678 1681 @property
1679 1682 def opener(self):
1680 1683 return self._revlog.opener
1681 1684
1682 1685 @opener.setter
1683 1686 def opener(self, value):
1684 1687 self._revlog.opener = value
1685 1688
1686 1689 @interfaceutil.implementer(repository.imanifestlog)
1687 1690 class manifestlog(object):
1688 1691 """A collection class representing the collection of manifest snapshots
1689 1692 referenced by commits in the repository.
1690 1693
1691 1694 In this situation, 'manifest' refers to the abstract concept of a snapshot
1692 1695 of the list of files in the given commit. Consumers of the output of this
1693 1696 class do not care about the implementation details of the actual manifests
1694 1697 they receive (i.e. tree or flat or lazily loaded, etc)."""
1695 1698 def __init__(self, opener, repo, rootstore, narrowmatch):
1696 1699 usetreemanifest = False
1697 1700 cachesize = 4
1698 1701
1699 1702 opts = getattr(opener, 'options', None)
1700 1703 if opts is not None:
1701 1704 usetreemanifest = opts.get('treemanifest', usetreemanifest)
1702 1705 cachesize = opts.get('manifestcachesize', cachesize)
1703 1706
1704 1707 self._treemanifests = usetreemanifest
1705 1708
1706 1709 self._rootstore = rootstore
1707 1710 self._rootstore._setupmanifestcachehooks(repo)
1708 1711 self._narrowmatch = narrowmatch
1709 1712
1710 1713 # A cache of the manifestctx or treemanifestctx for each directory
1711 1714 self._dirmancache = {}
1712 1715 self._dirmancache[''] = util.lrucachedict(cachesize)
1713 1716
1714 1717 self._cachesize = cachesize
1715 1718
1716 1719 def __getitem__(self, node):
1717 1720 """Retrieves the manifest instance for the given node. Throws a
1718 1721 LookupError if not found.
1719 1722 """
1720 1723 return self.get('', node)
1721 1724
1722 1725 def get(self, tree, node, verify=True):
1723 1726 """Retrieves the manifest instance for the given node. Throws a
1724 1727 LookupError if not found.
1725 1728
1726 1729 `verify` - if True an exception will be thrown if the node is not in
1727 1730 the revlog
1728 1731 """
1729 1732 if node in self._dirmancache.get(tree, ()):
1730 1733 return self._dirmancache[tree][node]
1731 1734
1732 1735 if not self._narrowmatch.always():
1733 1736 if not self._narrowmatch.visitdir(tree[:-1]):
1734 1737 return excludeddirmanifestctx(tree, node)
1735 1738 if tree:
1736 1739 if self._rootstore._treeondisk:
1737 1740 if verify:
1738 1741 # Side-effect is LookupError is raised if node doesn't
1739 1742 # exist.
1740 1743 self.getstorage(tree).rev(node)
1741 1744
1742 1745 m = treemanifestctx(self, tree, node)
1743 1746 else:
1744 1747 raise error.Abort(
1745 1748 _("cannot ask for manifest directory '%s' in a flat "
1746 1749 "manifest") % tree)
1747 1750 else:
1748 1751 if verify:
1749 1752 # Side-effect is LookupError is raised if node doesn't exist.
1750 1753 self._rootstore.rev(node)
1751 1754
1752 1755 if self._treemanifests:
1753 1756 m = treemanifestctx(self, '', node)
1754 1757 else:
1755 1758 m = manifestctx(self, node)
1756 1759
1757 1760 if node != nullid:
1758 1761 mancache = self._dirmancache.get(tree)
1759 1762 if not mancache:
1760 1763 mancache = util.lrucachedict(self._cachesize)
1761 1764 self._dirmancache[tree] = mancache
1762 1765 mancache[node] = m
1763 1766 return m
1764 1767
1765 1768 def getstorage(self, tree):
1766 1769 return self._rootstore.dirlog(tree)
1767 1770
1768 1771 def clearcaches(self, clear_persisted_data=False):
1769 1772 self._dirmancache.clear()
1770 1773 self._rootstore.clearcaches(clear_persisted_data=clear_persisted_data)
1771 1774
1772 1775 def rev(self, node):
1773 1776 return self._rootstore.rev(node)
1774 1777
1775 1778 @interfaceutil.implementer(repository.imanifestrevisionwritable)
1776 1779 class memmanifestctx(object):
1777 1780 def __init__(self, manifestlog):
1778 1781 self._manifestlog = manifestlog
1779 1782 self._manifestdict = manifestdict()
1780 1783
1781 1784 def _storage(self):
1782 1785 return self._manifestlog.getstorage(b'')
1783 1786
1784 1787 def new(self):
1785 1788 return memmanifestctx(self._manifestlog)
1786 1789
1787 1790 def copy(self):
1788 1791 memmf = memmanifestctx(self._manifestlog)
1789 1792 memmf._manifestdict = self.read().copy()
1790 1793 return memmf
1791 1794
1792 1795 def read(self):
1793 1796 return self._manifestdict
1794 1797
1795 1798 def write(self, transaction, link, p1, p2, added, removed, match=None):
1796 1799 return self._storage().add(self._manifestdict, transaction, link,
1797 1800 p1, p2, added, removed, match=match)
1798 1801
1799 1802 @interfaceutil.implementer(repository.imanifestrevisionstored)
1800 1803 class manifestctx(object):
1801 1804 """A class representing a single revision of a manifest, including its
1802 1805 contents, its parent revs, and its linkrev.
1803 1806 """
1804 1807 def __init__(self, manifestlog, node):
1805 1808 self._manifestlog = manifestlog
1806 1809 self._data = None
1807 1810
1808 1811 self._node = node
1809 1812
1810 1813 # TODO: We eventually want p1, p2, and linkrev exposed on this class,
1811 1814 # but let's add it later when something needs it and we can load it
1812 1815 # lazily.
1813 1816 #self.p1, self.p2 = store.parents(node)
1814 1817 #rev = store.rev(node)
1815 1818 #self.linkrev = store.linkrev(rev)
1816 1819
1817 1820 def _storage(self):
1818 1821 return self._manifestlog.getstorage(b'')
1819 1822
1820 1823 def node(self):
1821 1824 return self._node
1822 1825
1823 1826 def new(self):
1824 1827 return memmanifestctx(self._manifestlog)
1825 1828
1826 1829 def copy(self):
1827 1830 memmf = memmanifestctx(self._manifestlog)
1828 1831 memmf._manifestdict = self.read().copy()
1829 1832 return memmf
1830 1833
1831 1834 @propertycache
1832 1835 def parents(self):
1833 1836 return self._storage().parents(self._node)
1834 1837
1835 1838 def read(self):
1836 1839 if self._data is None:
1837 1840 if self._node == nullid:
1838 1841 self._data = manifestdict()
1839 1842 else:
1840 1843 store = self._storage()
1841 1844 if self._node in store.fulltextcache:
1842 1845 text = pycompat.bytestr(store.fulltextcache[self._node])
1843 1846 else:
1844 1847 text = store.revision(self._node)
1845 1848 arraytext = bytearray(text)
1846 1849 store.fulltextcache[self._node] = arraytext
1847 1850 self._data = manifestdict(text)
1848 1851 return self._data
1849 1852
1850 1853 def readfast(self, shallow=False):
1851 1854 '''Calls either readdelta or read, based on which would be less work.
1852 1855 readdelta is called if the delta is against the p1, and therefore can be
1853 1856 read quickly.
1854 1857
1855 1858 If `shallow` is True, nothing changes since this is a flat manifest.
1856 1859 '''
1857 1860 store = self._storage()
1858 1861 r = store.rev(self._node)
1859 1862 deltaparent = store.deltaparent(r)
1860 1863 if deltaparent != nullrev and deltaparent in store.parentrevs(r):
1861 1864 return self.readdelta()
1862 1865 return self.read()
1863 1866
1864 1867 def readdelta(self, shallow=False):
1865 1868 '''Returns a manifest containing just the entries that are present
1866 1869 in this manifest, but not in its p1 manifest. This is efficient to read
1867 1870 if the revlog delta is already p1.
1868 1871
1869 1872 Changing the value of `shallow` has no effect on flat manifests.
1870 1873 '''
1871 1874 store = self._storage()
1872 1875 r = store.rev(self._node)
1873 1876 d = mdiff.patchtext(store.revdiff(store.deltaparent(r), r))
1874 1877 return manifestdict(d)
1875 1878
1876 1879 def find(self, key):
1877 1880 return self.read().find(key)
1878 1881
1879 1882 @interfaceutil.implementer(repository.imanifestrevisionwritable)
1880 1883 class memtreemanifestctx(object):
1881 1884 def __init__(self, manifestlog, dir=''):
1882 1885 self._manifestlog = manifestlog
1883 1886 self._dir = dir
1884 1887 self._treemanifest = treemanifest()
1885 1888
1886 1889 def _storage(self):
1887 1890 return self._manifestlog.getstorage(b'')
1888 1891
1889 1892 def new(self, dir=''):
1890 1893 return memtreemanifestctx(self._manifestlog, dir=dir)
1891 1894
1892 1895 def copy(self):
1893 1896 memmf = memtreemanifestctx(self._manifestlog, dir=self._dir)
1894 1897 memmf._treemanifest = self._treemanifest.copy()
1895 1898 return memmf
1896 1899
1897 1900 def read(self):
1898 1901 return self._treemanifest
1899 1902
1900 1903 def write(self, transaction, link, p1, p2, added, removed, match=None):
1901 1904 def readtree(dir, node):
1902 1905 return self._manifestlog.get(dir, node).read()
1903 1906 return self._storage().add(self._treemanifest, transaction, link,
1904 1907 p1, p2, added, removed, readtree=readtree,
1905 1908 match=match)
1906 1909
1907 1910 @interfaceutil.implementer(repository.imanifestrevisionstored)
1908 1911 class treemanifestctx(object):
1909 1912 def __init__(self, manifestlog, dir, node):
1910 1913 self._manifestlog = manifestlog
1911 1914 self._dir = dir
1912 1915 self._data = None
1913 1916
1914 1917 self._node = node
1915 1918
1916 1919 # TODO: Load p1/p2/linkrev lazily. They need to be lazily loaded so that
1917 1920 # we can instantiate treemanifestctx objects for directories we don't
1918 1921 # have on disk.
1919 1922 #self.p1, self.p2 = store.parents(node)
1920 1923 #rev = store.rev(node)
1921 1924 #self.linkrev = store.linkrev(rev)
1922 1925
1923 1926 def _storage(self):
1924 1927 narrowmatch = self._manifestlog._narrowmatch
1925 1928 if not narrowmatch.always():
1926 1929 if not narrowmatch.visitdir(self._dir[:-1]):
1927 1930 return excludedmanifestrevlog(self._dir)
1928 1931 return self._manifestlog.getstorage(self._dir)
1929 1932
1930 1933 def read(self):
1931 1934 if self._data is None:
1932 1935 store = self._storage()
1933 1936 if self._node == nullid:
1934 1937 self._data = treemanifest()
1935 1938 # TODO accessing non-public API
1936 1939 elif store._treeondisk:
1937 1940 m = treemanifest(dir=self._dir)
1938 1941 def gettext():
1939 1942 return store.revision(self._node)
1940 1943 def readsubtree(dir, subm):
1941 1944 # Set verify to False since we need to be able to create
1942 1945 # subtrees for trees that don't exist on disk.
1943 1946 return self._manifestlog.get(dir, subm, verify=False).read()
1944 1947 m.read(gettext, readsubtree)
1945 1948 m.setnode(self._node)
1946 1949 self._data = m
1947 1950 else:
1948 1951 if self._node in store.fulltextcache:
1949 1952 text = pycompat.bytestr(store.fulltextcache[self._node])
1950 1953 else:
1951 1954 text = store.revision(self._node)
1952 1955 arraytext = bytearray(text)
1953 1956 store.fulltextcache[self._node] = arraytext
1954 1957 self._data = treemanifest(dir=self._dir, text=text)
1955 1958
1956 1959 return self._data
1957 1960
1958 1961 def node(self):
1959 1962 return self._node
1960 1963
1961 1964 def new(self, dir=''):
1962 1965 return memtreemanifestctx(self._manifestlog, dir=dir)
1963 1966
1964 1967 def copy(self):
1965 1968 memmf = memtreemanifestctx(self._manifestlog, dir=self._dir)
1966 1969 memmf._treemanifest = self.read().copy()
1967 1970 return memmf
1968 1971
1969 1972 @propertycache
1970 1973 def parents(self):
1971 1974 return self._storage().parents(self._node)
1972 1975
1973 1976 def readdelta(self, shallow=False):
1974 1977 '''Returns a manifest containing just the entries that are present
1975 1978 in this manifest, but not in its p1 manifest. This is efficient to read
1976 1979 if the revlog delta is already p1.
1977 1980
1978 1981 If `shallow` is True, this will read the delta for this directory,
1979 1982 without recursively reading subdirectory manifests. Instead, any
1980 1983 subdirectory entry will be reported as it appears in the manifest, i.e.
1981 1984 the subdirectory will be reported among files and distinguished only by
1982 1985 its 't' flag.
1983 1986 '''
1984 1987 store = self._storage()
1985 1988 if shallow:
1986 1989 r = store.rev(self._node)
1987 1990 d = mdiff.patchtext(store.revdiff(store.deltaparent(r), r))
1988 1991 return manifestdict(d)
1989 1992 else:
1990 1993 # Need to perform a slow delta
1991 1994 r0 = store.deltaparent(store.rev(self._node))
1992 1995 m0 = self._manifestlog.get(self._dir, store.node(r0)).read()
1993 1996 m1 = self.read()
1994 1997 md = treemanifest(dir=self._dir)
1995 1998 for f, ((n0, fl0), (n1, fl1)) in m0.diff(m1).iteritems():
1996 1999 if n1:
1997 2000 md[f] = n1
1998 2001 if fl1:
1999 2002 md.setflag(f, fl1)
2000 2003 return md
2001 2004
2002 2005 def readfast(self, shallow=False):
2003 2006 '''Calls either readdelta or read, based on which would be less work.
2004 2007 readdelta is called if the delta is against the p1, and therefore can be
2005 2008 read quickly.
2006 2009
2007 2010 If `shallow` is True, it only returns the entries from this manifest,
2008 2011 and not any submanifests.
2009 2012 '''
2010 2013 store = self._storage()
2011 2014 r = store.rev(self._node)
2012 2015 deltaparent = store.deltaparent(r)
2013 2016 if (deltaparent != nullrev and
2014 2017 deltaparent in store.parentrevs(r)):
2015 2018 return self.readdelta(shallow=shallow)
2016 2019
2017 2020 if shallow:
2018 2021 return manifestdict(store.revision(self._node))
2019 2022 else:
2020 2023 return self.read()
2021 2024
2022 2025 def find(self, key):
2023 2026 return self.read().find(key)
2024 2027
2025 2028 class excludeddir(treemanifest):
2026 2029 """Stand-in for a directory that is excluded from the repository.
2027 2030
2028 2031 With narrowing active on a repository that uses treemanifests,
2029 2032 some of the directory revlogs will be excluded from the resulting
2030 2033 clone. This is a huge storage win for clients, but means we need
2031 2034 some sort of pseudo-manifest to surface to internals so we can
2032 2035 detect a merge conflict outside the narrowspec. That's what this
2033 2036 class is: it stands in for a directory whose node is known, but
2034 2037 whose contents are unknown.
2035 2038 """
2036 2039 def __init__(self, dir, node):
2037 2040 super(excludeddir, self).__init__(dir)
2038 2041 self._node = node
2039 2042 # Add an empty file, which will be included by iterators and such,
2040 2043 # appearing as the directory itself (i.e. something like "dir/")
2041 2044 self._files[''] = node
2042 2045 self._flags[''] = 't'
2043 2046
2044 2047 # Manifests outside the narrowspec should never be modified, so avoid
2045 2048 # copying. This makes a noticeable difference when there are very many
2046 2049 # directories outside the narrowspec. Also, it makes sense for the copy to
2047 2050 # be of the same type as the original, which would not happen with the
2048 2051 # super type's copy().
2049 2052 def copy(self):
2050 2053 return self
2051 2054
2052 2055 class excludeddirmanifestctx(treemanifestctx):
2053 2056 """context wrapper for excludeddir - see that docstring for rationale"""
2054 2057 def __init__(self, dir, node):
2055 2058 self._dir = dir
2056 2059 self._node = node
2057 2060
2058 2061 def read(self):
2059 2062 return excludeddir(self._dir, self._node)
2060 2063
2061 2064 def write(self, *args):
2062 2065 raise error.ProgrammingError(
2063 2066 'attempt to write manifest from excluded dir %s' % self._dir)
2064 2067
2065 2068 class excludedmanifestrevlog(manifestrevlog):
2066 2069 """Stand-in for excluded treemanifest revlogs.
2067 2070
2068 2071 When narrowing is active on a treemanifest repository, we'll have
2069 2072 references to directories we can't see due to the revlog being
2070 2073 skipped. This class exists to conform to the manifestrevlog
2071 2074 interface for those directories and proactively prevent writes to
2072 2075 outside the narrowspec.
2073 2076 """
2074 2077
2075 2078 def __init__(self, dir):
2076 2079 self._dir = dir
2077 2080
2078 2081 def __len__(self):
2079 2082 raise error.ProgrammingError(
2080 2083 'attempt to get length of excluded dir %s' % self._dir)
2081 2084
2082 2085 def rev(self, node):
2083 2086 raise error.ProgrammingError(
2084 2087 'attempt to get rev from excluded dir %s' % self._dir)
2085 2088
2086 2089 def linkrev(self, node):
2087 2090 raise error.ProgrammingError(
2088 2091 'attempt to get linkrev from excluded dir %s' % self._dir)
2089 2092
2090 2093 def node(self, rev):
2091 2094 raise error.ProgrammingError(
2092 2095 'attempt to get node from excluded dir %s' % self._dir)
2093 2096
2094 2097 def add(self, *args, **kwargs):
2095 2098 # We should never write entries in dirlogs outside the narrow clone.
2096 2099 # However, the method still gets called from writesubtree() in
2097 2100 # _addtree(), so we need to handle it. We should possibly make that
2098 2101 # avoid calling add() with a clean manifest (_dirty is always False
2099 2102 # in excludeddir instances).
2100 2103 pass
@@ -1,1870 +1,1873
1 1 # repository.py - Interfaces and base classes for repositories and peers.
2 2 #
3 3 # Copyright 2017 Gregory Szorc <gregory.szorc@gmail.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 from __future__ import absolute_import
9 9
10 10 from .i18n import _
11 11 from . import (
12 12 error,
13 13 )
14 14 from .utils import (
15 15 interfaceutil,
16 16 )
17 17
18 18 # When narrowing is finalized and no longer subject to format changes,
19 19 # we should move this to just "narrow" or similar.
20 20 NARROW_REQUIREMENT = 'narrowhg-experimental'
21 21
22 22 # Local repository feature string.
23 23
24 24 # Revlogs are being used for file storage.
25 25 REPO_FEATURE_REVLOG_FILE_STORAGE = b'revlogfilestorage'
26 26 # The storage part of the repository is shared from an external source.
27 27 REPO_FEATURE_SHARED_STORAGE = b'sharedstore'
28 28 # LFS supported for backing file storage.
29 29 REPO_FEATURE_LFS = b'lfs'
30 30 # Repository supports being stream cloned.
31 31 REPO_FEATURE_STREAM_CLONE = b'streamclone'
32 32 # Files storage may lack data for all ancestors.
33 33 REPO_FEATURE_SHALLOW_FILE_STORAGE = b'shallowfilestorage'
34 34
35 35 REVISION_FLAG_CENSORED = 1 << 15
36 36 REVISION_FLAG_ELLIPSIS = 1 << 14
37 37 REVISION_FLAG_EXTSTORED = 1 << 13
38 38
39 39 REVISION_FLAGS_KNOWN = (
40 40 REVISION_FLAG_CENSORED | REVISION_FLAG_ELLIPSIS | REVISION_FLAG_EXTSTORED)
41 41
42 42 CG_DELTAMODE_STD = b'default'
43 43 CG_DELTAMODE_PREV = b'previous'
44 44 CG_DELTAMODE_FULL = b'fulltext'
45 45 CG_DELTAMODE_P1 = b'p1'
46 46
47 47 class ipeerconnection(interfaceutil.Interface):
48 48 """Represents a "connection" to a repository.
49 49
50 50 This is the base interface for representing a connection to a repository.
51 51 It holds basic properties and methods applicable to all peer types.
52 52
53 53 This is not a complete interface definition and should not be used
54 54 outside of this module.
55 55 """
56 56 ui = interfaceutil.Attribute("""ui.ui instance""")
57 57
58 58 def url():
59 59 """Returns a URL string representing this peer.
60 60
61 61 Currently, implementations expose the raw URL used to construct the
62 62 instance. It may contain credentials as part of the URL. The
63 63 expectations of the value aren't well-defined and this could lead to
64 64 data leakage.
65 65
66 66 TODO audit/clean consumers and more clearly define the contents of this
67 67 value.
68 68 """
69 69
70 70 def local():
71 71 """Returns a local repository instance.
72 72
73 73 If the peer represents a local repository, returns an object that
74 74 can be used to interface with it. Otherwise returns ``None``.
75 75 """
76 76
77 77 def peer():
78 78 """Returns an object conforming to this interface.
79 79
80 80 Most implementations will ``return self``.
81 81 """
82 82
83 83 def canpush():
84 84 """Returns a boolean indicating if this peer can be pushed to."""
85 85
86 86 def close():
87 87 """Close the connection to this peer.
88 88
89 89 This is called when the peer will no longer be used. Resources
90 90 associated with the peer should be cleaned up.
91 91 """
92 92
93 93 class ipeercapabilities(interfaceutil.Interface):
94 94 """Peer sub-interface related to capabilities."""
95 95
96 96 def capable(name):
97 97 """Determine support for a named capability.
98 98
99 99 Returns ``False`` if capability not supported.
100 100
101 101 Returns ``True`` if boolean capability is supported. Returns a string
102 102 if capability support is non-boolean.
103 103
104 104 Capability strings may or may not map to wire protocol capabilities.
105 105 """
106 106
107 107 def requirecap(name, purpose):
108 108 """Require a capability to be present.
109 109
110 110 Raises a ``CapabilityError`` if the capability isn't present.
111 111 """
112 112
113 113 class ipeercommands(interfaceutil.Interface):
114 114 """Client-side interface for communicating over the wire protocol.
115 115
116 116 This interface is used as a gateway to the Mercurial wire protocol.
117 117 methods commonly call wire protocol commands of the same name.
118 118 """
119 119
120 120 def branchmap():
121 121 """Obtain heads in named branches.
122 122
123 123 Returns a dict mapping branch name to an iterable of nodes that are
124 124 heads on that branch.
125 125 """
126 126
127 127 def capabilities():
128 128 """Obtain capabilities of the peer.
129 129
130 130 Returns a set of string capabilities.
131 131 """
132 132
133 133 def clonebundles():
134 134 """Obtains the clone bundles manifest for the repo.
135 135
136 136 Returns the manifest as unparsed bytes.
137 137 """
138 138
139 139 def debugwireargs(one, two, three=None, four=None, five=None):
140 140 """Used to facilitate debugging of arguments passed over the wire."""
141 141
142 142 def getbundle(source, **kwargs):
143 143 """Obtain remote repository data as a bundle.
144 144
145 145 This command is how the bulk of repository data is transferred from
146 146 the peer to the local repository
147 147
148 148 Returns a generator of bundle data.
149 149 """
150 150
151 151 def heads():
152 152 """Determine all known head revisions in the peer.
153 153
154 154 Returns an iterable of binary nodes.
155 155 """
156 156
157 157 def known(nodes):
158 158 """Determine whether multiple nodes are known.
159 159
160 160 Accepts an iterable of nodes whose presence to check for.
161 161
162 162 Returns an iterable of booleans indicating of the corresponding node
163 163 at that index is known to the peer.
164 164 """
165 165
166 166 def listkeys(namespace):
167 167 """Obtain all keys in a pushkey namespace.
168 168
169 169 Returns an iterable of key names.
170 170 """
171 171
172 172 def lookup(key):
173 173 """Resolve a value to a known revision.
174 174
175 175 Returns a binary node of the resolved revision on success.
176 176 """
177 177
178 178 def pushkey(namespace, key, old, new):
179 179 """Set a value using the ``pushkey`` protocol.
180 180
181 181 Arguments correspond to the pushkey namespace and key to operate on and
182 182 the old and new values for that key.
183 183
184 184 Returns a string with the peer result. The value inside varies by the
185 185 namespace.
186 186 """
187 187
188 188 def stream_out():
189 189 """Obtain streaming clone data.
190 190
191 191 Successful result should be a generator of data chunks.
192 192 """
193 193
194 194 def unbundle(bundle, heads, url):
195 195 """Transfer repository data to the peer.
196 196
197 197 This is how the bulk of data during a push is transferred.
198 198
199 199 Returns the integer number of heads added to the peer.
200 200 """
201 201
202 202 class ipeerlegacycommands(interfaceutil.Interface):
203 203 """Interface for implementing support for legacy wire protocol commands.
204 204
205 205 Wire protocol commands transition to legacy status when they are no longer
206 206 used by modern clients. To facilitate identifying which commands are
207 207 legacy, the interfaces are split.
208 208 """
209 209
210 210 def between(pairs):
211 211 """Obtain nodes between pairs of nodes.
212 212
213 213 ``pairs`` is an iterable of node pairs.
214 214
215 215 Returns an iterable of iterables of nodes corresponding to each
216 216 requested pair.
217 217 """
218 218
219 219 def branches(nodes):
220 220 """Obtain ancestor changesets of specific nodes back to a branch point.
221 221
222 222 For each requested node, the peer finds the first ancestor node that is
223 223 a DAG root or is a merge.
224 224
225 225 Returns an iterable of iterables with the resolved values for each node.
226 226 """
227 227
228 228 def changegroup(nodes, source):
229 229 """Obtain a changegroup with data for descendants of specified nodes."""
230 230
231 231 def changegroupsubset(bases, heads, source):
232 232 pass
233 233
234 234 class ipeercommandexecutor(interfaceutil.Interface):
235 235 """Represents a mechanism to execute remote commands.
236 236
237 237 This is the primary interface for requesting that wire protocol commands
238 238 be executed. Instances of this interface are active in a context manager
239 239 and have a well-defined lifetime. When the context manager exits, all
240 240 outstanding requests are waited on.
241 241 """
242 242
243 243 def callcommand(name, args):
244 244 """Request that a named command be executed.
245 245
246 246 Receives the command name and a dictionary of command arguments.
247 247
248 248 Returns a ``concurrent.futures.Future`` that will resolve to the
249 249 result of that command request. That exact value is left up to
250 250 the implementation and possibly varies by command.
251 251
252 252 Not all commands can coexist with other commands in an executor
253 253 instance: it depends on the underlying wire protocol transport being
254 254 used and the command itself.
255 255
256 256 Implementations MAY call ``sendcommands()`` automatically if the
257 257 requested command can not coexist with other commands in this executor.
258 258
259 259 Implementations MAY call ``sendcommands()`` automatically when the
260 260 future's ``result()`` is called. So, consumers using multiple
261 261 commands with an executor MUST ensure that ``result()`` is not called
262 262 until all command requests have been issued.
263 263 """
264 264
265 265 def sendcommands():
266 266 """Trigger submission of queued command requests.
267 267
268 268 Not all transports submit commands as soon as they are requested to
269 269 run. When called, this method forces queued command requests to be
270 270 issued. It will no-op if all commands have already been sent.
271 271
272 272 When called, no more new commands may be issued with this executor.
273 273 """
274 274
275 275 def close():
276 276 """Signal that this command request is finished.
277 277
278 278 When called, no more new commands may be issued. All outstanding
279 279 commands that have previously been issued are waited on before
280 280 returning. This not only includes waiting for the futures to resolve,
281 281 but also waiting for all response data to arrive. In other words,
282 282 calling this waits for all on-wire state for issued command requests
283 283 to finish.
284 284
285 285 When used as a context manager, this method is called when exiting the
286 286 context manager.
287 287
288 288 This method may call ``sendcommands()`` if there are buffered commands.
289 289 """
290 290
291 291 class ipeerrequests(interfaceutil.Interface):
292 292 """Interface for executing commands on a peer."""
293 293
294 294 limitedarguments = interfaceutil.Attribute(
295 295 """True if the peer cannot receive large argument value for commands."""
296 296 )
297 297
298 298 def commandexecutor():
299 299 """A context manager that resolves to an ipeercommandexecutor.
300 300
301 301 The object this resolves to can be used to issue command requests
302 302 to the peer.
303 303
304 304 Callers should call its ``callcommand`` method to issue command
305 305 requests.
306 306
307 307 A new executor should be obtained for each distinct set of commands
308 308 (possibly just a single command) that the consumer wants to execute
309 309 as part of a single operation or round trip. This is because some
310 310 peers are half-duplex and/or don't support persistent connections.
311 311 e.g. in the case of HTTP peers, commands sent to an executor represent
312 312 a single HTTP request. While some peers may support multiple command
313 313 sends over the wire per executor, consumers need to code to the least
314 314 capable peer. So it should be assumed that command executors buffer
315 315 called commands until they are told to send them and that each
316 316 command executor could result in a new connection or wire-level request
317 317 being issued.
318 318 """
319 319
320 320 class ipeerbase(ipeerconnection, ipeercapabilities, ipeerrequests):
321 321 """Unified interface for peer repositories.
322 322
323 323 All peer instances must conform to this interface.
324 324 """
325 325
326 326 class ipeerv2(ipeerconnection, ipeercapabilities, ipeerrequests):
327 327 """Unified peer interface for wire protocol version 2 peers."""
328 328
329 329 apidescriptor = interfaceutil.Attribute(
330 330 """Data structure holding description of server API.""")
331 331
332 332 @interfaceutil.implementer(ipeerbase)
333 333 class peer(object):
334 334 """Base class for peer repositories."""
335 335
336 336 limitedarguments = False
337 337
338 338 def capable(self, name):
339 339 caps = self.capabilities()
340 340 if name in caps:
341 341 return True
342 342
343 343 name = '%s=' % name
344 344 for cap in caps:
345 345 if cap.startswith(name):
346 346 return cap[len(name):]
347 347
348 348 return False
349 349
350 350 def requirecap(self, name, purpose):
351 351 if self.capable(name):
352 352 return
353 353
354 354 raise error.CapabilityError(
355 355 _('cannot %s; remote repository does not support the '
356 356 '\'%s\' capability') % (purpose, name))
357 357
358 358 class iverifyproblem(interfaceutil.Interface):
359 359 """Represents a problem with the integrity of the repository.
360 360
361 361 Instances of this interface are emitted to describe an integrity issue
362 362 with a repository (e.g. corrupt storage, missing data, etc).
363 363
364 364 Instances are essentially messages associated with severity.
365 365 """
366 366 warning = interfaceutil.Attribute(
367 367 """Message indicating a non-fatal problem.""")
368 368
369 369 error = interfaceutil.Attribute(
370 370 """Message indicating a fatal problem.""")
371 371
372 372 node = interfaceutil.Attribute(
373 373 """Revision encountering the problem.
374 374
375 375 ``None`` means the problem doesn't apply to a single revision.
376 376 """)
377 377
378 378 class irevisiondelta(interfaceutil.Interface):
379 379 """Represents a delta between one revision and another.
380 380
381 381 Instances convey enough information to allow a revision to be exchanged
382 382 with another repository.
383 383
384 384 Instances represent the fulltext revision data or a delta against
385 385 another revision. Therefore the ``revision`` and ``delta`` attributes
386 386 are mutually exclusive.
387 387
388 388 Typically used for changegroup generation.
389 389 """
390 390
391 391 node = interfaceutil.Attribute(
392 392 """20 byte node of this revision.""")
393 393
394 394 p1node = interfaceutil.Attribute(
395 395 """20 byte node of 1st parent of this revision.""")
396 396
397 397 p2node = interfaceutil.Attribute(
398 398 """20 byte node of 2nd parent of this revision.""")
399 399
400 400 linknode = interfaceutil.Attribute(
401 401 """20 byte node of the changelog revision this node is linked to.""")
402 402
403 403 flags = interfaceutil.Attribute(
404 404 """2 bytes of integer flags that apply to this revision.
405 405
406 406 This is a bitwise composition of the ``REVISION_FLAG_*`` constants.
407 407 """)
408 408
409 409 basenode = interfaceutil.Attribute(
410 410 """20 byte node of the revision this data is a delta against.
411 411
412 412 ``nullid`` indicates that the revision is a full revision and not
413 413 a delta.
414 414 """)
415 415
416 416 baserevisionsize = interfaceutil.Attribute(
417 417 """Size of base revision this delta is against.
418 418
419 419 May be ``None`` if ``basenode`` is ``nullid``.
420 420 """)
421 421
422 422 revision = interfaceutil.Attribute(
423 423 """Raw fulltext of revision data for this node.""")
424 424
425 425 delta = interfaceutil.Attribute(
426 426 """Delta between ``basenode`` and ``node``.
427 427
428 428 Stored in the bdiff delta format.
429 429 """)
430 430
431 431 class ifilerevisionssequence(interfaceutil.Interface):
432 432 """Contains index data for all revisions of a file.
433 433
434 434 Types implementing this behave like lists of tuples. The index
435 435 in the list corresponds to the revision number. The values contain
436 436 index metadata.
437 437
438 438 The *null* revision (revision number -1) is always the last item
439 439 in the index.
440 440 """
441 441
442 442 def __len__():
443 443 """The total number of revisions."""
444 444
445 445 def __getitem__(rev):
446 446 """Returns the object having a specific revision number.
447 447
448 448 Returns an 8-tuple with the following fields:
449 449
450 450 offset+flags
451 451 Contains the offset and flags for the revision. 64-bit unsigned
452 452 integer where first 6 bytes are the offset and the next 2 bytes
453 453 are flags. The offset can be 0 if it is not used by the store.
454 454 compressed size
455 455 Size of the revision data in the store. It can be 0 if it isn't
456 456 needed by the store.
457 457 uncompressed size
458 458 Fulltext size. It can be 0 if it isn't needed by the store.
459 459 base revision
460 460 Revision number of revision the delta for storage is encoded
461 461 against. -1 indicates not encoded against a base revision.
462 462 link revision
463 463 Revision number of changelog revision this entry is related to.
464 464 p1 revision
465 465 Revision number of 1st parent. -1 if no 1st parent.
466 466 p2 revision
467 467 Revision number of 2nd parent. -1 if no 1st parent.
468 468 node
469 469 Binary node value for this revision number.
470 470
471 471 Negative values should index off the end of the sequence. ``-1``
472 472 should return the null revision. ``-2`` should return the most
473 473 recent revision.
474 474 """
475 475
476 476 def __contains__(rev):
477 477 """Whether a revision number exists."""
478 478
479 479 def insert(self, i, entry):
480 480 """Add an item to the index at specific revision."""
481 481
482 482 class ifileindex(interfaceutil.Interface):
483 483 """Storage interface for index data of a single file.
484 484
485 485 File storage data is divided into index metadata and data storage.
486 486 This interface defines the index portion of the interface.
487 487
488 488 The index logically consists of:
489 489
490 490 * A mapping between revision numbers and nodes.
491 491 * DAG data (storing and querying the relationship between nodes).
492 492 * Metadata to facilitate storage.
493 493 """
494 494 def __len__():
495 495 """Obtain the number of revisions stored for this file."""
496 496
497 497 def __iter__():
498 498 """Iterate over revision numbers for this file."""
499 499
500 500 def hasnode(node):
501 501 """Returns a bool indicating if a node is known to this store.
502 502
503 503 Implementations must only return True for full, binary node values:
504 504 hex nodes, revision numbers, and partial node matches must be
505 505 rejected.
506 506
507 507 The null node is never present.
508 508 """
509 509
510 510 def revs(start=0, stop=None):
511 511 """Iterate over revision numbers for this file, with control."""
512 512
513 513 def parents(node):
514 514 """Returns a 2-tuple of parent nodes for a revision.
515 515
516 516 Values will be ``nullid`` if the parent is empty.
517 517 """
518 518
519 519 def parentrevs(rev):
520 520 """Like parents() but operates on revision numbers."""
521 521
522 522 def rev(node):
523 523 """Obtain the revision number given a node.
524 524
525 525 Raises ``error.LookupError`` if the node is not known.
526 526 """
527 527
528 528 def node(rev):
529 529 """Obtain the node value given a revision number.
530 530
531 531 Raises ``IndexError`` if the node is not known.
532 532 """
533 533
534 534 def lookup(node):
535 535 """Attempt to resolve a value to a node.
536 536
537 537 Value can be a binary node, hex node, revision number, or a string
538 538 that can be converted to an integer.
539 539
540 540 Raises ``error.LookupError`` if a node could not be resolved.
541 541 """
542 542
543 543 def linkrev(rev):
544 544 """Obtain the changeset revision number a revision is linked to."""
545 545
546 546 def iscensored(rev):
547 547 """Return whether a revision's content has been censored."""
548 548
549 549 def commonancestorsheads(node1, node2):
550 550 """Obtain an iterable of nodes containing heads of common ancestors.
551 551
552 552 See ``ancestor.commonancestorsheads()``.
553 553 """
554 554
555 555 def descendants(revs):
556 556 """Obtain descendant revision numbers for a set of revision numbers.
557 557
558 558 If ``nullrev`` is in the set, this is equivalent to ``revs()``.
559 559 """
560 560
561 561 def heads(start=None, stop=None):
562 562 """Obtain a list of nodes that are DAG heads, with control.
563 563
564 564 The set of revisions examined can be limited by specifying
565 565 ``start`` and ``stop``. ``start`` is a node. ``stop`` is an
566 566 iterable of nodes. DAG traversal starts at earlier revision
567 567 ``start`` and iterates forward until any node in ``stop`` is
568 568 encountered.
569 569 """
570 570
571 571 def children(node):
572 572 """Obtain nodes that are children of a node.
573 573
574 574 Returns a list of nodes.
575 575 """
576 576
577 577 class ifiledata(interfaceutil.Interface):
578 578 """Storage interface for data storage of a specific file.
579 579
580 580 This complements ``ifileindex`` and provides an interface for accessing
581 581 data for a tracked file.
582 582 """
583 583 def size(rev):
584 584 """Obtain the fulltext size of file data.
585 585
586 586 Any metadata is excluded from size measurements.
587 587 """
588 588
589 589 def revision(node, raw=False):
590 590 """"Obtain fulltext data for a node.
591 591
592 592 By default, any storage transformations are applied before the data
593 593 is returned. If ``raw`` is True, non-raw storage transformations
594 594 are not applied.
595 595
596 596 The fulltext data may contain a header containing metadata. Most
597 597 consumers should use ``read()`` to obtain the actual file data.
598 598 """
599 599
600 600 def read(node):
601 601 """Resolve file fulltext data.
602 602
603 603 This is similar to ``revision()`` except any metadata in the data
604 604 headers is stripped.
605 605 """
606 606
607 607 def renamed(node):
608 608 """Obtain copy metadata for a node.
609 609
610 610 Returns ``False`` if no copy metadata is stored or a 2-tuple of
611 611 (path, node) from which this revision was copied.
612 612 """
613 613
614 614 def cmp(node, fulltext):
615 615 """Compare fulltext to another revision.
616 616
617 617 Returns True if the fulltext is different from what is stored.
618 618
619 619 This takes copy metadata into account.
620 620
621 621 TODO better document the copy metadata and censoring logic.
622 622 """
623 623
624 624 def emitrevisions(nodes,
625 625 nodesorder=None,
626 626 revisiondata=False,
627 627 assumehaveparentrevisions=False,
628 628 deltamode=CG_DELTAMODE_STD):
629 629 """Produce ``irevisiondelta`` for revisions.
630 630
631 631 Given an iterable of nodes, emits objects conforming to the
632 632 ``irevisiondelta`` interface that describe revisions in storage.
633 633
634 634 This method is a generator.
635 635
636 636 The input nodes may be unordered. Implementations must ensure that a
637 637 node's parents are emitted before the node itself. Transitively, this
638 638 means that a node may only be emitted once all its ancestors in
639 639 ``nodes`` have also been emitted.
640 640
641 641 By default, emits "index" data (the ``node``, ``p1node``, and
642 642 ``p2node`` attributes). If ``revisiondata`` is set, revision data
643 643 will also be present on the emitted objects.
644 644
645 645 With default argument values, implementations can choose to emit
646 646 either fulltext revision data or a delta. When emitting deltas,
647 647 implementations must consider whether the delta's base revision
648 648 fulltext is available to the receiver.
649 649
650 650 The base revision fulltext is guaranteed to be available if any of
651 651 the following are met:
652 652
653 653 * Its fulltext revision was emitted by this method call.
654 654 * A delta for that revision was emitted by this method call.
655 655 * ``assumehaveparentrevisions`` is True and the base revision is a
656 656 parent of the node.
657 657
658 658 ``nodesorder`` can be used to control the order that revisions are
659 659 emitted. By default, revisions can be reordered as long as they are
660 660 in DAG topological order (see above). If the value is ``nodes``,
661 661 the iteration order from ``nodes`` should be used. If the value is
662 662 ``storage``, then the native order from the backing storage layer
663 663 is used. (Not all storage layers will have strong ordering and behavior
664 664 of this mode is storage-dependent.) ``nodes`` ordering can force
665 665 revisions to be emitted before their ancestors, so consumers should
666 666 use it with care.
667 667
668 668 The ``linknode`` attribute on the returned ``irevisiondelta`` may not
669 669 be set and it is the caller's responsibility to resolve it, if needed.
670 670
671 671 If ``deltamode`` is CG_DELTAMODE_PREV and revision data is requested,
672 672 all revision data should be emitted as deltas against the revision
673 673 emitted just prior. The initial revision should be a delta against its
674 674 1st parent.
675 675 """
676 676
677 677 class ifilemutation(interfaceutil.Interface):
678 678 """Storage interface for mutation events of a tracked file."""
679 679
680 680 def add(filedata, meta, transaction, linkrev, p1, p2):
681 681 """Add a new revision to the store.
682 682
683 683 Takes file data, dictionary of metadata, a transaction, linkrev,
684 684 and parent nodes.
685 685
686 686 Returns the node that was added.
687 687
688 688 May no-op if a revision matching the supplied data is already stored.
689 689 """
690 690
691 691 def addrevision(revisiondata, transaction, linkrev, p1, p2, node=None,
692 692 flags=0, cachedelta=None):
693 693 """Add a new revision to the store.
694 694
695 695 This is similar to ``add()`` except it operates at a lower level.
696 696
697 697 The data passed in already contains a metadata header, if any.
698 698
699 699 ``node`` and ``flags`` can be used to define the expected node and
700 700 the flags to use with storage. ``flags`` is a bitwise value composed
701 701 of the various ``REVISION_FLAG_*`` constants.
702 702
703 703 ``add()`` is usually called when adding files from e.g. the working
704 704 directory. ``addrevision()`` is often called by ``add()`` and for
705 705 scenarios where revision data has already been computed, such as when
706 706 applying raw data from a peer repo.
707 707 """
708 708
709 709 def addgroup(deltas, linkmapper, transaction, addrevisioncb=None,
710 710 maybemissingparents=False):
711 711 """Process a series of deltas for storage.
712 712
713 713 ``deltas`` is an iterable of 7-tuples of
714 714 (node, p1, p2, linknode, deltabase, delta, flags) defining revisions
715 715 to add.
716 716
717 717 The ``delta`` field contains ``mpatch`` data to apply to a base
718 718 revision, identified by ``deltabase``. The base node can be
719 719 ``nullid``, in which case the header from the delta can be ignored
720 720 and the delta used as the fulltext.
721 721
722 722 ``addrevisioncb`` should be called for each node as it is committed.
723 723
724 724 ``maybemissingparents`` is a bool indicating whether the incoming
725 725 data may reference parents/ancestor revisions that aren't present.
726 726 This flag is set when receiving data into a "shallow" store that
727 727 doesn't hold all history.
728 728
729 729 Returns a list of nodes that were processed. A node will be in the list
730 730 even if it existed in the store previously.
731 731 """
732 732
733 733 def censorrevision(tr, node, tombstone=b''):
734 734 """Remove the content of a single revision.
735 735
736 736 The specified ``node`` will have its content purged from storage.
737 737 Future attempts to access the revision data for this node will
738 738 result in failure.
739 739
740 740 A ``tombstone`` message can optionally be stored. This message may be
741 741 displayed to users when they attempt to access the missing revision
742 742 data.
743 743
744 744 Storage backends may have stored deltas against the previous content
745 745 in this revision. As part of censoring a revision, these storage
746 746 backends are expected to rewrite any internally stored deltas such
747 747 that they no longer reference the deleted content.
748 748 """
749 749
750 750 def getstrippoint(minlink):
751 751 """Find the minimum revision that must be stripped to strip a linkrev.
752 752
753 753 Returns a 2-tuple containing the minimum revision number and a set
754 754 of all revisions numbers that would be broken by this strip.
755 755
756 756 TODO this is highly revlog centric and should be abstracted into
757 757 a higher-level deletion API. ``repair.strip()`` relies on this.
758 758 """
759 759
760 760 def strip(minlink, transaction):
761 761 """Remove storage of items starting at a linkrev.
762 762
763 763 This uses ``getstrippoint()`` to determine the first node to remove.
764 764 Then it effectively truncates storage for all revisions after that.
765 765
766 766 TODO this is highly revlog centric and should be abstracted into a
767 767 higher-level deletion API.
768 768 """
769 769
770 770 class ifilestorage(ifileindex, ifiledata, ifilemutation):
771 771 """Complete storage interface for a single tracked file."""
772 772
773 773 def files():
774 774 """Obtain paths that are backing storage for this file.
775 775
776 776 TODO this is used heavily by verify code and there should probably
777 777 be a better API for that.
778 778 """
779 779
780 780 def storageinfo(exclusivefiles=False, sharedfiles=False,
781 781 revisionscount=False, trackedsize=False,
782 782 storedsize=False):
783 783 """Obtain information about storage for this file's data.
784 784
785 785 Returns a dict describing storage for this tracked path. The keys
786 786 in the dict map to arguments of the same. The arguments are bools
787 787 indicating whether to calculate and obtain that data.
788 788
789 789 exclusivefiles
790 790 Iterable of (vfs, path) describing files that are exclusively
791 791 used to back storage for this tracked path.
792 792
793 793 sharedfiles
794 794 Iterable of (vfs, path) describing files that are used to back
795 795 storage for this tracked path. Those files may also provide storage
796 796 for other stored entities.
797 797
798 798 revisionscount
799 799 Number of revisions available for retrieval.
800 800
801 801 trackedsize
802 802 Total size in bytes of all tracked revisions. This is a sum of the
803 803 length of the fulltext of all revisions.
804 804
805 805 storedsize
806 806 Total size in bytes used to store data for all tracked revisions.
807 807 This is commonly less than ``trackedsize`` due to internal usage
808 808 of deltas rather than fulltext revisions.
809 809
810 810 Not all storage backends may support all queries are have a reasonable
811 811 value to use. In that case, the value should be set to ``None`` and
812 812 callers are expected to handle this special value.
813 813 """
814 814
815 815 def verifyintegrity(state):
816 816 """Verifies the integrity of file storage.
817 817
818 818 ``state`` is a dict holding state of the verifier process. It can be
819 819 used to communicate data between invocations of multiple storage
820 820 primitives.
821 821
822 822 If individual revisions cannot have their revision content resolved,
823 823 the method is expected to set the ``skipread`` key to a set of nodes
824 824 that encountered problems.
825 825
826 826 The method yields objects conforming to the ``iverifyproblem``
827 827 interface.
828 828 """
829 829
830 830 class idirs(interfaceutil.Interface):
831 831 """Interface representing a collection of directories from paths.
832 832
833 833 This interface is essentially a derived data structure representing
834 834 directories from a collection of paths.
835 835 """
836 836
837 837 def addpath(path):
838 838 """Add a path to the collection.
839 839
840 840 All directories in the path will be added to the collection.
841 841 """
842 842
843 843 def delpath(path):
844 844 """Remove a path from the collection.
845 845
846 846 If the removal was the last path in a particular directory, the
847 847 directory is removed from the collection.
848 848 """
849 849
850 850 def __iter__():
851 851 """Iterate over the directories in this collection of paths."""
852 852
853 853 def __contains__(path):
854 854 """Whether a specific directory is in this collection."""
855 855
856 856 class imanifestdict(interfaceutil.Interface):
857 857 """Interface representing a manifest data structure.
858 858
859 859 A manifest is effectively a dict mapping paths to entries. Each entry
860 860 consists of a binary node and extra flags affecting that entry.
861 861 """
862 862
863 863 def __getitem__(path):
864 864 """Returns the binary node value for a path in the manifest.
865 865
866 866 Raises ``KeyError`` if the path does not exist in the manifest.
867 867
868 868 Equivalent to ``self.find(path)[0]``.
869 869 """
870 870
871 871 def find(path):
872 872 """Returns the entry for a path in the manifest.
873 873
874 874 Returns a 2-tuple of (node, flags).
875 875
876 876 Raises ``KeyError`` if the path does not exist in the manifest.
877 877 """
878 878
879 879 def __len__():
880 880 """Return the number of entries in the manifest."""
881 881
882 882 def __nonzero__():
883 883 """Returns True if the manifest has entries, False otherwise."""
884 884
885 885 __bool__ = __nonzero__
886 886
887 887 def __setitem__(path, node):
888 888 """Define the node value for a path in the manifest.
889 889
890 890 If the path is already in the manifest, its flags will be copied to
891 891 the new entry.
892 892 """
893 893
894 894 def __contains__(path):
895 895 """Whether a path exists in the manifest."""
896 896
897 897 def __delitem__(path):
898 898 """Remove a path from the manifest.
899 899
900 900 Raises ``KeyError`` if the path is not in the manifest.
901 901 """
902 902
903 903 def __iter__():
904 904 """Iterate over paths in the manifest."""
905 905
906 906 def iterkeys():
907 907 """Iterate over paths in the manifest."""
908 908
909 909 def keys():
910 910 """Obtain a list of paths in the manifest."""
911 911
912 912 def filesnotin(other, match=None):
913 913 """Obtain the set of paths in this manifest but not in another.
914 914
915 915 ``match`` is an optional matcher function to be applied to both
916 916 manifests.
917 917
918 918 Returns a set of paths.
919 919 """
920 920
921 921 def dirs():
922 922 """Returns an object implementing the ``idirs`` interface."""
923 923
924 924 def hasdir(dir):
925 925 """Returns a bool indicating if a directory is in this manifest."""
926 926
927 927 def matches(match):
928 928 """Generate a new manifest filtered through a matcher.
929 929
930 930 Returns an object conforming to the ``imanifestdict`` interface.
931 931 """
932 932
933 933 def walk(match):
934 934 """Generator of paths in manifest satisfying a matcher.
935 935
936 936 This is equivalent to ``self.matches(match).iterkeys()`` except a new
937 937 manifest object is not created.
938 938
939 939 If the matcher has explicit files listed and they don't exist in
940 940 the manifest, ``match.bad()`` is called for each missing file.
941 941 """
942 942
943 943 def diff(other, match=None, clean=False):
944 944 """Find differences between this manifest and another.
945 945
946 946 This manifest is compared to ``other``.
947 947
948 948 If ``match`` is provided, the two manifests are filtered against this
949 949 matcher and only entries satisfying the matcher are compared.
950 950
951 951 If ``clean`` is True, unchanged files are included in the returned
952 952 object.
953 953
954 954 Returns a dict with paths as keys and values of 2-tuples of 2-tuples of
955 955 the form ``((node1, flag1), (node2, flag2))`` where ``(node1, flag1)``
956 956 represents the node and flags for this manifest and ``(node2, flag2)``
957 957 are the same for the other manifest.
958 958 """
959 959
960 960 def setflag(path, flag):
961 961 """Set the flag value for a given path.
962 962
963 963 Raises ``KeyError`` if the path is not already in the manifest.
964 964 """
965 965
966 966 def get(path, default=None):
967 967 """Obtain the node value for a path or a default value if missing."""
968 968
969 969 def flags(path, default=''):
970 970 """Return the flags value for a path or a default value if missing."""
971 971
972 972 def copy():
973 973 """Return a copy of this manifest."""
974 974
975 975 def items():
976 976 """Returns an iterable of (path, node) for items in this manifest."""
977 977
978 978 def iteritems():
979 979 """Identical to items()."""
980 980
981 981 def iterentries():
982 982 """Returns an iterable of (path, node, flags) for this manifest.
983 983
984 984 Similar to ``iteritems()`` except items are a 3-tuple and include
985 985 flags.
986 986 """
987 987
988 988 def text():
989 989 """Obtain the raw data representation for this manifest.
990 990
991 991 Result is used to create a manifest revision.
992 992 """
993 993
994 994 def fastdelta(base, changes):
995 995 """Obtain a delta between this manifest and another given changes.
996 996
997 997 ``base`` in the raw data representation for another manifest.
998 998
999 999 ``changes`` is an iterable of ``(path, to_delete)``.
1000 1000
1001 1001 Returns a 2-tuple containing ``bytearray(self.text())`` and the
1002 1002 delta between ``base`` and this manifest.
1003 1003 """
1004 1004
1005 1005 class imanifestrevisionbase(interfaceutil.Interface):
1006 1006 """Base interface representing a single revision of a manifest.
1007 1007
1008 1008 Should not be used as a primary interface: should always be inherited
1009 1009 as part of a larger interface.
1010 1010 """
1011 1011
1012 1012 def new():
1013 1013 """Obtain a new manifest instance.
1014 1014
1015 1015 Returns an object conforming to the ``imanifestrevisionwritable``
1016 1016 interface. The instance will be associated with the same
1017 1017 ``imanifestlog`` collection as this instance.
1018 1018 """
1019 1019
1020 1020 def copy():
1021 1021 """Obtain a copy of this manifest instance.
1022 1022
1023 1023 Returns an object conforming to the ``imanifestrevisionwritable``
1024 1024 interface. The instance will be associated with the same
1025 1025 ``imanifestlog`` collection as this instance.
1026 1026 """
1027 1027
1028 1028 def read():
1029 1029 """Obtain the parsed manifest data structure.
1030 1030
1031 1031 The returned object conforms to the ``imanifestdict`` interface.
1032 1032 """
1033 1033
1034 1034 class imanifestrevisionstored(imanifestrevisionbase):
1035 1035 """Interface representing a manifest revision committed to storage."""
1036 1036
1037 1037 def node():
1038 1038 """The binary node for this manifest."""
1039 1039
1040 1040 parents = interfaceutil.Attribute(
1041 1041 """List of binary nodes that are parents for this manifest revision."""
1042 1042 )
1043 1043
1044 1044 def readdelta(shallow=False):
1045 1045 """Obtain the manifest data structure representing changes from parent.
1046 1046
1047 1047 This manifest is compared to its 1st parent. A new manifest representing
1048 1048 those differences is constructed.
1049 1049
1050 1050 The returned object conforms to the ``imanifestdict`` interface.
1051 1051 """
1052 1052
1053 1053 def readfast(shallow=False):
1054 1054 """Calls either ``read()`` or ``readdelta()``.
1055 1055
1056 1056 The faster of the two options is called.
1057 1057 """
1058 1058
1059 1059 def find(key):
1060 1060 """Calls self.read().find(key)``.
1061 1061
1062 1062 Returns a 2-tuple of ``(node, flags)`` or raises ``KeyError``.
1063 1063 """
1064 1064
1065 1065 class imanifestrevisionwritable(imanifestrevisionbase):
1066 1066 """Interface representing a manifest revision that can be committed."""
1067 1067
1068 1068 def write(transaction, linkrev, p1node, p2node, added, removed, match=None):
1069 1069 """Add this revision to storage.
1070 1070
1071 1071 Takes a transaction object, the changeset revision number it will
1072 1072 be associated with, its parent nodes, and lists of added and
1073 1073 removed paths.
1074 1074
1075 1075 If match is provided, storage can choose not to inspect or write out
1076 1076 items that do not match. Storage is still required to be able to provide
1077 1077 the full manifest in the future for any directories written (these
1078 1078 manifests should not be "narrowed on disk").
1079 1079
1080 1080 Returns the binary node of the created revision.
1081 1081 """
1082 1082
1083 1083 class imanifeststorage(interfaceutil.Interface):
1084 1084 """Storage interface for manifest data."""
1085 1085
1086 1086 tree = interfaceutil.Attribute(
1087 1087 """The path to the directory this manifest tracks.
1088 1088
1089 1089 The empty bytestring represents the root manifest.
1090 1090 """)
1091 1091
1092 1092 index = interfaceutil.Attribute(
1093 1093 """An ``ifilerevisionssequence`` instance.""")
1094 1094
1095 1095 indexfile = interfaceutil.Attribute(
1096 1096 """Path of revlog index file.
1097 1097
1098 1098 TODO this is revlog specific and should not be exposed.
1099 1099 """)
1100 1100
1101 1101 opener = interfaceutil.Attribute(
1102 1102 """VFS opener to use to access underlying files used for storage.
1103 1103
1104 1104 TODO this is revlog specific and should not be exposed.
1105 1105 """)
1106 1106
1107 1107 version = interfaceutil.Attribute(
1108 1108 """Revlog version number.
1109 1109
1110 1110 TODO this is revlog specific and should not be exposed.
1111 1111 """)
1112 1112
1113 1113 _generaldelta = interfaceutil.Attribute(
1114 1114 """Whether generaldelta storage is being used.
1115 1115
1116 1116 TODO this is revlog specific and should not be exposed.
1117 1117 """)
1118 1118
1119 1119 fulltextcache = interfaceutil.Attribute(
1120 1120 """Dict with cache of fulltexts.
1121 1121
1122 1122 TODO this doesn't feel appropriate for the storage interface.
1123 1123 """)
1124 1124
1125 1125 def __len__():
1126 1126 """Obtain the number of revisions stored for this manifest."""
1127 1127
1128 1128 def __iter__():
1129 1129 """Iterate over revision numbers for this manifest."""
1130 1130
1131 1131 def rev(node):
1132 1132 """Obtain the revision number given a binary node.
1133 1133
1134 1134 Raises ``error.LookupError`` if the node is not known.
1135 1135 """
1136 1136
1137 1137 def node(rev):
1138 1138 """Obtain the node value given a revision number.
1139 1139
1140 1140 Raises ``error.LookupError`` if the revision is not known.
1141 1141 """
1142 1142
1143 1143 def lookup(value):
1144 1144 """Attempt to resolve a value to a node.
1145 1145
1146 1146 Value can be a binary node, hex node, revision number, or a bytes
1147 1147 that can be converted to an integer.
1148 1148
1149 1149 Raises ``error.LookupError`` if a ndoe could not be resolved.
1150 1150 """
1151 1151
1152 1152 def parents(node):
1153 1153 """Returns a 2-tuple of parent nodes for a node.
1154 1154
1155 1155 Values will be ``nullid`` if the parent is empty.
1156 1156 """
1157 1157
1158 1158 def parentrevs(rev):
1159 1159 """Like parents() but operates on revision numbers."""
1160 1160
1161 1161 def linkrev(rev):
1162 1162 """Obtain the changeset revision number a revision is linked to."""
1163 1163
1164 1164 def revision(node, _df=None, raw=False):
1165 1165 """Obtain fulltext data for a node."""
1166 1166
1167 def rawdata(node, _df=None):
1168 """Obtain raw data for a node."""
1169
1167 1170 def revdiff(rev1, rev2):
1168 1171 """Obtain a delta between two revision numbers.
1169 1172
1170 1173 The returned data is the result of ``bdiff.bdiff()`` on the raw
1171 1174 revision data.
1172 1175 """
1173 1176
1174 1177 def cmp(node, fulltext):
1175 1178 """Compare fulltext to another revision.
1176 1179
1177 1180 Returns True if the fulltext is different from what is stored.
1178 1181 """
1179 1182
1180 1183 def emitrevisions(nodes,
1181 1184 nodesorder=None,
1182 1185 revisiondata=False,
1183 1186 assumehaveparentrevisions=False):
1184 1187 """Produce ``irevisiondelta`` describing revisions.
1185 1188
1186 1189 See the documentation for ``ifiledata`` for more.
1187 1190 """
1188 1191
1189 1192 def addgroup(deltas, linkmapper, transaction, addrevisioncb=None):
1190 1193 """Process a series of deltas for storage.
1191 1194
1192 1195 See the documentation in ``ifilemutation`` for more.
1193 1196 """
1194 1197
1195 1198 def rawsize(rev):
1196 1199 """Obtain the size of tracked data.
1197 1200
1198 1201 Is equivalent to ``len(m.revision(node, raw=True))``.
1199 1202
1200 1203 TODO this method is only used by upgrade code and may be removed.
1201 1204 """
1202 1205
1203 1206 def getstrippoint(minlink):
1204 1207 """Find minimum revision that must be stripped to strip a linkrev.
1205 1208
1206 1209 See the documentation in ``ifilemutation`` for more.
1207 1210 """
1208 1211
1209 1212 def strip(minlink, transaction):
1210 1213 """Remove storage of items starting at a linkrev.
1211 1214
1212 1215 See the documentation in ``ifilemutation`` for more.
1213 1216 """
1214 1217
1215 1218 def checksize():
1216 1219 """Obtain the expected sizes of backing files.
1217 1220
1218 1221 TODO this is used by verify and it should not be part of the interface.
1219 1222 """
1220 1223
1221 1224 def files():
1222 1225 """Obtain paths that are backing storage for this manifest.
1223 1226
1224 1227 TODO this is used by verify and there should probably be a better API
1225 1228 for this functionality.
1226 1229 """
1227 1230
1228 1231 def deltaparent(rev):
1229 1232 """Obtain the revision that a revision is delta'd against.
1230 1233
1231 1234 TODO delta encoding is an implementation detail of storage and should
1232 1235 not be exposed to the storage interface.
1233 1236 """
1234 1237
1235 1238 def clone(tr, dest, **kwargs):
1236 1239 """Clone this instance to another."""
1237 1240
1238 1241 def clearcaches(clear_persisted_data=False):
1239 1242 """Clear any caches associated with this instance."""
1240 1243
1241 1244 def dirlog(d):
1242 1245 """Obtain a manifest storage instance for a tree."""
1243 1246
1244 1247 def add(m, transaction, link, p1, p2, added, removed, readtree=None,
1245 1248 match=None):
1246 1249 """Add a revision to storage.
1247 1250
1248 1251 ``m`` is an object conforming to ``imanifestdict``.
1249 1252
1250 1253 ``link`` is the linkrev revision number.
1251 1254
1252 1255 ``p1`` and ``p2`` are the parent revision numbers.
1253 1256
1254 1257 ``added`` and ``removed`` are iterables of added and removed paths,
1255 1258 respectively.
1256 1259
1257 1260 ``readtree`` is a function that can be used to read the child tree(s)
1258 1261 when recursively writing the full tree structure when using
1259 1262 treemanifets.
1260 1263
1261 1264 ``match`` is a matcher that can be used to hint to storage that not all
1262 1265 paths must be inspected; this is an optimization and can be safely
1263 1266 ignored. Note that the storage must still be able to reproduce a full
1264 1267 manifest including files that did not match.
1265 1268 """
1266 1269
1267 1270 def storageinfo(exclusivefiles=False, sharedfiles=False,
1268 1271 revisionscount=False, trackedsize=False,
1269 1272 storedsize=False):
1270 1273 """Obtain information about storage for this manifest's data.
1271 1274
1272 1275 See ``ifilestorage.storageinfo()`` for a description of this method.
1273 1276 This one behaves the same way, except for manifest data.
1274 1277 """
1275 1278
1276 1279 class imanifestlog(interfaceutil.Interface):
1277 1280 """Interface representing a collection of manifest snapshots.
1278 1281
1279 1282 Represents the root manifest in a repository.
1280 1283
1281 1284 Also serves as a means to access nested tree manifests and to cache
1282 1285 tree manifests.
1283 1286 """
1284 1287
1285 1288 def __getitem__(node):
1286 1289 """Obtain a manifest instance for a given binary node.
1287 1290
1288 1291 Equivalent to calling ``self.get('', node)``.
1289 1292
1290 1293 The returned object conforms to the ``imanifestrevisionstored``
1291 1294 interface.
1292 1295 """
1293 1296
1294 1297 def get(tree, node, verify=True):
1295 1298 """Retrieve the manifest instance for a given directory and binary node.
1296 1299
1297 1300 ``node`` always refers to the node of the root manifest (which will be
1298 1301 the only manifest if flat manifests are being used).
1299 1302
1300 1303 If ``tree`` is the empty string, the root manifest is returned.
1301 1304 Otherwise the manifest for the specified directory will be returned
1302 1305 (requires tree manifests).
1303 1306
1304 1307 If ``verify`` is True, ``LookupError`` is raised if the node is not
1305 1308 known.
1306 1309
1307 1310 The returned object conforms to the ``imanifestrevisionstored``
1308 1311 interface.
1309 1312 """
1310 1313
1311 1314 def getstorage(tree):
1312 1315 """Retrieve an interface to storage for a particular tree.
1313 1316
1314 1317 If ``tree`` is the empty bytestring, storage for the root manifest will
1315 1318 be returned. Otherwise storage for a tree manifest is returned.
1316 1319
1317 1320 TODO formalize interface for returned object.
1318 1321 """
1319 1322
1320 1323 def clearcaches():
1321 1324 """Clear caches associated with this collection."""
1322 1325
1323 1326 def rev(node):
1324 1327 """Obtain the revision number for a binary node.
1325 1328
1326 1329 Raises ``error.LookupError`` if the node is not known.
1327 1330 """
1328 1331
1329 1332 class ilocalrepositoryfilestorage(interfaceutil.Interface):
1330 1333 """Local repository sub-interface providing access to tracked file storage.
1331 1334
1332 1335 This interface defines how a repository accesses storage for a single
1333 1336 tracked file path.
1334 1337 """
1335 1338
1336 1339 def file(f):
1337 1340 """Obtain a filelog for a tracked path.
1338 1341
1339 1342 The returned type conforms to the ``ifilestorage`` interface.
1340 1343 """
1341 1344
1342 1345 class ilocalrepositorymain(interfaceutil.Interface):
1343 1346 """Main interface for local repositories.
1344 1347
1345 1348 This currently captures the reality of things - not how things should be.
1346 1349 """
1347 1350
1348 1351 supportedformats = interfaceutil.Attribute(
1349 1352 """Set of requirements that apply to stream clone.
1350 1353
1351 1354 This is actually a class attribute and is shared among all instances.
1352 1355 """)
1353 1356
1354 1357 supported = interfaceutil.Attribute(
1355 1358 """Set of requirements that this repo is capable of opening.""")
1356 1359
1357 1360 requirements = interfaceutil.Attribute(
1358 1361 """Set of requirements this repo uses.""")
1359 1362
1360 1363 features = interfaceutil.Attribute(
1361 1364 """Set of "features" this repository supports.
1362 1365
1363 1366 A "feature" is a loosely-defined term. It can refer to a feature
1364 1367 in the classical sense or can describe an implementation detail
1365 1368 of the repository. For example, a ``readonly`` feature may denote
1366 1369 the repository as read-only. Or a ``revlogfilestore`` feature may
1367 1370 denote that the repository is using revlogs for file storage.
1368 1371
1369 1372 The intent of features is to provide a machine-queryable mechanism
1370 1373 for repo consumers to test for various repository characteristics.
1371 1374
1372 1375 Features are similar to ``requirements``. The main difference is that
1373 1376 requirements are stored on-disk and represent requirements to open the
1374 1377 repository. Features are more run-time capabilities of the repository
1375 1378 and more granular capabilities (which may be derived from requirements).
1376 1379 """)
1377 1380
1378 1381 filtername = interfaceutil.Attribute(
1379 1382 """Name of the repoview that is active on this repo.""")
1380 1383
1381 1384 wvfs = interfaceutil.Attribute(
1382 1385 """VFS used to access the working directory.""")
1383 1386
1384 1387 vfs = interfaceutil.Attribute(
1385 1388 """VFS rooted at the .hg directory.
1386 1389
1387 1390 Used to access repository data not in the store.
1388 1391 """)
1389 1392
1390 1393 svfs = interfaceutil.Attribute(
1391 1394 """VFS rooted at the store.
1392 1395
1393 1396 Used to access repository data in the store. Typically .hg/store.
1394 1397 But can point elsewhere if the store is shared.
1395 1398 """)
1396 1399
1397 1400 root = interfaceutil.Attribute(
1398 1401 """Path to the root of the working directory.""")
1399 1402
1400 1403 path = interfaceutil.Attribute(
1401 1404 """Path to the .hg directory.""")
1402 1405
1403 1406 origroot = interfaceutil.Attribute(
1404 1407 """The filesystem path that was used to construct the repo.""")
1405 1408
1406 1409 auditor = interfaceutil.Attribute(
1407 1410 """A pathauditor for the working directory.
1408 1411
1409 1412 This checks if a path refers to a nested repository.
1410 1413
1411 1414 Operates on the filesystem.
1412 1415 """)
1413 1416
1414 1417 nofsauditor = interfaceutil.Attribute(
1415 1418 """A pathauditor for the working directory.
1416 1419
1417 1420 This is like ``auditor`` except it doesn't do filesystem checks.
1418 1421 """)
1419 1422
1420 1423 baseui = interfaceutil.Attribute(
1421 1424 """Original ui instance passed into constructor.""")
1422 1425
1423 1426 ui = interfaceutil.Attribute(
1424 1427 """Main ui instance for this instance.""")
1425 1428
1426 1429 sharedpath = interfaceutil.Attribute(
1427 1430 """Path to the .hg directory of the repo this repo was shared from.""")
1428 1431
1429 1432 store = interfaceutil.Attribute(
1430 1433 """A store instance.""")
1431 1434
1432 1435 spath = interfaceutil.Attribute(
1433 1436 """Path to the store.""")
1434 1437
1435 1438 sjoin = interfaceutil.Attribute(
1436 1439 """Alias to self.store.join.""")
1437 1440
1438 1441 cachevfs = interfaceutil.Attribute(
1439 1442 """A VFS used to access the cache directory.
1440 1443
1441 1444 Typically .hg/cache.
1442 1445 """)
1443 1446
1444 1447 wcachevfs = interfaceutil.Attribute(
1445 1448 """A VFS used to access the cache directory dedicated to working copy
1446 1449
1447 1450 Typically .hg/wcache.
1448 1451 """)
1449 1452
1450 1453 filteredrevcache = interfaceutil.Attribute(
1451 1454 """Holds sets of revisions to be filtered.""")
1452 1455
1453 1456 names = interfaceutil.Attribute(
1454 1457 """A ``namespaces`` instance.""")
1455 1458
1456 1459 def close():
1457 1460 """Close the handle on this repository."""
1458 1461
1459 1462 def peer():
1460 1463 """Obtain an object conforming to the ``peer`` interface."""
1461 1464
1462 1465 def unfiltered():
1463 1466 """Obtain an unfiltered/raw view of this repo."""
1464 1467
1465 1468 def filtered(name, visibilityexceptions=None):
1466 1469 """Obtain a named view of this repository."""
1467 1470
1468 1471 obsstore = interfaceutil.Attribute(
1469 1472 """A store of obsolescence data.""")
1470 1473
1471 1474 changelog = interfaceutil.Attribute(
1472 1475 """A handle on the changelog revlog.""")
1473 1476
1474 1477 manifestlog = interfaceutil.Attribute(
1475 1478 """An instance conforming to the ``imanifestlog`` interface.
1476 1479
1477 1480 Provides access to manifests for the repository.
1478 1481 """)
1479 1482
1480 1483 dirstate = interfaceutil.Attribute(
1481 1484 """Working directory state.""")
1482 1485
1483 1486 narrowpats = interfaceutil.Attribute(
1484 1487 """Matcher patterns for this repository's narrowspec.""")
1485 1488
1486 1489 def narrowmatch(match=None, includeexact=False):
1487 1490 """Obtain a matcher for the narrowspec."""
1488 1491
1489 1492 def setnarrowpats(newincludes, newexcludes):
1490 1493 """Define the narrowspec for this repository."""
1491 1494
1492 1495 def __getitem__(changeid):
1493 1496 """Try to resolve a changectx."""
1494 1497
1495 1498 def __contains__(changeid):
1496 1499 """Whether a changeset exists."""
1497 1500
1498 1501 def __nonzero__():
1499 1502 """Always returns True."""
1500 1503 return True
1501 1504
1502 1505 __bool__ = __nonzero__
1503 1506
1504 1507 def __len__():
1505 1508 """Returns the number of changesets in the repo."""
1506 1509
1507 1510 def __iter__():
1508 1511 """Iterate over revisions in the changelog."""
1509 1512
1510 1513 def revs(expr, *args):
1511 1514 """Evaluate a revset.
1512 1515
1513 1516 Emits revisions.
1514 1517 """
1515 1518
1516 1519 def set(expr, *args):
1517 1520 """Evaluate a revset.
1518 1521
1519 1522 Emits changectx instances.
1520 1523 """
1521 1524
1522 1525 def anyrevs(specs, user=False, localalias=None):
1523 1526 """Find revisions matching one of the given revsets."""
1524 1527
1525 1528 def url():
1526 1529 """Returns a string representing the location of this repo."""
1527 1530
1528 1531 def hook(name, throw=False, **args):
1529 1532 """Call a hook."""
1530 1533
1531 1534 def tags():
1532 1535 """Return a mapping of tag to node."""
1533 1536
1534 1537 def tagtype(tagname):
1535 1538 """Return the type of a given tag."""
1536 1539
1537 1540 def tagslist():
1538 1541 """Return a list of tags ordered by revision."""
1539 1542
1540 1543 def nodetags(node):
1541 1544 """Return the tags associated with a node."""
1542 1545
1543 1546 def nodebookmarks(node):
1544 1547 """Return the list of bookmarks pointing to the specified node."""
1545 1548
1546 1549 def branchmap():
1547 1550 """Return a mapping of branch to heads in that branch."""
1548 1551
1549 1552 def revbranchcache():
1550 1553 pass
1551 1554
1552 1555 def branchtip(branchtip, ignoremissing=False):
1553 1556 """Return the tip node for a given branch."""
1554 1557
1555 1558 def lookup(key):
1556 1559 """Resolve the node for a revision."""
1557 1560
1558 1561 def lookupbranch(key):
1559 1562 """Look up the branch name of the given revision or branch name."""
1560 1563
1561 1564 def known(nodes):
1562 1565 """Determine whether a series of nodes is known.
1563 1566
1564 1567 Returns a list of bools.
1565 1568 """
1566 1569
1567 1570 def local():
1568 1571 """Whether the repository is local."""
1569 1572 return True
1570 1573
1571 1574 def publishing():
1572 1575 """Whether the repository is a publishing repository."""
1573 1576
1574 1577 def cancopy():
1575 1578 pass
1576 1579
1577 1580 def shared():
1578 1581 """The type of shared repository or None."""
1579 1582
1580 1583 def wjoin(f, *insidef):
1581 1584 """Calls self.vfs.reljoin(self.root, f, *insidef)"""
1582 1585
1583 1586 def setparents(p1, p2):
1584 1587 """Set the parent nodes of the working directory."""
1585 1588
1586 1589 def filectx(path, changeid=None, fileid=None):
1587 1590 """Obtain a filectx for the given file revision."""
1588 1591
1589 1592 def getcwd():
1590 1593 """Obtain the current working directory from the dirstate."""
1591 1594
1592 1595 def pathto(f, cwd=None):
1593 1596 """Obtain the relative path to a file."""
1594 1597
1595 1598 def adddatafilter(name, fltr):
1596 1599 pass
1597 1600
1598 1601 def wread(filename):
1599 1602 """Read a file from wvfs, using data filters."""
1600 1603
1601 1604 def wwrite(filename, data, flags, backgroundclose=False, **kwargs):
1602 1605 """Write data to a file in the wvfs, using data filters."""
1603 1606
1604 1607 def wwritedata(filename, data):
1605 1608 """Resolve data for writing to the wvfs, using data filters."""
1606 1609
1607 1610 def currenttransaction():
1608 1611 """Obtain the current transaction instance or None."""
1609 1612
1610 1613 def transaction(desc, report=None):
1611 1614 """Open a new transaction to write to the repository."""
1612 1615
1613 1616 def undofiles():
1614 1617 """Returns a list of (vfs, path) for files to undo transactions."""
1615 1618
1616 1619 def recover():
1617 1620 """Roll back an interrupted transaction."""
1618 1621
1619 1622 def rollback(dryrun=False, force=False):
1620 1623 """Undo the last transaction.
1621 1624
1622 1625 DANGEROUS.
1623 1626 """
1624 1627
1625 1628 def updatecaches(tr=None, full=False):
1626 1629 """Warm repo caches."""
1627 1630
1628 1631 def invalidatecaches():
1629 1632 """Invalidate cached data due to the repository mutating."""
1630 1633
1631 1634 def invalidatevolatilesets():
1632 1635 pass
1633 1636
1634 1637 def invalidatedirstate():
1635 1638 """Invalidate the dirstate."""
1636 1639
1637 1640 def invalidate(clearfilecache=False):
1638 1641 pass
1639 1642
1640 1643 def invalidateall():
1641 1644 pass
1642 1645
1643 1646 def lock(wait=True):
1644 1647 """Lock the repository store and return a lock instance."""
1645 1648
1646 1649 def wlock(wait=True):
1647 1650 """Lock the non-store parts of the repository."""
1648 1651
1649 1652 def currentwlock():
1650 1653 """Return the wlock if it's held or None."""
1651 1654
1652 1655 def checkcommitpatterns(wctx, vdirs, match, status, fail):
1653 1656 pass
1654 1657
1655 1658 def commit(text='', user=None, date=None, match=None, force=False,
1656 1659 editor=False, extra=None):
1657 1660 """Add a new revision to the repository."""
1658 1661
1659 1662 def commitctx(ctx, error=False, origctx=None):
1660 1663 """Commit a commitctx instance to the repository."""
1661 1664
1662 1665 def destroying():
1663 1666 """Inform the repository that nodes are about to be destroyed."""
1664 1667
1665 1668 def destroyed():
1666 1669 """Inform the repository that nodes have been destroyed."""
1667 1670
1668 1671 def status(node1='.', node2=None, match=None, ignored=False,
1669 1672 clean=False, unknown=False, listsubrepos=False):
1670 1673 """Convenience method to call repo[x].status()."""
1671 1674
1672 1675 def addpostdsstatus(ps):
1673 1676 pass
1674 1677
1675 1678 def postdsstatus():
1676 1679 pass
1677 1680
1678 1681 def clearpostdsstatus():
1679 1682 pass
1680 1683
1681 1684 def heads(start=None):
1682 1685 """Obtain list of nodes that are DAG heads."""
1683 1686
1684 1687 def branchheads(branch=None, start=None, closed=False):
1685 1688 pass
1686 1689
1687 1690 def branches(nodes):
1688 1691 pass
1689 1692
1690 1693 def between(pairs):
1691 1694 pass
1692 1695
1693 1696 def checkpush(pushop):
1694 1697 pass
1695 1698
1696 1699 prepushoutgoinghooks = interfaceutil.Attribute(
1697 1700 """util.hooks instance.""")
1698 1701
1699 1702 def pushkey(namespace, key, old, new):
1700 1703 pass
1701 1704
1702 1705 def listkeys(namespace):
1703 1706 pass
1704 1707
1705 1708 def debugwireargs(one, two, three=None, four=None, five=None):
1706 1709 pass
1707 1710
1708 1711 def savecommitmessage(text):
1709 1712 pass
1710 1713
1711 1714 class completelocalrepository(ilocalrepositorymain,
1712 1715 ilocalrepositoryfilestorage):
1713 1716 """Complete interface for a local repository."""
1714 1717
1715 1718 class iwireprotocolcommandcacher(interfaceutil.Interface):
1716 1719 """Represents a caching backend for wire protocol commands.
1717 1720
1718 1721 Wire protocol version 2 supports transparent caching of many commands.
1719 1722 To leverage this caching, servers can activate objects that cache
1720 1723 command responses. Objects handle both cache writing and reading.
1721 1724 This interface defines how that response caching mechanism works.
1722 1725
1723 1726 Wire protocol version 2 commands emit a series of objects that are
1724 1727 serialized and sent to the client. The caching layer exists between
1725 1728 the invocation of the command function and the sending of its output
1726 1729 objects to an output layer.
1727 1730
1728 1731 Instances of this interface represent a binding to a cache that
1729 1732 can serve a response (in place of calling a command function) and/or
1730 1733 write responses to a cache for subsequent use.
1731 1734
1732 1735 When a command request arrives, the following happens with regards
1733 1736 to this interface:
1734 1737
1735 1738 1. The server determines whether the command request is cacheable.
1736 1739 2. If it is, an instance of this interface is spawned.
1737 1740 3. The cacher is activated in a context manager (``__enter__`` is called).
1738 1741 4. A cache *key* for that request is derived. This will call the
1739 1742 instance's ``adjustcachekeystate()`` method so the derivation
1740 1743 can be influenced.
1741 1744 5. The cacher is informed of the derived cache key via a call to
1742 1745 ``setcachekey()``.
1743 1746 6. The cacher's ``lookup()`` method is called to test for presence of
1744 1747 the derived key in the cache.
1745 1748 7. If ``lookup()`` returns a hit, that cached result is used in place
1746 1749 of invoking the command function. ``__exit__`` is called and the instance
1747 1750 is discarded.
1748 1751 8. The command function is invoked.
1749 1752 9. ``onobject()`` is called for each object emitted by the command
1750 1753 function.
1751 1754 10. After the final object is seen, ``onfinished()`` is called.
1752 1755 11. ``__exit__`` is called to signal the end of use of the instance.
1753 1756
1754 1757 Cache *key* derivation can be influenced by the instance.
1755 1758
1756 1759 Cache keys are initially derived by a deterministic representation of
1757 1760 the command request. This includes the command name, arguments, protocol
1758 1761 version, etc. This initial key derivation is performed by CBOR-encoding a
1759 1762 data structure and feeding that output into a hasher.
1760 1763
1761 1764 Instances of this interface can influence this initial key derivation
1762 1765 via ``adjustcachekeystate()``.
1763 1766
1764 1767 The instance is informed of the derived cache key via a call to
1765 1768 ``setcachekey()``. The instance must store the key locally so it can
1766 1769 be consulted on subsequent operations that may require it.
1767 1770
1768 1771 When constructed, the instance has access to a callable that can be used
1769 1772 for encoding response objects. This callable receives as its single
1770 1773 argument an object emitted by a command function. It returns an iterable
1771 1774 of bytes chunks representing the encoded object. Unless the cacher is
1772 1775 caching native Python objects in memory or has a way of reconstructing
1773 1776 the original Python objects, implementations typically call this function
1774 1777 to produce bytes from the output objects and then store those bytes in
1775 1778 the cache. When it comes time to re-emit those bytes, they are wrapped
1776 1779 in a ``wireprototypes.encodedresponse`` instance to tell the output
1777 1780 layer that they are pre-encoded.
1778 1781
1779 1782 When receiving the objects emitted by the command function, instances
1780 1783 can choose what to do with those objects. The simplest thing to do is
1781 1784 re-emit the original objects. They will be forwarded to the output
1782 1785 layer and will be processed as if the cacher did not exist.
1783 1786
1784 1787 Implementations could also choose to not emit objects - instead locally
1785 1788 buffering objects or their encoded representation. They could then emit
1786 1789 a single "coalesced" object when ``onfinished()`` is called. In
1787 1790 this way, the implementation would function as a filtering layer of
1788 1791 sorts.
1789 1792
1790 1793 When caching objects, typically the encoded form of the object will
1791 1794 be stored. Keep in mind that if the original object is forwarded to
1792 1795 the output layer, it will need to be encoded there as well. For large
1793 1796 output, this redundant encoding could add overhead. Implementations
1794 1797 could wrap the encoded object data in ``wireprototypes.encodedresponse``
1795 1798 instances to avoid this overhead.
1796 1799 """
1797 1800 def __enter__():
1798 1801 """Marks the instance as active.
1799 1802
1800 1803 Should return self.
1801 1804 """
1802 1805
1803 1806 def __exit__(exctype, excvalue, exctb):
1804 1807 """Called when cacher is no longer used.
1805 1808
1806 1809 This can be used by implementations to perform cleanup actions (e.g.
1807 1810 disconnecting network sockets, aborting a partially cached response.
1808 1811 """
1809 1812
1810 1813 def adjustcachekeystate(state):
1811 1814 """Influences cache key derivation by adjusting state to derive key.
1812 1815
1813 1816 A dict defining the state used to derive the cache key is passed.
1814 1817
1815 1818 Implementations can modify this dict to record additional state that
1816 1819 is wanted to influence key derivation.
1817 1820
1818 1821 Implementations are *highly* encouraged to not modify or delete
1819 1822 existing keys.
1820 1823 """
1821 1824
1822 1825 def setcachekey(key):
1823 1826 """Record the derived cache key for this request.
1824 1827
1825 1828 Instances may mutate the key for internal usage, as desired. e.g.
1826 1829 instances may wish to prepend the repo name, introduce path
1827 1830 components for filesystem or URL addressing, etc. Behavior is up to
1828 1831 the cache.
1829 1832
1830 1833 Returns a bool indicating if the request is cacheable by this
1831 1834 instance.
1832 1835 """
1833 1836
1834 1837 def lookup():
1835 1838 """Attempt to resolve an entry in the cache.
1836 1839
1837 1840 The instance is instructed to look for the cache key that it was
1838 1841 informed about via the call to ``setcachekey()``.
1839 1842
1840 1843 If there's no cache hit or the cacher doesn't wish to use the cached
1841 1844 entry, ``None`` should be returned.
1842 1845
1843 1846 Else, a dict defining the cached result should be returned. The
1844 1847 dict may have the following keys:
1845 1848
1846 1849 objs
1847 1850 An iterable of objects that should be sent to the client. That
1848 1851 iterable of objects is expected to be what the command function
1849 1852 would return if invoked or an equivalent representation thereof.
1850 1853 """
1851 1854
1852 1855 def onobject(obj):
1853 1856 """Called when a new object is emitted from the command function.
1854 1857
1855 1858 Receives as its argument the object that was emitted from the
1856 1859 command function.
1857 1860
1858 1861 This method returns an iterator of objects to forward to the output
1859 1862 layer. The easiest implementation is a generator that just
1860 1863 ``yield obj``.
1861 1864 """
1862 1865
1863 1866 def onfinished():
1864 1867 """Called after all objects have been emitted from the command function.
1865 1868
1866 1869 Implementations should return an iterator of objects to forward to
1867 1870 the output layer.
1868 1871
1869 1872 This method can be a generator.
1870 1873 """
General Comments 0
You need to be logged in to leave comments. Login now