##// END OF EJS Templates
revlog: remove legacy usage of `_lazydeltabase`...
marmoute -
r51960:26dcdbe1 default
parent child Browse files
Show More
@@ -1,4623 +1,4626 b''
1 1 # perf.py - performance test routines
2 2 '''helper extension to measure performance
3 3
4 4 Configurations
5 5 ==============
6 6
7 7 ``perf``
8 8 --------
9 9
10 10 ``all-timing``
11 11 When set, additional statistics will be reported for each benchmark: best,
12 12 worst, median average. If not set only the best timing is reported
13 13 (default: off).
14 14
15 15 ``presleep``
16 16 number of second to wait before any group of runs (default: 1)
17 17
18 18 ``pre-run``
19 19 number of run to perform before starting measurement.
20 20
21 21 ``profile-benchmark``
22 22 Enable profiling for the benchmarked section.
23 23 (The first iteration is benchmarked)
24 24
25 25 ``run-limits``
26 26 Control the number of runs each benchmark will perform. The option value
27 27 should be a list of `<time>-<numberofrun>` pairs. After each run the
28 28 conditions are considered in order with the following logic:
29 29
30 30 If benchmark has been running for <time> seconds, and we have performed
31 31 <numberofrun> iterations, stop the benchmark,
32 32
33 33 The default value is: `3.0-100, 10.0-3`
34 34
35 35 ``stub``
36 36 When set, benchmarks will only be run once, useful for testing
37 37 (default: off)
38 38 '''
39 39
40 40 # "historical portability" policy of perf.py:
41 41 #
42 42 # We have to do:
43 43 # - make perf.py "loadable" with as wide Mercurial version as possible
44 44 # This doesn't mean that perf commands work correctly with that Mercurial.
45 45 # BTW, perf.py itself has been available since 1.1 (or eb240755386d).
46 46 # - make historical perf command work correctly with as wide Mercurial
47 47 # version as possible
48 48 #
49 49 # We have to do, if possible with reasonable cost:
50 50 # - make recent perf command for historical feature work correctly
51 51 # with early Mercurial
52 52 #
53 53 # We don't have to do:
54 54 # - make perf command for recent feature work correctly with early
55 55 # Mercurial
56 56
57 57 import contextlib
58 58 import functools
59 59 import gc
60 60 import os
61 61 import random
62 62 import shutil
63 63 import struct
64 64 import sys
65 65 import tempfile
66 66 import threading
67 67 import time
68 68
69 69 import mercurial.revlog
70 70 from mercurial import (
71 71 changegroup,
72 72 cmdutil,
73 73 commands,
74 74 copies,
75 75 error,
76 76 extensions,
77 77 hg,
78 78 mdiff,
79 79 merge,
80 80 util,
81 81 )
82 82
83 83 # for "historical portability":
84 84 # try to import modules separately (in dict order), and ignore
85 85 # failure, because these aren't available with early Mercurial
86 86 try:
87 87 from mercurial import branchmap # since 2.5 (or bcee63733aad)
88 88 except ImportError:
89 89 pass
90 90 try:
91 91 from mercurial import obsolete # since 2.3 (or ad0d6c2b3279)
92 92 except ImportError:
93 93 pass
94 94 try:
95 95 from mercurial import registrar # since 3.7 (or 37d50250b696)
96 96
97 97 dir(registrar) # forcibly load it
98 98 except ImportError:
99 99 registrar = None
100 100 try:
101 101 from mercurial import repoview # since 2.5 (or 3a6ddacb7198)
102 102 except ImportError:
103 103 pass
104 104 try:
105 105 from mercurial.utils import repoviewutil # since 5.0
106 106 except ImportError:
107 107 repoviewutil = None
108 108 try:
109 109 from mercurial import scmutil # since 1.9 (or 8b252e826c68)
110 110 except ImportError:
111 111 pass
112 112 try:
113 113 from mercurial import setdiscovery # since 1.9 (or cb98fed52495)
114 114 except ImportError:
115 115 pass
116 116
117 117 try:
118 118 from mercurial import profiling
119 119 except ImportError:
120 120 profiling = None
121 121
122 122 try:
123 123 from mercurial.revlogutils import constants as revlog_constants
124 124
125 125 perf_rl_kind = (revlog_constants.KIND_OTHER, b'created-by-perf')
126 126
127 127 def revlog(opener, *args, **kwargs):
128 128 return mercurial.revlog.revlog(opener, perf_rl_kind, *args, **kwargs)
129 129
130 130
131 131 except (ImportError, AttributeError):
132 132 perf_rl_kind = None
133 133
134 134 def revlog(opener, *args, **kwargs):
135 135 return mercurial.revlog.revlog(opener, *args, **kwargs)
136 136
137 137
138 138 def identity(a):
139 139 return a
140 140
141 141
142 142 try:
143 143 from mercurial import pycompat
144 144
145 145 getargspec = pycompat.getargspec # added to module after 4.5
146 146 _byteskwargs = pycompat.byteskwargs # since 4.1 (or fbc3f73dc802)
147 147 _sysstr = pycompat.sysstr # since 4.0 (or 2219f4f82ede)
148 148 _bytestr = pycompat.bytestr # since 4.2 (or b70407bd84d5)
149 149 _xrange = pycompat.xrange # since 4.8 (or 7eba8f83129b)
150 150 fsencode = pycompat.fsencode # since 3.9 (or f4a5e0e86a7e)
151 151 if pycompat.ispy3:
152 152 _maxint = sys.maxsize # per py3 docs for replacing maxint
153 153 else:
154 154 _maxint = sys.maxint
155 155 except (NameError, ImportError, AttributeError):
156 156 import inspect
157 157
158 158 getargspec = inspect.getargspec
159 159 _byteskwargs = identity
160 160 _bytestr = str
161 161 fsencode = identity # no py3 support
162 162 _maxint = sys.maxint # no py3 support
163 163 _sysstr = lambda x: x # no py3 support
164 164 _xrange = xrange
165 165
166 166 try:
167 167 # 4.7+
168 168 queue = pycompat.queue.Queue
169 169 except (NameError, AttributeError, ImportError):
170 170 # <4.7.
171 171 try:
172 172 queue = pycompat.queue
173 173 except (NameError, AttributeError, ImportError):
174 174 import Queue as queue
175 175
176 176 try:
177 177 from mercurial import logcmdutil
178 178
179 179 makelogtemplater = logcmdutil.maketemplater
180 180 except (AttributeError, ImportError):
181 181 try:
182 182 makelogtemplater = cmdutil.makelogtemplater
183 183 except (AttributeError, ImportError):
184 184 makelogtemplater = None
185 185
186 186 # for "historical portability":
187 187 # define util.safehasattr forcibly, because util.safehasattr has been
188 188 # available since 1.9.3 (or 94b200a11cf7)
189 189 _undefined = object()
190 190
191 191
192 192 def safehasattr(thing, attr):
193 193 return getattr(thing, _sysstr(attr), _undefined) is not _undefined
194 194
195 195
196 196 setattr(util, 'safehasattr', safehasattr)
197 197
198 198 # for "historical portability":
199 199 # define util.timer forcibly, because util.timer has been available
200 200 # since ae5d60bb70c9
201 201 if safehasattr(time, 'perf_counter'):
202 202 util.timer = time.perf_counter
203 203 elif os.name == b'nt':
204 204 util.timer = time.clock
205 205 else:
206 206 util.timer = time.time
207 207
208 208 # for "historical portability":
209 209 # use locally defined empty option list, if formatteropts isn't
210 210 # available, because commands.formatteropts has been available since
211 211 # 3.2 (or 7a7eed5176a4), even though formatting itself has been
212 212 # available since 2.2 (or ae5f92e154d3)
213 213 formatteropts = getattr(
214 214 cmdutil, "formatteropts", getattr(commands, "formatteropts", [])
215 215 )
216 216
217 217 # for "historical portability":
218 218 # use locally defined option list, if debugrevlogopts isn't available,
219 219 # because commands.debugrevlogopts has been available since 3.7 (or
220 220 # 5606f7d0d063), even though cmdutil.openrevlog() has been available
221 221 # since 1.9 (or a79fea6b3e77).
222 222 revlogopts = getattr(
223 223 cmdutil,
224 224 "debugrevlogopts",
225 225 getattr(
226 226 commands,
227 227 "debugrevlogopts",
228 228 [
229 229 (b'c', b'changelog', False, b'open changelog'),
230 230 (b'm', b'manifest', False, b'open manifest'),
231 231 (b'', b'dir', False, b'open directory manifest'),
232 232 ],
233 233 ),
234 234 )
235 235
236 236 cmdtable = {}
237 237
238 238
239 239 # for "historical portability":
240 240 # define parsealiases locally, because cmdutil.parsealiases has been
241 241 # available since 1.5 (or 6252852b4332)
242 242 def parsealiases(cmd):
243 243 return cmd.split(b"|")
244 244
245 245
246 246 if safehasattr(registrar, 'command'):
247 247 command = registrar.command(cmdtable)
248 248 elif safehasattr(cmdutil, 'command'):
249 249 command = cmdutil.command(cmdtable)
250 250 if 'norepo' not in getargspec(command).args:
251 251 # for "historical portability":
252 252 # wrap original cmdutil.command, because "norepo" option has
253 253 # been available since 3.1 (or 75a96326cecb)
254 254 _command = command
255 255
256 256 def command(name, options=(), synopsis=None, norepo=False):
257 257 if norepo:
258 258 commands.norepo += b' %s' % b' '.join(parsealiases(name))
259 259 return _command(name, list(options), synopsis)
260 260
261 261
262 262 else:
263 263 # for "historical portability":
264 264 # define "@command" annotation locally, because cmdutil.command
265 265 # has been available since 1.9 (or 2daa5179e73f)
266 266 def command(name, options=(), synopsis=None, norepo=False):
267 267 def decorator(func):
268 268 if synopsis:
269 269 cmdtable[name] = func, list(options), synopsis
270 270 else:
271 271 cmdtable[name] = func, list(options)
272 272 if norepo:
273 273 commands.norepo += b' %s' % b' '.join(parsealiases(name))
274 274 return func
275 275
276 276 return decorator
277 277
278 278
279 279 try:
280 280 import mercurial.registrar
281 281 import mercurial.configitems
282 282
283 283 configtable = {}
284 284 configitem = mercurial.registrar.configitem(configtable)
285 285 configitem(
286 286 b'perf',
287 287 b'presleep',
288 288 default=mercurial.configitems.dynamicdefault,
289 289 experimental=True,
290 290 )
291 291 configitem(
292 292 b'perf',
293 293 b'stub',
294 294 default=mercurial.configitems.dynamicdefault,
295 295 experimental=True,
296 296 )
297 297 configitem(
298 298 b'perf',
299 299 b'parentscount',
300 300 default=mercurial.configitems.dynamicdefault,
301 301 experimental=True,
302 302 )
303 303 configitem(
304 304 b'perf',
305 305 b'all-timing',
306 306 default=mercurial.configitems.dynamicdefault,
307 307 experimental=True,
308 308 )
309 309 configitem(
310 310 b'perf',
311 311 b'pre-run',
312 312 default=mercurial.configitems.dynamicdefault,
313 313 )
314 314 configitem(
315 315 b'perf',
316 316 b'profile-benchmark',
317 317 default=mercurial.configitems.dynamicdefault,
318 318 )
319 319 configitem(
320 320 b'perf',
321 321 b'run-limits',
322 322 default=mercurial.configitems.dynamicdefault,
323 323 experimental=True,
324 324 )
325 325 except (ImportError, AttributeError):
326 326 pass
327 327 except TypeError:
328 328 # compatibility fix for a11fd395e83f
329 329 # hg version: 5.2
330 330 configitem(
331 331 b'perf',
332 332 b'presleep',
333 333 default=mercurial.configitems.dynamicdefault,
334 334 )
335 335 configitem(
336 336 b'perf',
337 337 b'stub',
338 338 default=mercurial.configitems.dynamicdefault,
339 339 )
340 340 configitem(
341 341 b'perf',
342 342 b'parentscount',
343 343 default=mercurial.configitems.dynamicdefault,
344 344 )
345 345 configitem(
346 346 b'perf',
347 347 b'all-timing',
348 348 default=mercurial.configitems.dynamicdefault,
349 349 )
350 350 configitem(
351 351 b'perf',
352 352 b'pre-run',
353 353 default=mercurial.configitems.dynamicdefault,
354 354 )
355 355 configitem(
356 356 b'perf',
357 357 b'profile-benchmark',
358 358 default=mercurial.configitems.dynamicdefault,
359 359 )
360 360 configitem(
361 361 b'perf',
362 362 b'run-limits',
363 363 default=mercurial.configitems.dynamicdefault,
364 364 )
365 365
366 366
367 367 def getlen(ui):
368 368 if ui.configbool(b"perf", b"stub", False):
369 369 return lambda x: 1
370 370 return len
371 371
372 372
373 373 class noop:
374 374 """dummy context manager"""
375 375
376 376 def __enter__(self):
377 377 pass
378 378
379 379 def __exit__(self, *args):
380 380 pass
381 381
382 382
383 383 NOOPCTX = noop()
384 384
385 385
386 386 def gettimer(ui, opts=None):
387 387 """return a timer function and formatter: (timer, formatter)
388 388
389 389 This function exists to gather the creation of formatter in a single
390 390 place instead of duplicating it in all performance commands."""
391 391
392 392 # enforce an idle period before execution to counteract power management
393 393 # experimental config: perf.presleep
394 394 time.sleep(getint(ui, b"perf", b"presleep", 1))
395 395
396 396 if opts is None:
397 397 opts = {}
398 398 # redirect all to stderr unless buffer api is in use
399 399 if not ui._buffers:
400 400 ui = ui.copy()
401 401 uifout = safeattrsetter(ui, b'fout', ignoremissing=True)
402 402 if uifout:
403 403 # for "historical portability":
404 404 # ui.fout/ferr have been available since 1.9 (or 4e1ccd4c2b6d)
405 405 uifout.set(ui.ferr)
406 406
407 407 # get a formatter
408 408 uiformatter = getattr(ui, 'formatter', None)
409 409 if uiformatter:
410 410 fm = uiformatter(b'perf', opts)
411 411 else:
412 412 # for "historical portability":
413 413 # define formatter locally, because ui.formatter has been
414 414 # available since 2.2 (or ae5f92e154d3)
415 415 from mercurial import node
416 416
417 417 class defaultformatter:
418 418 """Minimized composition of baseformatter and plainformatter"""
419 419
420 420 def __init__(self, ui, topic, opts):
421 421 self._ui = ui
422 422 if ui.debugflag:
423 423 self.hexfunc = node.hex
424 424 else:
425 425 self.hexfunc = node.short
426 426
427 427 def __nonzero__(self):
428 428 return False
429 429
430 430 __bool__ = __nonzero__
431 431
432 432 def startitem(self):
433 433 pass
434 434
435 435 def data(self, **data):
436 436 pass
437 437
438 438 def write(self, fields, deftext, *fielddata, **opts):
439 439 self._ui.write(deftext % fielddata, **opts)
440 440
441 441 def condwrite(self, cond, fields, deftext, *fielddata, **opts):
442 442 if cond:
443 443 self._ui.write(deftext % fielddata, **opts)
444 444
445 445 def plain(self, text, **opts):
446 446 self._ui.write(text, **opts)
447 447
448 448 def end(self):
449 449 pass
450 450
451 451 fm = defaultformatter(ui, b'perf', opts)
452 452
453 453 # stub function, runs code only once instead of in a loop
454 454 # experimental config: perf.stub
455 455 if ui.configbool(b"perf", b"stub", False):
456 456 return functools.partial(stub_timer, fm), fm
457 457
458 458 # experimental config: perf.all-timing
459 459 displayall = ui.configbool(b"perf", b"all-timing", True)
460 460
461 461 # experimental config: perf.run-limits
462 462 limitspec = ui.configlist(b"perf", b"run-limits", [])
463 463 limits = []
464 464 for item in limitspec:
465 465 parts = item.split(b'-', 1)
466 466 if len(parts) < 2:
467 467 ui.warn((b'malformatted run limit entry, missing "-": %s\n' % item))
468 468 continue
469 469 try:
470 470 time_limit = float(_sysstr(parts[0]))
471 471 except ValueError as e:
472 472 ui.warn(
473 473 (
474 474 b'malformatted run limit entry, %s: %s\n'
475 475 % (_bytestr(e), item)
476 476 )
477 477 )
478 478 continue
479 479 try:
480 480 run_limit = int(_sysstr(parts[1]))
481 481 except ValueError as e:
482 482 ui.warn(
483 483 (
484 484 b'malformatted run limit entry, %s: %s\n'
485 485 % (_bytestr(e), item)
486 486 )
487 487 )
488 488 continue
489 489 limits.append((time_limit, run_limit))
490 490 if not limits:
491 491 limits = DEFAULTLIMITS
492 492
493 493 profiler = None
494 494 if profiling is not None:
495 495 if ui.configbool(b"perf", b"profile-benchmark", False):
496 496 profiler = profiling.profile(ui)
497 497
498 498 prerun = getint(ui, b"perf", b"pre-run", 0)
499 499 t = functools.partial(
500 500 _timer,
501 501 fm,
502 502 displayall=displayall,
503 503 limits=limits,
504 504 prerun=prerun,
505 505 profiler=profiler,
506 506 )
507 507 return t, fm
508 508
509 509
510 510 def stub_timer(fm, func, setup=None, title=None):
511 511 if setup is not None:
512 512 setup()
513 513 func()
514 514
515 515
516 516 @contextlib.contextmanager
517 517 def timeone():
518 518 r = []
519 519 ostart = os.times()
520 520 cstart = util.timer()
521 521 yield r
522 522 cstop = util.timer()
523 523 ostop = os.times()
524 524 a, b = ostart, ostop
525 525 r.append((cstop - cstart, b[0] - a[0], b[1] - a[1]))
526 526
527 527
528 528 # list of stop condition (elapsed time, minimal run count)
529 529 DEFAULTLIMITS = (
530 530 (3.0, 100),
531 531 (10.0, 3),
532 532 )
533 533
534 534
535 535 @contextlib.contextmanager
536 536 def noop_context():
537 537 yield
538 538
539 539
540 540 def _timer(
541 541 fm,
542 542 func,
543 543 setup=None,
544 544 context=noop_context,
545 545 title=None,
546 546 displayall=False,
547 547 limits=DEFAULTLIMITS,
548 548 prerun=0,
549 549 profiler=None,
550 550 ):
551 551 gc.collect()
552 552 results = []
553 553 begin = util.timer()
554 554 count = 0
555 555 if profiler is None:
556 556 profiler = NOOPCTX
557 557 for i in range(prerun):
558 558 if setup is not None:
559 559 setup()
560 560 with context():
561 561 func()
562 562 keepgoing = True
563 563 while keepgoing:
564 564 if setup is not None:
565 565 setup()
566 566 with context():
567 567 with profiler:
568 568 with timeone() as item:
569 569 r = func()
570 570 profiler = NOOPCTX
571 571 count += 1
572 572 results.append(item[0])
573 573 cstop = util.timer()
574 574 # Look for a stop condition.
575 575 elapsed = cstop - begin
576 576 for t, mincount in limits:
577 577 if elapsed >= t and count >= mincount:
578 578 keepgoing = False
579 579 break
580 580
581 581 formatone(fm, results, title=title, result=r, displayall=displayall)
582 582
583 583
584 584 def formatone(fm, timings, title=None, result=None, displayall=False):
585 585 count = len(timings)
586 586
587 587 fm.startitem()
588 588
589 589 if title:
590 590 fm.write(b'title', b'! %s\n', title)
591 591 if result:
592 592 fm.write(b'result', b'! result: %s\n', result)
593 593
594 594 def display(role, entry):
595 595 prefix = b''
596 596 if role != b'best':
597 597 prefix = b'%s.' % role
598 598 fm.plain(b'!')
599 599 fm.write(prefix + b'wall', b' wall %f', entry[0])
600 600 fm.write(prefix + b'comb', b' comb %f', entry[1] + entry[2])
601 601 fm.write(prefix + b'user', b' user %f', entry[1])
602 602 fm.write(prefix + b'sys', b' sys %f', entry[2])
603 603 fm.write(prefix + b'count', b' (%s of %%d)' % role, count)
604 604 fm.plain(b'\n')
605 605
606 606 timings.sort()
607 607 min_val = timings[0]
608 608 display(b'best', min_val)
609 609 if displayall:
610 610 max_val = timings[-1]
611 611 display(b'max', max_val)
612 612 avg = tuple([sum(x) / count for x in zip(*timings)])
613 613 display(b'avg', avg)
614 614 median = timings[len(timings) // 2]
615 615 display(b'median', median)
616 616
617 617
618 618 # utilities for historical portability
619 619
620 620
621 621 def getint(ui, section, name, default):
622 622 # for "historical portability":
623 623 # ui.configint has been available since 1.9 (or fa2b596db182)
624 624 v = ui.config(section, name, None)
625 625 if v is None:
626 626 return default
627 627 try:
628 628 return int(v)
629 629 except ValueError:
630 630 raise error.ConfigError(
631 631 b"%s.%s is not an integer ('%s')" % (section, name, v)
632 632 )
633 633
634 634
635 635 def safeattrsetter(obj, name, ignoremissing=False):
636 636 """Ensure that 'obj' has 'name' attribute before subsequent setattr
637 637
638 638 This function is aborted, if 'obj' doesn't have 'name' attribute
639 639 at runtime. This avoids overlooking removal of an attribute, which
640 640 breaks assumption of performance measurement, in the future.
641 641
642 642 This function returns the object to (1) assign a new value, and
643 643 (2) restore an original value to the attribute.
644 644
645 645 If 'ignoremissing' is true, missing 'name' attribute doesn't cause
646 646 abortion, and this function returns None. This is useful to
647 647 examine an attribute, which isn't ensured in all Mercurial
648 648 versions.
649 649 """
650 650 if not util.safehasattr(obj, name):
651 651 if ignoremissing:
652 652 return None
653 653 raise error.Abort(
654 654 (
655 655 b"missing attribute %s of %s might break assumption"
656 656 b" of performance measurement"
657 657 )
658 658 % (name, obj)
659 659 )
660 660
661 661 origvalue = getattr(obj, _sysstr(name))
662 662
663 663 class attrutil:
664 664 def set(self, newvalue):
665 665 setattr(obj, _sysstr(name), newvalue)
666 666
667 667 def restore(self):
668 668 setattr(obj, _sysstr(name), origvalue)
669 669
670 670 return attrutil()
671 671
672 672
673 673 # utilities to examine each internal API changes
674 674
675 675
676 676 def getbranchmapsubsettable():
677 677 # for "historical portability":
678 678 # subsettable is defined in:
679 679 # - branchmap since 2.9 (or 175c6fd8cacc)
680 680 # - repoview since 2.5 (or 59a9f18d4587)
681 681 # - repoviewutil since 5.0
682 682 for mod in (branchmap, repoview, repoviewutil):
683 683 subsettable = getattr(mod, 'subsettable', None)
684 684 if subsettable:
685 685 return subsettable
686 686
687 687 # bisecting in bcee63733aad::59a9f18d4587 can reach here (both
688 688 # branchmap and repoview modules exist, but subsettable attribute
689 689 # doesn't)
690 690 raise error.Abort(
691 691 b"perfbranchmap not available with this Mercurial",
692 692 hint=b"use 2.5 or later",
693 693 )
694 694
695 695
696 696 def getsvfs(repo):
697 697 """Return appropriate object to access files under .hg/store"""
698 698 # for "historical portability":
699 699 # repo.svfs has been available since 2.3 (or 7034365089bf)
700 700 svfs = getattr(repo, 'svfs', None)
701 701 if svfs:
702 702 return svfs
703 703 else:
704 704 return getattr(repo, 'sopener')
705 705
706 706
707 707 def getvfs(repo):
708 708 """Return appropriate object to access files under .hg"""
709 709 # for "historical portability":
710 710 # repo.vfs has been available since 2.3 (or 7034365089bf)
711 711 vfs = getattr(repo, 'vfs', None)
712 712 if vfs:
713 713 return vfs
714 714 else:
715 715 return getattr(repo, 'opener')
716 716
717 717
718 718 def repocleartagscachefunc(repo):
719 719 """Return the function to clear tags cache according to repo internal API"""
720 720 if util.safehasattr(repo, b'_tagscache'): # since 2.0 (or 9dca7653b525)
721 721 # in this case, setattr(repo, '_tagscache', None) or so isn't
722 722 # correct way to clear tags cache, because existing code paths
723 723 # expect _tagscache to be a structured object.
724 724 def clearcache():
725 725 # _tagscache has been filteredpropertycache since 2.5 (or
726 726 # 98c867ac1330), and delattr() can't work in such case
727 727 if '_tagscache' in vars(repo):
728 728 del repo.__dict__['_tagscache']
729 729
730 730 return clearcache
731 731
732 732 repotags = safeattrsetter(repo, b'_tags', ignoremissing=True)
733 733 if repotags: # since 1.4 (or 5614a628d173)
734 734 return lambda: repotags.set(None)
735 735
736 736 repotagscache = safeattrsetter(repo, b'tagscache', ignoremissing=True)
737 737 if repotagscache: # since 0.6 (or d7df759d0e97)
738 738 return lambda: repotagscache.set(None)
739 739
740 740 # Mercurial earlier than 0.6 (or d7df759d0e97) logically reaches
741 741 # this point, but it isn't so problematic, because:
742 742 # - repo.tags of such Mercurial isn't "callable", and repo.tags()
743 743 # in perftags() causes failure soon
744 744 # - perf.py itself has been available since 1.1 (or eb240755386d)
745 745 raise error.Abort(b"tags API of this hg command is unknown")
746 746
747 747
748 748 # utilities to clear cache
749 749
750 750
751 751 def clearfilecache(obj, attrname):
752 752 unfiltered = getattr(obj, 'unfiltered', None)
753 753 if unfiltered is not None:
754 754 obj = obj.unfiltered()
755 755 if attrname in vars(obj):
756 756 delattr(obj, attrname)
757 757 obj._filecache.pop(attrname, None)
758 758
759 759
760 760 def clearchangelog(repo):
761 761 if repo is not repo.unfiltered():
762 762 object.__setattr__(repo, '_clcachekey', None)
763 763 object.__setattr__(repo, '_clcache', None)
764 764 clearfilecache(repo.unfiltered(), 'changelog')
765 765
766 766
767 767 # perf commands
768 768
769 769
770 770 @command(b'perf::walk|perfwalk', formatteropts)
771 771 def perfwalk(ui, repo, *pats, **opts):
772 772 opts = _byteskwargs(opts)
773 773 timer, fm = gettimer(ui, opts)
774 774 m = scmutil.match(repo[None], pats, {})
775 775 timer(
776 776 lambda: len(
777 777 list(
778 778 repo.dirstate.walk(m, subrepos=[], unknown=True, ignored=False)
779 779 )
780 780 )
781 781 )
782 782 fm.end()
783 783
784 784
785 785 @command(b'perf::annotate|perfannotate', formatteropts)
786 786 def perfannotate(ui, repo, f, **opts):
787 787 opts = _byteskwargs(opts)
788 788 timer, fm = gettimer(ui, opts)
789 789 fc = repo[b'.'][f]
790 790 timer(lambda: len(fc.annotate(True)))
791 791 fm.end()
792 792
793 793
794 794 @command(
795 795 b'perf::status|perfstatus',
796 796 [
797 797 (b'u', b'unknown', False, b'ask status to look for unknown files'),
798 798 (b'', b'dirstate', False, b'benchmark the internal dirstate call'),
799 799 ]
800 800 + formatteropts,
801 801 )
802 802 def perfstatus(ui, repo, **opts):
803 803 """benchmark the performance of a single status call
804 804
805 805 The repository data are preserved between each call.
806 806
807 807 By default, only the status of the tracked file are requested. If
808 808 `--unknown` is passed, the "unknown" files are also tracked.
809 809 """
810 810 opts = _byteskwargs(opts)
811 811 # m = match.always(repo.root, repo.getcwd())
812 812 # timer(lambda: sum(map(len, repo.dirstate.status(m, [], False, False,
813 813 # False))))
814 814 timer, fm = gettimer(ui, opts)
815 815 if opts[b'dirstate']:
816 816 dirstate = repo.dirstate
817 817 m = scmutil.matchall(repo)
818 818 unknown = opts[b'unknown']
819 819
820 820 def status_dirstate():
821 821 s = dirstate.status(
822 822 m, subrepos=[], ignored=False, clean=False, unknown=unknown
823 823 )
824 824 sum(map(bool, s))
825 825
826 826 if util.safehasattr(dirstate, 'running_status'):
827 827 with dirstate.running_status(repo):
828 828 timer(status_dirstate)
829 829 dirstate.invalidate()
830 830 else:
831 831 timer(status_dirstate)
832 832 else:
833 833 timer(lambda: sum(map(len, repo.status(unknown=opts[b'unknown']))))
834 834 fm.end()
835 835
836 836
837 837 @command(b'perf::addremove|perfaddremove', formatteropts)
838 838 def perfaddremove(ui, repo, **opts):
839 839 opts = _byteskwargs(opts)
840 840 timer, fm = gettimer(ui, opts)
841 841 try:
842 842 oldquiet = repo.ui.quiet
843 843 repo.ui.quiet = True
844 844 matcher = scmutil.match(repo[None])
845 845 opts[b'dry_run'] = True
846 846 if 'uipathfn' in getargspec(scmutil.addremove).args:
847 847 uipathfn = scmutil.getuipathfn(repo)
848 848 timer(lambda: scmutil.addremove(repo, matcher, b"", uipathfn, opts))
849 849 else:
850 850 timer(lambda: scmutil.addremove(repo, matcher, b"", opts))
851 851 finally:
852 852 repo.ui.quiet = oldquiet
853 853 fm.end()
854 854
855 855
856 856 def clearcaches(cl):
857 857 # behave somewhat consistently across internal API changes
858 858 if util.safehasattr(cl, b'clearcaches'):
859 859 cl.clearcaches()
860 860 elif util.safehasattr(cl, b'_nodecache'):
861 861 # <= hg-5.2
862 862 from mercurial.node import nullid, nullrev
863 863
864 864 cl._nodecache = {nullid: nullrev}
865 865 cl._nodepos = None
866 866
867 867
868 868 @command(b'perf::heads|perfheads', formatteropts)
869 869 def perfheads(ui, repo, **opts):
870 870 """benchmark the computation of a changelog heads"""
871 871 opts = _byteskwargs(opts)
872 872 timer, fm = gettimer(ui, opts)
873 873 cl = repo.changelog
874 874
875 875 def s():
876 876 clearcaches(cl)
877 877
878 878 def d():
879 879 len(cl.headrevs())
880 880
881 881 timer(d, setup=s)
882 882 fm.end()
883 883
884 884
885 885 def _default_clear_on_disk_tags_cache(repo):
886 886 from mercurial import tags
887 887
888 888 repo.cachevfs.tryunlink(tags._filename(repo))
889 889
890 890
891 891 def _default_clear_on_disk_tags_fnodes_cache(repo):
892 892 from mercurial import tags
893 893
894 894 repo.cachevfs.tryunlink(tags._fnodescachefile)
895 895
896 896
897 897 def _default_forget_fnodes(repo, revs):
898 898 """function used by the perf extension to prune some entries from the
899 899 fnodes cache"""
900 900 from mercurial import tags
901 901
902 902 missing_1 = b'\xff' * 4
903 903 missing_2 = b'\xff' * 20
904 904 cache = tags.hgtagsfnodescache(repo.unfiltered())
905 905 for r in revs:
906 906 cache._writeentry(r * tags._fnodesrecsize, missing_1, missing_2)
907 907 cache.write()
908 908
909 909
910 910 @command(
911 911 b'perf::tags|perftags',
912 912 formatteropts
913 913 + [
914 914 (b'', b'clear-revlogs', False, b'refresh changelog and manifest'),
915 915 (
916 916 b'',
917 917 b'clear-on-disk-cache',
918 918 False,
919 919 b'clear on disk tags cache (DESTRUCTIVE)',
920 920 ),
921 921 (
922 922 b'',
923 923 b'clear-fnode-cache-all',
924 924 False,
925 925 b'clear on disk file node cache (DESTRUCTIVE),',
926 926 ),
927 927 (
928 928 b'',
929 929 b'clear-fnode-cache-rev',
930 930 [],
931 931 b'clear on disk file node cache (DESTRUCTIVE),',
932 932 b'REVS',
933 933 ),
934 934 (
935 935 b'',
936 936 b'update-last',
937 937 b'',
938 938 b'simulate an update over the last N revisions (DESTRUCTIVE),',
939 939 b'N',
940 940 ),
941 941 ],
942 942 )
943 943 def perftags(ui, repo, **opts):
944 944 """Benchmark tags retrieval in various situation
945 945
946 946 The option marked as (DESTRUCTIVE) will alter the on-disk cache, possibly
947 947 altering performance after the command was run. However, it does not
948 948 destroy any stored data.
949 949 """
950 950 from mercurial import tags
951 951
952 952 opts = _byteskwargs(opts)
953 953 timer, fm = gettimer(ui, opts)
954 954 repocleartagscache = repocleartagscachefunc(repo)
955 955 clearrevlogs = opts[b'clear_revlogs']
956 956 clear_disk = opts[b'clear_on_disk_cache']
957 957 clear_fnode = opts[b'clear_fnode_cache_all']
958 958
959 959 clear_fnode_revs = opts[b'clear_fnode_cache_rev']
960 960 update_last_str = opts[b'update_last']
961 961 update_last = None
962 962 if update_last_str:
963 963 try:
964 964 update_last = int(update_last_str)
965 965 except ValueError:
966 966 msg = b'could not parse value for update-last: "%s"'
967 967 msg %= update_last_str
968 968 hint = b'value should be an integer'
969 969 raise error.Abort(msg, hint=hint)
970 970
971 971 clear_disk_fn = getattr(
972 972 tags,
973 973 "clear_cache_on_disk",
974 974 _default_clear_on_disk_tags_cache,
975 975 )
976 976 clear_fnodes_fn = getattr(
977 977 tags,
978 978 "clear_cache_fnodes",
979 979 _default_clear_on_disk_tags_fnodes_cache,
980 980 )
981 981 clear_fnodes_rev_fn = getattr(
982 982 tags,
983 983 "forget_fnodes",
984 984 _default_forget_fnodes,
985 985 )
986 986
987 987 clear_revs = []
988 988 if clear_fnode_revs:
989 989 clear_revs.extends(scmutil.revrange(repo, clear_fnode_revs))
990 990
991 991 if update_last:
992 992 revset = b'last(all(), %d)' % update_last
993 993 last_revs = repo.unfiltered().revs(revset)
994 994 clear_revs.extend(last_revs)
995 995
996 996 from mercurial import repoview
997 997
998 998 rev_filter = {(b'experimental', b'extra-filter-revs'): revset}
999 999 with repo.ui.configoverride(rev_filter, source=b"perf"):
1000 1000 filter_id = repoview.extrafilter(repo.ui)
1001 1001
1002 1002 filter_name = b'%s%%%s' % (repo.filtername, filter_id)
1003 1003 pre_repo = repo.filtered(filter_name)
1004 1004 pre_repo.tags() # warm the cache
1005 1005 old_tags_path = repo.cachevfs.join(tags._filename(pre_repo))
1006 1006 new_tags_path = repo.cachevfs.join(tags._filename(repo))
1007 1007
1008 1008 clear_revs = sorted(set(clear_revs))
1009 1009
1010 1010 def s():
1011 1011 if update_last:
1012 1012 util.copyfile(old_tags_path, new_tags_path)
1013 1013 if clearrevlogs:
1014 1014 clearchangelog(repo)
1015 1015 clearfilecache(repo.unfiltered(), 'manifest')
1016 1016 if clear_disk:
1017 1017 clear_disk_fn(repo)
1018 1018 if clear_fnode:
1019 1019 clear_fnodes_fn(repo)
1020 1020 elif clear_revs:
1021 1021 clear_fnodes_rev_fn(repo, clear_revs)
1022 1022 repocleartagscache()
1023 1023
1024 1024 def t():
1025 1025 len(repo.tags())
1026 1026
1027 1027 timer(t, setup=s)
1028 1028 fm.end()
1029 1029
1030 1030
1031 1031 @command(b'perf::ancestors|perfancestors', formatteropts)
1032 1032 def perfancestors(ui, repo, **opts):
1033 1033 opts = _byteskwargs(opts)
1034 1034 timer, fm = gettimer(ui, opts)
1035 1035 heads = repo.changelog.headrevs()
1036 1036
1037 1037 def d():
1038 1038 for a in repo.changelog.ancestors(heads):
1039 1039 pass
1040 1040
1041 1041 timer(d)
1042 1042 fm.end()
1043 1043
1044 1044
1045 1045 @command(b'perf::ancestorset|perfancestorset', formatteropts)
1046 1046 def perfancestorset(ui, repo, revset, **opts):
1047 1047 opts = _byteskwargs(opts)
1048 1048 timer, fm = gettimer(ui, opts)
1049 1049 revs = repo.revs(revset)
1050 1050 heads = repo.changelog.headrevs()
1051 1051
1052 1052 def d():
1053 1053 s = repo.changelog.ancestors(heads)
1054 1054 for rev in revs:
1055 1055 rev in s
1056 1056
1057 1057 timer(d)
1058 1058 fm.end()
1059 1059
1060 1060
1061 1061 @command(
1062 1062 b'perf::delta-find',
1063 1063 revlogopts + formatteropts,
1064 1064 b'-c|-m|FILE REV',
1065 1065 )
1066 1066 def perf_delta_find(ui, repo, arg_1, arg_2=None, **opts):
1067 1067 """benchmark the process of finding a valid delta for a revlog revision
1068 1068
1069 1069 When a revlog receives a new revision (e.g. from a commit, or from an
1070 1070 incoming bundle), it searches for a suitable delta-base to produce a delta.
1071 1071 This perf command measures how much time we spend in this process. It
1072 1072 operates on an already stored revision.
1073 1073
1074 1074 See `hg help debug-delta-find` for another related command.
1075 1075 """
1076 1076 from mercurial import revlogutils
1077 1077 import mercurial.revlogutils.deltas as deltautil
1078 1078
1079 1079 opts = _byteskwargs(opts)
1080 1080 if arg_2 is None:
1081 1081 file_ = None
1082 1082 rev = arg_1
1083 1083 else:
1084 1084 file_ = arg_1
1085 1085 rev = arg_2
1086 1086
1087 1087 repo = repo.unfiltered()
1088 1088
1089 1089 timer, fm = gettimer(ui, opts)
1090 1090
1091 1091 rev = int(rev)
1092 1092
1093 1093 revlog = cmdutil.openrevlog(repo, b'perf::delta-find', file_, opts)
1094 1094
1095 1095 deltacomputer = deltautil.deltacomputer(revlog)
1096 1096
1097 1097 node = revlog.node(rev)
1098 1098 p1r, p2r = revlog.parentrevs(rev)
1099 1099 p1 = revlog.node(p1r)
1100 1100 p2 = revlog.node(p2r)
1101 1101 full_text = revlog.revision(rev)
1102 1102 textlen = len(full_text)
1103 1103 cachedelta = None
1104 1104 flags = revlog.flags(rev)
1105 1105
1106 1106 revinfo = revlogutils.revisioninfo(
1107 1107 node,
1108 1108 p1,
1109 1109 p2,
1110 1110 [full_text], # btext
1111 1111 textlen,
1112 1112 cachedelta,
1113 1113 flags,
1114 1114 )
1115 1115
1116 1116 # Note: we should probably purge the potential caches (like the full
1117 1117 # manifest cache) between runs.
1118 1118 def find_one():
1119 1119 with revlog._datafp() as fh:
1120 1120 deltacomputer.finddeltainfo(revinfo, fh, target_rev=rev)
1121 1121
1122 1122 timer(find_one)
1123 1123 fm.end()
1124 1124
1125 1125
1126 1126 @command(b'perf::discovery|perfdiscovery', formatteropts, b'PATH')
1127 1127 def perfdiscovery(ui, repo, path, **opts):
1128 1128 """benchmark discovery between local repo and the peer at given path"""
1129 1129 repos = [repo, None]
1130 1130 timer, fm = gettimer(ui, opts)
1131 1131
1132 1132 try:
1133 1133 from mercurial.utils.urlutil import get_unique_pull_path_obj
1134 1134
1135 1135 path = get_unique_pull_path_obj(b'perfdiscovery', ui, path)
1136 1136 except ImportError:
1137 1137 try:
1138 1138 from mercurial.utils.urlutil import get_unique_pull_path
1139 1139
1140 1140 path = get_unique_pull_path(b'perfdiscovery', repo, ui, path)[0]
1141 1141 except ImportError:
1142 1142 path = ui.expandpath(path)
1143 1143
1144 1144 def s():
1145 1145 repos[1] = hg.peer(ui, opts, path)
1146 1146
1147 1147 def d():
1148 1148 setdiscovery.findcommonheads(ui, *repos)
1149 1149
1150 1150 timer(d, setup=s)
1151 1151 fm.end()
1152 1152
1153 1153
1154 1154 @command(
1155 1155 b'perf::bookmarks|perfbookmarks',
1156 1156 formatteropts
1157 1157 + [
1158 1158 (b'', b'clear-revlogs', False, b'refresh changelog and manifest'),
1159 1159 ],
1160 1160 )
1161 1161 def perfbookmarks(ui, repo, **opts):
1162 1162 """benchmark parsing bookmarks from disk to memory"""
1163 1163 opts = _byteskwargs(opts)
1164 1164 timer, fm = gettimer(ui, opts)
1165 1165
1166 1166 clearrevlogs = opts[b'clear_revlogs']
1167 1167
1168 1168 def s():
1169 1169 if clearrevlogs:
1170 1170 clearchangelog(repo)
1171 1171 clearfilecache(repo, b'_bookmarks')
1172 1172
1173 1173 def d():
1174 1174 repo._bookmarks
1175 1175
1176 1176 timer(d, setup=s)
1177 1177 fm.end()
1178 1178
1179 1179
1180 1180 @command(
1181 1181 b'perf::bundle',
1182 1182 [
1183 1183 (
1184 1184 b'r',
1185 1185 b'rev',
1186 1186 [],
1187 1187 b'changesets to bundle',
1188 1188 b'REV',
1189 1189 ),
1190 1190 (
1191 1191 b't',
1192 1192 b'type',
1193 1193 b'none',
1194 1194 b'bundlespec to use (see `hg help bundlespec`)',
1195 1195 b'TYPE',
1196 1196 ),
1197 1197 ]
1198 1198 + formatteropts,
1199 1199 b'REVS',
1200 1200 )
1201 1201 def perfbundle(ui, repo, *revs, **opts):
1202 1202 """benchmark the creation of a bundle from a repository
1203 1203
1204 1204 For now, this only supports "none" compression.
1205 1205 """
1206 1206 try:
1207 1207 from mercurial import bundlecaches
1208 1208
1209 1209 parsebundlespec = bundlecaches.parsebundlespec
1210 1210 except ImportError:
1211 1211 from mercurial import exchange
1212 1212
1213 1213 parsebundlespec = exchange.parsebundlespec
1214 1214
1215 1215 from mercurial import discovery
1216 1216 from mercurial import bundle2
1217 1217
1218 1218 opts = _byteskwargs(opts)
1219 1219 timer, fm = gettimer(ui, opts)
1220 1220
1221 1221 cl = repo.changelog
1222 1222 revs = list(revs)
1223 1223 revs.extend(opts.get(b'rev', ()))
1224 1224 revs = scmutil.revrange(repo, revs)
1225 1225 if not revs:
1226 1226 raise error.Abort(b"not revision specified")
1227 1227 # make it a consistent set (ie: without topological gaps)
1228 1228 old_len = len(revs)
1229 1229 revs = list(repo.revs(b"%ld::%ld", revs, revs))
1230 1230 if old_len != len(revs):
1231 1231 new_count = len(revs) - old_len
1232 1232 msg = b"add %d new revisions to make it a consistent set\n"
1233 1233 ui.write_err(msg % new_count)
1234 1234
1235 1235 targets = [cl.node(r) for r in repo.revs(b"heads(::%ld)", revs)]
1236 1236 bases = [cl.node(r) for r in repo.revs(b"heads(::%ld - %ld)", revs, revs)]
1237 1237 outgoing = discovery.outgoing(repo, bases, targets)
1238 1238
1239 1239 bundle_spec = opts.get(b'type')
1240 1240
1241 1241 bundle_spec = parsebundlespec(repo, bundle_spec, strict=False)
1242 1242
1243 1243 cgversion = bundle_spec.params.get(b"cg.version")
1244 1244 if cgversion is None:
1245 1245 if bundle_spec.version == b'v1':
1246 1246 cgversion = b'01'
1247 1247 if bundle_spec.version == b'v2':
1248 1248 cgversion = b'02'
1249 1249 if cgversion not in changegroup.supportedoutgoingversions(repo):
1250 1250 err = b"repository does not support bundle version %s"
1251 1251 raise error.Abort(err % cgversion)
1252 1252
1253 1253 if cgversion == b'01': # bundle1
1254 1254 bversion = b'HG10' + bundle_spec.wirecompression
1255 1255 bcompression = None
1256 1256 elif cgversion in (b'02', b'03'):
1257 1257 bversion = b'HG20'
1258 1258 bcompression = bundle_spec.wirecompression
1259 1259 else:
1260 1260 err = b'perf::bundle: unexpected changegroup version %s'
1261 1261 raise error.ProgrammingError(err % cgversion)
1262 1262
1263 1263 if bcompression is None:
1264 1264 bcompression = b'UN'
1265 1265
1266 1266 if bcompression != b'UN':
1267 1267 err = b'perf::bundle: compression currently unsupported: %s'
1268 1268 raise error.ProgrammingError(err % bcompression)
1269 1269
1270 1270 def do_bundle():
1271 1271 bundle2.writenewbundle(
1272 1272 ui,
1273 1273 repo,
1274 1274 b'perf::bundle',
1275 1275 os.devnull,
1276 1276 bversion,
1277 1277 outgoing,
1278 1278 bundle_spec.params,
1279 1279 )
1280 1280
1281 1281 timer(do_bundle)
1282 1282 fm.end()
1283 1283
1284 1284
1285 1285 @command(b'perf::bundleread|perfbundleread', formatteropts, b'BUNDLE')
1286 1286 def perfbundleread(ui, repo, bundlepath, **opts):
1287 1287 """Benchmark reading of bundle files.
1288 1288
1289 1289 This command is meant to isolate the I/O part of bundle reading as
1290 1290 much as possible.
1291 1291 """
1292 1292 from mercurial import (
1293 1293 bundle2,
1294 1294 exchange,
1295 1295 streamclone,
1296 1296 )
1297 1297
1298 1298 opts = _byteskwargs(opts)
1299 1299
1300 1300 def makebench(fn):
1301 1301 def run():
1302 1302 with open(bundlepath, b'rb') as fh:
1303 1303 bundle = exchange.readbundle(ui, fh, bundlepath)
1304 1304 fn(bundle)
1305 1305
1306 1306 return run
1307 1307
1308 1308 def makereadnbytes(size):
1309 1309 def run():
1310 1310 with open(bundlepath, b'rb') as fh:
1311 1311 bundle = exchange.readbundle(ui, fh, bundlepath)
1312 1312 while bundle.read(size):
1313 1313 pass
1314 1314
1315 1315 return run
1316 1316
1317 1317 def makestdioread(size):
1318 1318 def run():
1319 1319 with open(bundlepath, b'rb') as fh:
1320 1320 while fh.read(size):
1321 1321 pass
1322 1322
1323 1323 return run
1324 1324
1325 1325 # bundle1
1326 1326
1327 1327 def deltaiter(bundle):
1328 1328 for delta in bundle.deltaiter():
1329 1329 pass
1330 1330
1331 1331 def iterchunks(bundle):
1332 1332 for chunk in bundle.getchunks():
1333 1333 pass
1334 1334
1335 1335 # bundle2
1336 1336
1337 1337 def forwardchunks(bundle):
1338 1338 for chunk in bundle._forwardchunks():
1339 1339 pass
1340 1340
1341 1341 def iterparts(bundle):
1342 1342 for part in bundle.iterparts():
1343 1343 pass
1344 1344
1345 1345 def iterpartsseekable(bundle):
1346 1346 for part in bundle.iterparts(seekable=True):
1347 1347 pass
1348 1348
1349 1349 def seek(bundle):
1350 1350 for part in bundle.iterparts(seekable=True):
1351 1351 part.seek(0, os.SEEK_END)
1352 1352
1353 1353 def makepartreadnbytes(size):
1354 1354 def run():
1355 1355 with open(bundlepath, b'rb') as fh:
1356 1356 bundle = exchange.readbundle(ui, fh, bundlepath)
1357 1357 for part in bundle.iterparts():
1358 1358 while part.read(size):
1359 1359 pass
1360 1360
1361 1361 return run
1362 1362
1363 1363 benches = [
1364 1364 (makestdioread(8192), b'read(8k)'),
1365 1365 (makestdioread(16384), b'read(16k)'),
1366 1366 (makestdioread(32768), b'read(32k)'),
1367 1367 (makestdioread(131072), b'read(128k)'),
1368 1368 ]
1369 1369
1370 1370 with open(bundlepath, b'rb') as fh:
1371 1371 bundle = exchange.readbundle(ui, fh, bundlepath)
1372 1372
1373 1373 if isinstance(bundle, changegroup.cg1unpacker):
1374 1374 benches.extend(
1375 1375 [
1376 1376 (makebench(deltaiter), b'cg1 deltaiter()'),
1377 1377 (makebench(iterchunks), b'cg1 getchunks()'),
1378 1378 (makereadnbytes(8192), b'cg1 read(8k)'),
1379 1379 (makereadnbytes(16384), b'cg1 read(16k)'),
1380 1380 (makereadnbytes(32768), b'cg1 read(32k)'),
1381 1381 (makereadnbytes(131072), b'cg1 read(128k)'),
1382 1382 ]
1383 1383 )
1384 1384 elif isinstance(bundle, bundle2.unbundle20):
1385 1385 benches.extend(
1386 1386 [
1387 1387 (makebench(forwardchunks), b'bundle2 forwardchunks()'),
1388 1388 (makebench(iterparts), b'bundle2 iterparts()'),
1389 1389 (
1390 1390 makebench(iterpartsseekable),
1391 1391 b'bundle2 iterparts() seekable',
1392 1392 ),
1393 1393 (makebench(seek), b'bundle2 part seek()'),
1394 1394 (makepartreadnbytes(8192), b'bundle2 part read(8k)'),
1395 1395 (makepartreadnbytes(16384), b'bundle2 part read(16k)'),
1396 1396 (makepartreadnbytes(32768), b'bundle2 part read(32k)'),
1397 1397 (makepartreadnbytes(131072), b'bundle2 part read(128k)'),
1398 1398 ]
1399 1399 )
1400 1400 elif isinstance(bundle, streamclone.streamcloneapplier):
1401 1401 raise error.Abort(b'stream clone bundles not supported')
1402 1402 else:
1403 1403 raise error.Abort(b'unhandled bundle type: %s' % type(bundle))
1404 1404
1405 1405 for fn, title in benches:
1406 1406 timer, fm = gettimer(ui, opts)
1407 1407 timer(fn, title=title)
1408 1408 fm.end()
1409 1409
1410 1410
1411 1411 @command(
1412 1412 b'perf::changegroupchangelog|perfchangegroupchangelog',
1413 1413 formatteropts
1414 1414 + [
1415 1415 (b'', b'cgversion', b'02', b'changegroup version'),
1416 1416 (b'r', b'rev', b'', b'revisions to add to changegroup'),
1417 1417 ],
1418 1418 )
1419 1419 def perfchangegroupchangelog(ui, repo, cgversion=b'02', rev=None, **opts):
1420 1420 """Benchmark producing a changelog group for a changegroup.
1421 1421
1422 1422 This measures the time spent processing the changelog during a
1423 1423 bundle operation. This occurs during `hg bundle` and on a server
1424 1424 processing a `getbundle` wire protocol request (handles clones
1425 1425 and pull requests).
1426 1426
1427 1427 By default, all revisions are added to the changegroup.
1428 1428 """
1429 1429 opts = _byteskwargs(opts)
1430 1430 cl = repo.changelog
1431 1431 nodes = [cl.lookup(r) for r in repo.revs(rev or b'all()')]
1432 1432 bundler = changegroup.getbundler(cgversion, repo)
1433 1433
1434 1434 def d():
1435 1435 state, chunks = bundler._generatechangelog(cl, nodes)
1436 1436 for chunk in chunks:
1437 1437 pass
1438 1438
1439 1439 timer, fm = gettimer(ui, opts)
1440 1440
1441 1441 # Terminal printing can interfere with timing. So disable it.
1442 1442 with ui.configoverride({(b'progress', b'disable'): True}):
1443 1443 timer(d)
1444 1444
1445 1445 fm.end()
1446 1446
1447 1447
1448 1448 @command(b'perf::dirs|perfdirs', formatteropts)
1449 1449 def perfdirs(ui, repo, **opts):
1450 1450 opts = _byteskwargs(opts)
1451 1451 timer, fm = gettimer(ui, opts)
1452 1452 dirstate = repo.dirstate
1453 1453 b'a' in dirstate
1454 1454
1455 1455 def d():
1456 1456 dirstate.hasdir(b'a')
1457 1457 try:
1458 1458 del dirstate._map._dirs
1459 1459 except AttributeError:
1460 1460 pass
1461 1461
1462 1462 timer(d)
1463 1463 fm.end()
1464 1464
1465 1465
1466 1466 @command(
1467 1467 b'perf::dirstate|perfdirstate',
1468 1468 [
1469 1469 (
1470 1470 b'',
1471 1471 b'iteration',
1472 1472 None,
1473 1473 b'benchmark a full iteration for the dirstate',
1474 1474 ),
1475 1475 (
1476 1476 b'',
1477 1477 b'contains',
1478 1478 None,
1479 1479 b'benchmark a large amount of `nf in dirstate` calls',
1480 1480 ),
1481 1481 ]
1482 1482 + formatteropts,
1483 1483 )
1484 1484 def perfdirstate(ui, repo, **opts):
1485 1485 """benchmap the time of various distate operations
1486 1486
1487 1487 By default benchmark the time necessary to load a dirstate from scratch.
1488 1488 The dirstate is loaded to the point were a "contains" request can be
1489 1489 answered.
1490 1490 """
1491 1491 opts = _byteskwargs(opts)
1492 1492 timer, fm = gettimer(ui, opts)
1493 1493 b"a" in repo.dirstate
1494 1494
1495 1495 if opts[b'iteration'] and opts[b'contains']:
1496 1496 msg = b'only specify one of --iteration or --contains'
1497 1497 raise error.Abort(msg)
1498 1498
1499 1499 if opts[b'iteration']:
1500 1500 setup = None
1501 1501 dirstate = repo.dirstate
1502 1502
1503 1503 def d():
1504 1504 for f in dirstate:
1505 1505 pass
1506 1506
1507 1507 elif opts[b'contains']:
1508 1508 setup = None
1509 1509 dirstate = repo.dirstate
1510 1510 allfiles = list(dirstate)
1511 1511 # also add file path that will be "missing" from the dirstate
1512 1512 allfiles.extend([f[::-1] for f in allfiles])
1513 1513
1514 1514 def d():
1515 1515 for f in allfiles:
1516 1516 f in dirstate
1517 1517
1518 1518 else:
1519 1519
1520 1520 def setup():
1521 1521 repo.dirstate.invalidate()
1522 1522
1523 1523 def d():
1524 1524 b"a" in repo.dirstate
1525 1525
1526 1526 timer(d, setup=setup)
1527 1527 fm.end()
1528 1528
1529 1529
1530 1530 @command(b'perf::dirstatedirs|perfdirstatedirs', formatteropts)
1531 1531 def perfdirstatedirs(ui, repo, **opts):
1532 1532 """benchmap a 'dirstate.hasdir' call from an empty `dirs` cache"""
1533 1533 opts = _byteskwargs(opts)
1534 1534 timer, fm = gettimer(ui, opts)
1535 1535 repo.dirstate.hasdir(b"a")
1536 1536
1537 1537 def setup():
1538 1538 try:
1539 1539 del repo.dirstate._map._dirs
1540 1540 except AttributeError:
1541 1541 pass
1542 1542
1543 1543 def d():
1544 1544 repo.dirstate.hasdir(b"a")
1545 1545
1546 1546 timer(d, setup=setup)
1547 1547 fm.end()
1548 1548
1549 1549
1550 1550 @command(b'perf::dirstatefoldmap|perfdirstatefoldmap', formatteropts)
1551 1551 def perfdirstatefoldmap(ui, repo, **opts):
1552 1552 """benchmap a `dirstate._map.filefoldmap.get()` request
1553 1553
1554 1554 The dirstate filefoldmap cache is dropped between every request.
1555 1555 """
1556 1556 opts = _byteskwargs(opts)
1557 1557 timer, fm = gettimer(ui, opts)
1558 1558 dirstate = repo.dirstate
1559 1559 dirstate._map.filefoldmap.get(b'a')
1560 1560
1561 1561 def setup():
1562 1562 del dirstate._map.filefoldmap
1563 1563
1564 1564 def d():
1565 1565 dirstate._map.filefoldmap.get(b'a')
1566 1566
1567 1567 timer(d, setup=setup)
1568 1568 fm.end()
1569 1569
1570 1570
1571 1571 @command(b'perf::dirfoldmap|perfdirfoldmap', formatteropts)
1572 1572 def perfdirfoldmap(ui, repo, **opts):
1573 1573 """benchmap a `dirstate._map.dirfoldmap.get()` request
1574 1574
1575 1575 The dirstate dirfoldmap cache is dropped between every request.
1576 1576 """
1577 1577 opts = _byteskwargs(opts)
1578 1578 timer, fm = gettimer(ui, opts)
1579 1579 dirstate = repo.dirstate
1580 1580 dirstate._map.dirfoldmap.get(b'a')
1581 1581
1582 1582 def setup():
1583 1583 del dirstate._map.dirfoldmap
1584 1584 try:
1585 1585 del dirstate._map._dirs
1586 1586 except AttributeError:
1587 1587 pass
1588 1588
1589 1589 def d():
1590 1590 dirstate._map.dirfoldmap.get(b'a')
1591 1591
1592 1592 timer(d, setup=setup)
1593 1593 fm.end()
1594 1594
1595 1595
1596 1596 @command(b'perf::dirstatewrite|perfdirstatewrite', formatteropts)
1597 1597 def perfdirstatewrite(ui, repo, **opts):
1598 1598 """benchmap the time it take to write a dirstate on disk"""
1599 1599 opts = _byteskwargs(opts)
1600 1600 timer, fm = gettimer(ui, opts)
1601 1601 ds = repo.dirstate
1602 1602 b"a" in ds
1603 1603
1604 1604 def setup():
1605 1605 ds._dirty = True
1606 1606
1607 1607 def d():
1608 1608 ds.write(repo.currenttransaction())
1609 1609
1610 1610 with repo.wlock():
1611 1611 timer(d, setup=setup)
1612 1612 fm.end()
1613 1613
1614 1614
1615 1615 def _getmergerevs(repo, opts):
1616 1616 """parse command argument to return rev involved in merge
1617 1617
1618 1618 input: options dictionnary with `rev`, `from` and `bse`
1619 1619 output: (localctx, otherctx, basectx)
1620 1620 """
1621 1621 if opts[b'from']:
1622 1622 fromrev = scmutil.revsingle(repo, opts[b'from'])
1623 1623 wctx = repo[fromrev]
1624 1624 else:
1625 1625 wctx = repo[None]
1626 1626 # we don't want working dir files to be stat'd in the benchmark, so
1627 1627 # prime that cache
1628 1628 wctx.dirty()
1629 1629 rctx = scmutil.revsingle(repo, opts[b'rev'], opts[b'rev'])
1630 1630 if opts[b'base']:
1631 1631 fromrev = scmutil.revsingle(repo, opts[b'base'])
1632 1632 ancestor = repo[fromrev]
1633 1633 else:
1634 1634 ancestor = wctx.ancestor(rctx)
1635 1635 return (wctx, rctx, ancestor)
1636 1636
1637 1637
1638 1638 @command(
1639 1639 b'perf::mergecalculate|perfmergecalculate',
1640 1640 [
1641 1641 (b'r', b'rev', b'.', b'rev to merge against'),
1642 1642 (b'', b'from', b'', b'rev to merge from'),
1643 1643 (b'', b'base', b'', b'the revision to use as base'),
1644 1644 ]
1645 1645 + formatteropts,
1646 1646 )
1647 1647 def perfmergecalculate(ui, repo, **opts):
1648 1648 opts = _byteskwargs(opts)
1649 1649 timer, fm = gettimer(ui, opts)
1650 1650
1651 1651 wctx, rctx, ancestor = _getmergerevs(repo, opts)
1652 1652
1653 1653 def d():
1654 1654 # acceptremote is True because we don't want prompts in the middle of
1655 1655 # our benchmark
1656 1656 merge.calculateupdates(
1657 1657 repo,
1658 1658 wctx,
1659 1659 rctx,
1660 1660 [ancestor],
1661 1661 branchmerge=False,
1662 1662 force=False,
1663 1663 acceptremote=True,
1664 1664 followcopies=True,
1665 1665 )
1666 1666
1667 1667 timer(d)
1668 1668 fm.end()
1669 1669
1670 1670
1671 1671 @command(
1672 1672 b'perf::mergecopies|perfmergecopies',
1673 1673 [
1674 1674 (b'r', b'rev', b'.', b'rev to merge against'),
1675 1675 (b'', b'from', b'', b'rev to merge from'),
1676 1676 (b'', b'base', b'', b'the revision to use as base'),
1677 1677 ]
1678 1678 + formatteropts,
1679 1679 )
1680 1680 def perfmergecopies(ui, repo, **opts):
1681 1681 """measure runtime of `copies.mergecopies`"""
1682 1682 opts = _byteskwargs(opts)
1683 1683 timer, fm = gettimer(ui, opts)
1684 1684 wctx, rctx, ancestor = _getmergerevs(repo, opts)
1685 1685
1686 1686 def d():
1687 1687 # acceptremote is True because we don't want prompts in the middle of
1688 1688 # our benchmark
1689 1689 copies.mergecopies(repo, wctx, rctx, ancestor)
1690 1690
1691 1691 timer(d)
1692 1692 fm.end()
1693 1693
1694 1694
1695 1695 @command(b'perf::pathcopies|perfpathcopies', [], b"REV REV")
1696 1696 def perfpathcopies(ui, repo, rev1, rev2, **opts):
1697 1697 """benchmark the copy tracing logic"""
1698 1698 opts = _byteskwargs(opts)
1699 1699 timer, fm = gettimer(ui, opts)
1700 1700 ctx1 = scmutil.revsingle(repo, rev1, rev1)
1701 1701 ctx2 = scmutil.revsingle(repo, rev2, rev2)
1702 1702
1703 1703 def d():
1704 1704 copies.pathcopies(ctx1, ctx2)
1705 1705
1706 1706 timer(d)
1707 1707 fm.end()
1708 1708
1709 1709
1710 1710 @command(
1711 1711 b'perf::phases|perfphases',
1712 1712 [
1713 1713 (b'', b'full', False, b'include file reading time too'),
1714 1714 ],
1715 1715 b"",
1716 1716 )
1717 1717 def perfphases(ui, repo, **opts):
1718 1718 """benchmark phasesets computation"""
1719 1719 opts = _byteskwargs(opts)
1720 1720 timer, fm = gettimer(ui, opts)
1721 1721 _phases = repo._phasecache
1722 1722 full = opts.get(b'full')
1723 1723
1724 1724 def d():
1725 1725 phases = _phases
1726 1726 if full:
1727 1727 clearfilecache(repo, b'_phasecache')
1728 1728 phases = repo._phasecache
1729 1729 phases.invalidate()
1730 1730 phases.loadphaserevs(repo)
1731 1731
1732 1732 timer(d)
1733 1733 fm.end()
1734 1734
1735 1735
1736 1736 @command(b'perf::phasesremote|perfphasesremote', [], b"[DEST]")
1737 1737 def perfphasesremote(ui, repo, dest=None, **opts):
1738 1738 """benchmark time needed to analyse phases of the remote server"""
1739 1739 from mercurial.node import bin
1740 1740 from mercurial import (
1741 1741 exchange,
1742 1742 hg,
1743 1743 phases,
1744 1744 )
1745 1745
1746 1746 opts = _byteskwargs(opts)
1747 1747 timer, fm = gettimer(ui, opts)
1748 1748
1749 1749 path = ui.getpath(dest, default=(b'default-push', b'default'))
1750 1750 if not path:
1751 1751 raise error.Abort(
1752 1752 b'default repository not configured!',
1753 1753 hint=b"see 'hg help config.paths'",
1754 1754 )
1755 1755 if util.safehasattr(path, 'main_path'):
1756 1756 path = path.get_push_variant()
1757 1757 dest = path.loc
1758 1758 else:
1759 1759 dest = path.pushloc or path.loc
1760 1760 ui.statusnoi18n(b'analysing phase of %s\n' % util.hidepassword(dest))
1761 1761 other = hg.peer(repo, opts, dest)
1762 1762
1763 1763 # easier to perform discovery through the operation
1764 1764 op = exchange.pushoperation(repo, other)
1765 1765 exchange._pushdiscoverychangeset(op)
1766 1766
1767 1767 remotesubset = op.fallbackheads
1768 1768
1769 1769 with other.commandexecutor() as e:
1770 1770 remotephases = e.callcommand(
1771 1771 b'listkeys', {b'namespace': b'phases'}
1772 1772 ).result()
1773 1773 del other
1774 1774 publishing = remotephases.get(b'publishing', False)
1775 1775 if publishing:
1776 1776 ui.statusnoi18n(b'publishing: yes\n')
1777 1777 else:
1778 1778 ui.statusnoi18n(b'publishing: no\n')
1779 1779
1780 1780 has_node = getattr(repo.changelog.index, 'has_node', None)
1781 1781 if has_node is None:
1782 1782 has_node = repo.changelog.nodemap.__contains__
1783 1783 nonpublishroots = 0
1784 1784 for nhex, phase in remotephases.iteritems():
1785 1785 if nhex == b'publishing': # ignore data related to publish option
1786 1786 continue
1787 1787 node = bin(nhex)
1788 1788 if has_node(node) and int(phase):
1789 1789 nonpublishroots += 1
1790 1790 ui.statusnoi18n(b'number of roots: %d\n' % len(remotephases))
1791 1791 ui.statusnoi18n(b'number of known non public roots: %d\n' % nonpublishroots)
1792 1792
1793 1793 def d():
1794 1794 phases.remotephasessummary(repo, remotesubset, remotephases)
1795 1795
1796 1796 timer(d)
1797 1797 fm.end()
1798 1798
1799 1799
1800 1800 @command(
1801 1801 b'perf::manifest|perfmanifest',
1802 1802 [
1803 1803 (b'm', b'manifest-rev', False, b'Look up a manifest node revision'),
1804 1804 (b'', b'clear-disk', False, b'clear on-disk caches too'),
1805 1805 ]
1806 1806 + formatteropts,
1807 1807 b'REV|NODE',
1808 1808 )
1809 1809 def perfmanifest(ui, repo, rev, manifest_rev=False, clear_disk=False, **opts):
1810 1810 """benchmark the time to read a manifest from disk and return a usable
1811 1811 dict-like object
1812 1812
1813 1813 Manifest caches are cleared before retrieval."""
1814 1814 opts = _byteskwargs(opts)
1815 1815 timer, fm = gettimer(ui, opts)
1816 1816 if not manifest_rev:
1817 1817 ctx = scmutil.revsingle(repo, rev, rev)
1818 1818 t = ctx.manifestnode()
1819 1819 else:
1820 1820 from mercurial.node import bin
1821 1821
1822 1822 if len(rev) == 40:
1823 1823 t = bin(rev)
1824 1824 else:
1825 1825 try:
1826 1826 rev = int(rev)
1827 1827
1828 1828 if util.safehasattr(repo.manifestlog, b'getstorage'):
1829 1829 t = repo.manifestlog.getstorage(b'').node(rev)
1830 1830 else:
1831 1831 t = repo.manifestlog._revlog.lookup(rev)
1832 1832 except ValueError:
1833 1833 raise error.Abort(
1834 1834 b'manifest revision must be integer or full node'
1835 1835 )
1836 1836
1837 1837 def d():
1838 1838 repo.manifestlog.clearcaches(clear_persisted_data=clear_disk)
1839 1839 repo.manifestlog[t].read()
1840 1840
1841 1841 timer(d)
1842 1842 fm.end()
1843 1843
1844 1844
1845 1845 @command(b'perf::changeset|perfchangeset', formatteropts)
1846 1846 def perfchangeset(ui, repo, rev, **opts):
1847 1847 opts = _byteskwargs(opts)
1848 1848 timer, fm = gettimer(ui, opts)
1849 1849 n = scmutil.revsingle(repo, rev).node()
1850 1850
1851 1851 def d():
1852 1852 repo.changelog.read(n)
1853 1853 # repo.changelog._cache = None
1854 1854
1855 1855 timer(d)
1856 1856 fm.end()
1857 1857
1858 1858
1859 1859 @command(b'perf::ignore|perfignore', formatteropts)
1860 1860 def perfignore(ui, repo, **opts):
1861 1861 """benchmark operation related to computing ignore"""
1862 1862 opts = _byteskwargs(opts)
1863 1863 timer, fm = gettimer(ui, opts)
1864 1864 dirstate = repo.dirstate
1865 1865
1866 1866 def setupone():
1867 1867 dirstate.invalidate()
1868 1868 clearfilecache(dirstate, b'_ignore')
1869 1869
1870 1870 def runone():
1871 1871 dirstate._ignore
1872 1872
1873 1873 timer(runone, setup=setupone, title=b"load")
1874 1874 fm.end()
1875 1875
1876 1876
1877 1877 @command(
1878 1878 b'perf::index|perfindex',
1879 1879 [
1880 1880 (b'', b'rev', [], b'revision to be looked up (default tip)'),
1881 1881 (b'', b'no-lookup', None, b'do not revision lookup post creation'),
1882 1882 ]
1883 1883 + formatteropts,
1884 1884 )
1885 1885 def perfindex(ui, repo, **opts):
1886 1886 """benchmark index creation time followed by a lookup
1887 1887
1888 1888 The default is to look `tip` up. Depending on the index implementation,
1889 1889 the revision looked up can matters. For example, an implementation
1890 1890 scanning the index will have a faster lookup time for `--rev tip` than for
1891 1891 `--rev 0`. The number of looked up revisions and their order can also
1892 1892 matters.
1893 1893
1894 1894 Example of useful set to test:
1895 1895
1896 1896 * tip
1897 1897 * 0
1898 1898 * -10:
1899 1899 * :10
1900 1900 * -10: + :10
1901 1901 * :10: + -10:
1902 1902 * -10000:
1903 1903 * -10000: + 0
1904 1904
1905 1905 It is not currently possible to check for lookup of a missing node. For
1906 1906 deeper lookup benchmarking, checkout the `perfnodemap` command."""
1907 1907 import mercurial.revlog
1908 1908
1909 1909 opts = _byteskwargs(opts)
1910 1910 timer, fm = gettimer(ui, opts)
1911 1911 mercurial.revlog._prereadsize = 2 ** 24 # disable lazy parser in old hg
1912 1912 if opts[b'no_lookup']:
1913 1913 if opts['rev']:
1914 1914 raise error.Abort('--no-lookup and --rev are mutually exclusive')
1915 1915 nodes = []
1916 1916 elif not opts[b'rev']:
1917 1917 nodes = [repo[b"tip"].node()]
1918 1918 else:
1919 1919 revs = scmutil.revrange(repo, opts[b'rev'])
1920 1920 cl = repo.changelog
1921 1921 nodes = [cl.node(r) for r in revs]
1922 1922
1923 1923 unfi = repo.unfiltered()
1924 1924 # find the filecache func directly
1925 1925 # This avoid polluting the benchmark with the filecache logic
1926 1926 makecl = unfi.__class__.changelog.func
1927 1927
1928 1928 def setup():
1929 1929 # probably not necessary, but for good measure
1930 1930 clearchangelog(unfi)
1931 1931
1932 1932 def d():
1933 1933 cl = makecl(unfi)
1934 1934 for n in nodes:
1935 1935 cl.rev(n)
1936 1936
1937 1937 timer(d, setup=setup)
1938 1938 fm.end()
1939 1939
1940 1940
1941 1941 @command(
1942 1942 b'perf::nodemap|perfnodemap',
1943 1943 [
1944 1944 (b'', b'rev', [], b'revision to be looked up (default tip)'),
1945 1945 (b'', b'clear-caches', True, b'clear revlog cache between calls'),
1946 1946 ]
1947 1947 + formatteropts,
1948 1948 )
1949 1949 def perfnodemap(ui, repo, **opts):
1950 1950 """benchmark the time necessary to look up revision from a cold nodemap
1951 1951
1952 1952 Depending on the implementation, the amount and order of revision we look
1953 1953 up can varies. Example of useful set to test:
1954 1954 * tip
1955 1955 * 0
1956 1956 * -10:
1957 1957 * :10
1958 1958 * -10: + :10
1959 1959 * :10: + -10:
1960 1960 * -10000:
1961 1961 * -10000: + 0
1962 1962
1963 1963 The command currently focus on valid binary lookup. Benchmarking for
1964 1964 hexlookup, prefix lookup and missing lookup would also be valuable.
1965 1965 """
1966 1966 import mercurial.revlog
1967 1967
1968 1968 opts = _byteskwargs(opts)
1969 1969 timer, fm = gettimer(ui, opts)
1970 1970 mercurial.revlog._prereadsize = 2 ** 24 # disable lazy parser in old hg
1971 1971
1972 1972 unfi = repo.unfiltered()
1973 1973 clearcaches = opts[b'clear_caches']
1974 1974 # find the filecache func directly
1975 1975 # This avoid polluting the benchmark with the filecache logic
1976 1976 makecl = unfi.__class__.changelog.func
1977 1977 if not opts[b'rev']:
1978 1978 raise error.Abort(b'use --rev to specify revisions to look up')
1979 1979 revs = scmutil.revrange(repo, opts[b'rev'])
1980 1980 cl = repo.changelog
1981 1981 nodes = [cl.node(r) for r in revs]
1982 1982
1983 1983 # use a list to pass reference to a nodemap from one closure to the next
1984 1984 nodeget = [None]
1985 1985
1986 1986 def setnodeget():
1987 1987 # probably not necessary, but for good measure
1988 1988 clearchangelog(unfi)
1989 1989 cl = makecl(unfi)
1990 1990 if util.safehasattr(cl.index, 'get_rev'):
1991 1991 nodeget[0] = cl.index.get_rev
1992 1992 else:
1993 1993 nodeget[0] = cl.nodemap.get
1994 1994
1995 1995 def d():
1996 1996 get = nodeget[0]
1997 1997 for n in nodes:
1998 1998 get(n)
1999 1999
2000 2000 setup = None
2001 2001 if clearcaches:
2002 2002
2003 2003 def setup():
2004 2004 setnodeget()
2005 2005
2006 2006 else:
2007 2007 setnodeget()
2008 2008 d() # prewarm the data structure
2009 2009 timer(d, setup=setup)
2010 2010 fm.end()
2011 2011
2012 2012
2013 2013 @command(b'perf::startup|perfstartup', formatteropts)
2014 2014 def perfstartup(ui, repo, **opts):
2015 2015 opts = _byteskwargs(opts)
2016 2016 timer, fm = gettimer(ui, opts)
2017 2017
2018 2018 def d():
2019 2019 if os.name != 'nt':
2020 2020 os.system(
2021 2021 b"HGRCPATH= %s version -q > /dev/null" % fsencode(sys.argv[0])
2022 2022 )
2023 2023 else:
2024 2024 os.environ['HGRCPATH'] = r' '
2025 2025 os.system("%s version -q > NUL" % sys.argv[0])
2026 2026
2027 2027 timer(d)
2028 2028 fm.end()
2029 2029
2030 2030
2031 2031 def _find_stream_generator(version):
2032 2032 """find the proper generator function for this stream version"""
2033 2033 import mercurial.streamclone
2034 2034
2035 2035 available = {}
2036 2036
2037 2037 # try to fetch a v1 generator
2038 2038 generatev1 = getattr(mercurial.streamclone, "generatev1", None)
2039 2039 if generatev1 is not None:
2040 2040
2041 2041 def generate(repo):
2042 2042 entries, bytes, data = generatev2(repo, None, None, True)
2043 2043 return data
2044 2044
2045 2045 available[b'v1'] = generatev1
2046 2046 # try to fetch a v2 generator
2047 2047 generatev2 = getattr(mercurial.streamclone, "generatev2", None)
2048 2048 if generatev2 is not None:
2049 2049
2050 2050 def generate(repo):
2051 2051 entries, bytes, data = generatev2(repo, None, None, True)
2052 2052 return data
2053 2053
2054 2054 available[b'v2'] = generate
2055 2055 # try to fetch a v3 generator
2056 2056 generatev3 = getattr(mercurial.streamclone, "generatev3", None)
2057 2057 if generatev3 is not None:
2058 2058
2059 2059 def generate(repo):
2060 2060 entries, bytes, data = generatev3(repo, None, None, True)
2061 2061 return data
2062 2062
2063 2063 available[b'v3-exp'] = generate
2064 2064
2065 2065 # resolve the request
2066 2066 if version == b"latest":
2067 2067 # latest is the highest non experimental version
2068 2068 latest_key = max(v for v in available if b'-exp' not in v)
2069 2069 return available[latest_key]
2070 2070 elif version in available:
2071 2071 return available[version]
2072 2072 else:
2073 2073 msg = b"unkown or unavailable version: %s"
2074 2074 msg %= version
2075 2075 hint = b"available versions: %s"
2076 2076 hint %= b', '.join(sorted(available))
2077 2077 raise error.Abort(msg, hint=hint)
2078 2078
2079 2079
2080 2080 @command(
2081 2081 b'perf::stream-locked-section',
2082 2082 [
2083 2083 (
2084 2084 b'',
2085 2085 b'stream-version',
2086 2086 b'latest',
2087 2087 b'stream version to use ("v1", "v2", "v3" or "latest", (the default))',
2088 2088 ),
2089 2089 ]
2090 2090 + formatteropts,
2091 2091 )
2092 2092 def perf_stream_clone_scan(ui, repo, stream_version, **opts):
2093 2093 """benchmark the initial, repo-locked, section of a stream-clone"""
2094 2094
2095 2095 opts = _byteskwargs(opts)
2096 2096 timer, fm = gettimer(ui, opts)
2097 2097
2098 2098 # deletion of the generator may trigger some cleanup that we do not want to
2099 2099 # measure
2100 2100 result_holder = [None]
2101 2101
2102 2102 def setupone():
2103 2103 result_holder[0] = None
2104 2104
2105 2105 generate = _find_stream_generator(stream_version)
2106 2106
2107 2107 def runone():
2108 2108 # the lock is held for the duration the initialisation
2109 2109 result_holder[0] = generate(repo)
2110 2110
2111 2111 timer(runone, setup=setupone, title=b"load")
2112 2112 fm.end()
2113 2113
2114 2114
2115 2115 @command(
2116 2116 b'perf::stream-generate',
2117 2117 [
2118 2118 (
2119 2119 b'',
2120 2120 b'stream-version',
2121 2121 b'latest',
2122 2122 b'stream version to us ("v1", "v2" or "latest", (the default))',
2123 2123 ),
2124 2124 ]
2125 2125 + formatteropts,
2126 2126 )
2127 2127 def perf_stream_clone_generate(ui, repo, stream_version, **opts):
2128 2128 """benchmark the full generation of a stream clone"""
2129 2129
2130 2130 opts = _byteskwargs(opts)
2131 2131 timer, fm = gettimer(ui, opts)
2132 2132
2133 2133 # deletion of the generator may trigger some cleanup that we do not want to
2134 2134 # measure
2135 2135
2136 2136 generate = _find_stream_generator(stream_version)
2137 2137
2138 2138 def runone():
2139 2139 # the lock is held for the duration the initialisation
2140 2140 for chunk in generate(repo):
2141 2141 pass
2142 2142
2143 2143 timer(runone, title=b"generate")
2144 2144 fm.end()
2145 2145
2146 2146
2147 2147 @command(
2148 2148 b'perf::stream-consume',
2149 2149 formatteropts,
2150 2150 )
2151 2151 def perf_stream_clone_consume(ui, repo, filename, **opts):
2152 2152 """benchmark the full application of a stream clone
2153 2153
2154 2154 This include the creation of the repository
2155 2155 """
2156 2156 # try except to appease check code
2157 2157 msg = b"mercurial too old, missing necessary module: %s"
2158 2158 try:
2159 2159 from mercurial import bundle2
2160 2160 except ImportError as exc:
2161 2161 msg %= _bytestr(exc)
2162 2162 raise error.Abort(msg)
2163 2163 try:
2164 2164 from mercurial import exchange
2165 2165 except ImportError as exc:
2166 2166 msg %= _bytestr(exc)
2167 2167 raise error.Abort(msg)
2168 2168 try:
2169 2169 from mercurial import hg
2170 2170 except ImportError as exc:
2171 2171 msg %= _bytestr(exc)
2172 2172 raise error.Abort(msg)
2173 2173 try:
2174 2174 from mercurial import localrepo
2175 2175 except ImportError as exc:
2176 2176 msg %= _bytestr(exc)
2177 2177 raise error.Abort(msg)
2178 2178
2179 2179 opts = _byteskwargs(opts)
2180 2180 timer, fm = gettimer(ui, opts)
2181 2181
2182 2182 # deletion of the generator may trigger some cleanup that we do not want to
2183 2183 # measure
2184 2184 if not (os.path.isfile(filename) and os.access(filename, os.R_OK)):
2185 2185 raise error.Abort("not a readable file: %s" % filename)
2186 2186
2187 2187 run_variables = [None, None]
2188 2188
2189 2189 @contextlib.contextmanager
2190 2190 def context():
2191 2191 with open(filename, mode='rb') as bundle:
2192 2192 with tempfile.TemporaryDirectory() as tmp_dir:
2193 2193 tmp_dir = fsencode(tmp_dir)
2194 2194 run_variables[0] = bundle
2195 2195 run_variables[1] = tmp_dir
2196 2196 yield
2197 2197 run_variables[0] = None
2198 2198 run_variables[1] = None
2199 2199
2200 2200 def runone():
2201 2201 bundle = run_variables[0]
2202 2202 tmp_dir = run_variables[1]
2203 2203 # only pass ui when no srcrepo
2204 2204 localrepo.createrepository(
2205 2205 repo.ui, tmp_dir, requirements=repo.requirements
2206 2206 )
2207 2207 target = hg.repository(repo.ui, tmp_dir)
2208 2208 gen = exchange.readbundle(target.ui, bundle, bundle.name)
2209 2209 # stream v1
2210 2210 if util.safehasattr(gen, 'apply'):
2211 2211 gen.apply(target)
2212 2212 else:
2213 2213 with target.transaction(b"perf::stream-consume") as tr:
2214 2214 bundle2.applybundle(
2215 2215 target,
2216 2216 gen,
2217 2217 tr,
2218 2218 source=b'unbundle',
2219 2219 url=filename,
2220 2220 )
2221 2221
2222 2222 timer(runone, context=context, title=b"consume")
2223 2223 fm.end()
2224 2224
2225 2225
2226 2226 @command(b'perf::parents|perfparents', formatteropts)
2227 2227 def perfparents(ui, repo, **opts):
2228 2228 """benchmark the time necessary to fetch one changeset's parents.
2229 2229
2230 2230 The fetch is done using the `node identifier`, traversing all object layers
2231 2231 from the repository object. The first N revisions will be used for this
2232 2232 benchmark. N is controlled by the ``perf.parentscount`` config option
2233 2233 (default: 1000).
2234 2234 """
2235 2235 opts = _byteskwargs(opts)
2236 2236 timer, fm = gettimer(ui, opts)
2237 2237 # control the number of commits perfparents iterates over
2238 2238 # experimental config: perf.parentscount
2239 2239 count = getint(ui, b"perf", b"parentscount", 1000)
2240 2240 if len(repo.changelog) < count:
2241 2241 raise error.Abort(b"repo needs %d commits for this test" % count)
2242 2242 repo = repo.unfiltered()
2243 2243 nl = [repo.changelog.node(i) for i in _xrange(count)]
2244 2244
2245 2245 def d():
2246 2246 for n in nl:
2247 2247 repo.changelog.parents(n)
2248 2248
2249 2249 timer(d)
2250 2250 fm.end()
2251 2251
2252 2252
2253 2253 @command(b'perf::ctxfiles|perfctxfiles', formatteropts)
2254 2254 def perfctxfiles(ui, repo, x, **opts):
2255 2255 opts = _byteskwargs(opts)
2256 2256 x = int(x)
2257 2257 timer, fm = gettimer(ui, opts)
2258 2258
2259 2259 def d():
2260 2260 len(repo[x].files())
2261 2261
2262 2262 timer(d)
2263 2263 fm.end()
2264 2264
2265 2265
2266 2266 @command(b'perf::rawfiles|perfrawfiles', formatteropts)
2267 2267 def perfrawfiles(ui, repo, x, **opts):
2268 2268 opts = _byteskwargs(opts)
2269 2269 x = int(x)
2270 2270 timer, fm = gettimer(ui, opts)
2271 2271 cl = repo.changelog
2272 2272
2273 2273 def d():
2274 2274 len(cl.read(x)[3])
2275 2275
2276 2276 timer(d)
2277 2277 fm.end()
2278 2278
2279 2279
2280 2280 @command(b'perf::lookup|perflookup', formatteropts)
2281 2281 def perflookup(ui, repo, rev, **opts):
2282 2282 opts = _byteskwargs(opts)
2283 2283 timer, fm = gettimer(ui, opts)
2284 2284 timer(lambda: len(repo.lookup(rev)))
2285 2285 fm.end()
2286 2286
2287 2287
2288 2288 @command(
2289 2289 b'perf::linelogedits|perflinelogedits',
2290 2290 [
2291 2291 (b'n', b'edits', 10000, b'number of edits'),
2292 2292 (b'', b'max-hunk-lines', 10, b'max lines in a hunk'),
2293 2293 ],
2294 2294 norepo=True,
2295 2295 )
2296 2296 def perflinelogedits(ui, **opts):
2297 2297 from mercurial import linelog
2298 2298
2299 2299 opts = _byteskwargs(opts)
2300 2300
2301 2301 edits = opts[b'edits']
2302 2302 maxhunklines = opts[b'max_hunk_lines']
2303 2303
2304 2304 maxb1 = 100000
2305 2305 random.seed(0)
2306 2306 randint = random.randint
2307 2307 currentlines = 0
2308 2308 arglist = []
2309 2309 for rev in _xrange(edits):
2310 2310 a1 = randint(0, currentlines)
2311 2311 a2 = randint(a1, min(currentlines, a1 + maxhunklines))
2312 2312 b1 = randint(0, maxb1)
2313 2313 b2 = randint(b1, b1 + maxhunklines)
2314 2314 currentlines += (b2 - b1) - (a2 - a1)
2315 2315 arglist.append((rev, a1, a2, b1, b2))
2316 2316
2317 2317 def d():
2318 2318 ll = linelog.linelog()
2319 2319 for args in arglist:
2320 2320 ll.replacelines(*args)
2321 2321
2322 2322 timer, fm = gettimer(ui, opts)
2323 2323 timer(d)
2324 2324 fm.end()
2325 2325
2326 2326
2327 2327 @command(b'perf::revrange|perfrevrange', formatteropts)
2328 2328 def perfrevrange(ui, repo, *specs, **opts):
2329 2329 opts = _byteskwargs(opts)
2330 2330 timer, fm = gettimer(ui, opts)
2331 2331 revrange = scmutil.revrange
2332 2332 timer(lambda: len(revrange(repo, specs)))
2333 2333 fm.end()
2334 2334
2335 2335
2336 2336 @command(b'perf::nodelookup|perfnodelookup', formatteropts)
2337 2337 def perfnodelookup(ui, repo, rev, **opts):
2338 2338 opts = _byteskwargs(opts)
2339 2339 timer, fm = gettimer(ui, opts)
2340 2340 import mercurial.revlog
2341 2341
2342 2342 mercurial.revlog._prereadsize = 2 ** 24 # disable lazy parser in old hg
2343 2343 n = scmutil.revsingle(repo, rev).node()
2344 2344
2345 2345 try:
2346 2346 cl = revlog(getsvfs(repo), radix=b"00changelog")
2347 2347 except TypeError:
2348 2348 cl = revlog(getsvfs(repo), indexfile=b"00changelog.i")
2349 2349
2350 2350 def d():
2351 2351 cl.rev(n)
2352 2352 clearcaches(cl)
2353 2353
2354 2354 timer(d)
2355 2355 fm.end()
2356 2356
2357 2357
2358 2358 @command(
2359 2359 b'perf::log|perflog',
2360 2360 [(b'', b'rename', False, b'ask log to follow renames')] + formatteropts,
2361 2361 )
2362 2362 def perflog(ui, repo, rev=None, **opts):
2363 2363 opts = _byteskwargs(opts)
2364 2364 if rev is None:
2365 2365 rev = []
2366 2366 timer, fm = gettimer(ui, opts)
2367 2367 ui.pushbuffer()
2368 2368 timer(
2369 2369 lambda: commands.log(
2370 2370 ui, repo, rev=rev, date=b'', user=b'', copies=opts.get(b'rename')
2371 2371 )
2372 2372 )
2373 2373 ui.popbuffer()
2374 2374 fm.end()
2375 2375
2376 2376
2377 2377 @command(b'perf::moonwalk|perfmoonwalk', formatteropts)
2378 2378 def perfmoonwalk(ui, repo, **opts):
2379 2379 """benchmark walking the changelog backwards
2380 2380
2381 2381 This also loads the changelog data for each revision in the changelog.
2382 2382 """
2383 2383 opts = _byteskwargs(opts)
2384 2384 timer, fm = gettimer(ui, opts)
2385 2385
2386 2386 def moonwalk():
2387 2387 for i in repo.changelog.revs(start=(len(repo) - 1), stop=-1):
2388 2388 ctx = repo[i]
2389 2389 ctx.branch() # read changelog data (in addition to the index)
2390 2390
2391 2391 timer(moonwalk)
2392 2392 fm.end()
2393 2393
2394 2394
2395 2395 @command(
2396 2396 b'perf::templating|perftemplating',
2397 2397 [
2398 2398 (b'r', b'rev', [], b'revisions to run the template on'),
2399 2399 ]
2400 2400 + formatteropts,
2401 2401 )
2402 2402 def perftemplating(ui, repo, testedtemplate=None, **opts):
2403 2403 """test the rendering time of a given template"""
2404 2404 if makelogtemplater is None:
2405 2405 raise error.Abort(
2406 2406 b"perftemplating not available with this Mercurial",
2407 2407 hint=b"use 4.3 or later",
2408 2408 )
2409 2409
2410 2410 opts = _byteskwargs(opts)
2411 2411
2412 2412 nullui = ui.copy()
2413 2413 nullui.fout = open(os.devnull, 'wb')
2414 2414 nullui.disablepager()
2415 2415 revs = opts.get(b'rev')
2416 2416 if not revs:
2417 2417 revs = [b'all()']
2418 2418 revs = list(scmutil.revrange(repo, revs))
2419 2419
2420 2420 defaulttemplate = (
2421 2421 b'{date|shortdate} [{rev}:{node|short}]'
2422 2422 b' {author|person}: {desc|firstline}\n'
2423 2423 )
2424 2424 if testedtemplate is None:
2425 2425 testedtemplate = defaulttemplate
2426 2426 displayer = makelogtemplater(nullui, repo, testedtemplate)
2427 2427
2428 2428 def format():
2429 2429 for r in revs:
2430 2430 ctx = repo[r]
2431 2431 displayer.show(ctx)
2432 2432 displayer.flush(ctx)
2433 2433
2434 2434 timer, fm = gettimer(ui, opts)
2435 2435 timer(format)
2436 2436 fm.end()
2437 2437
2438 2438
2439 2439 def _displaystats(ui, opts, entries, data):
2440 2440 # use a second formatter because the data are quite different, not sure
2441 2441 # how it flies with the templater.
2442 2442 fm = ui.formatter(b'perf-stats', opts)
2443 2443 for key, title in entries:
2444 2444 values = data[key]
2445 2445 nbvalues = len(data)
2446 2446 values.sort()
2447 2447 stats = {
2448 2448 'key': key,
2449 2449 'title': title,
2450 2450 'nbitems': len(values),
2451 2451 'min': values[0][0],
2452 2452 '10%': values[(nbvalues * 10) // 100][0],
2453 2453 '25%': values[(nbvalues * 25) // 100][0],
2454 2454 '50%': values[(nbvalues * 50) // 100][0],
2455 2455 '75%': values[(nbvalues * 75) // 100][0],
2456 2456 '80%': values[(nbvalues * 80) // 100][0],
2457 2457 '85%': values[(nbvalues * 85) // 100][0],
2458 2458 '90%': values[(nbvalues * 90) // 100][0],
2459 2459 '95%': values[(nbvalues * 95) // 100][0],
2460 2460 '99%': values[(nbvalues * 99) // 100][0],
2461 2461 'max': values[-1][0],
2462 2462 }
2463 2463 fm.startitem()
2464 2464 fm.data(**stats)
2465 2465 # make node pretty for the human output
2466 2466 fm.plain('### %s (%d items)\n' % (title, len(values)))
2467 2467 lines = [
2468 2468 'min',
2469 2469 '10%',
2470 2470 '25%',
2471 2471 '50%',
2472 2472 '75%',
2473 2473 '80%',
2474 2474 '85%',
2475 2475 '90%',
2476 2476 '95%',
2477 2477 '99%',
2478 2478 'max',
2479 2479 ]
2480 2480 for l in lines:
2481 2481 fm.plain('%s: %s\n' % (l, stats[l]))
2482 2482 fm.end()
2483 2483
2484 2484
2485 2485 @command(
2486 2486 b'perf::helper-mergecopies|perfhelper-mergecopies',
2487 2487 formatteropts
2488 2488 + [
2489 2489 (b'r', b'revs', [], b'restrict search to these revisions'),
2490 2490 (b'', b'timing', False, b'provides extra data (costly)'),
2491 2491 (b'', b'stats', False, b'provides statistic about the measured data'),
2492 2492 ],
2493 2493 )
2494 2494 def perfhelpermergecopies(ui, repo, revs=[], **opts):
2495 2495 """find statistics about potential parameters for `perfmergecopies`
2496 2496
2497 2497 This command find (base, p1, p2) triplet relevant for copytracing
2498 2498 benchmarking in the context of a merge. It reports values for some of the
2499 2499 parameters that impact merge copy tracing time during merge.
2500 2500
2501 2501 If `--timing` is set, rename detection is run and the associated timing
2502 2502 will be reported. The extra details come at the cost of slower command
2503 2503 execution.
2504 2504
2505 2505 Since rename detection is only run once, other factors might easily
2506 2506 affect the precision of the timing. However it should give a good
2507 2507 approximation of which revision triplets are very costly.
2508 2508 """
2509 2509 opts = _byteskwargs(opts)
2510 2510 fm = ui.formatter(b'perf', opts)
2511 2511 dotiming = opts[b'timing']
2512 2512 dostats = opts[b'stats']
2513 2513
2514 2514 output_template = [
2515 2515 ("base", "%(base)12s"),
2516 2516 ("p1", "%(p1.node)12s"),
2517 2517 ("p2", "%(p2.node)12s"),
2518 2518 ("p1.nb-revs", "%(p1.nbrevs)12d"),
2519 2519 ("p1.nb-files", "%(p1.nbmissingfiles)12d"),
2520 2520 ("p1.renames", "%(p1.renamedfiles)12d"),
2521 2521 ("p1.time", "%(p1.time)12.3f"),
2522 2522 ("p2.nb-revs", "%(p2.nbrevs)12d"),
2523 2523 ("p2.nb-files", "%(p2.nbmissingfiles)12d"),
2524 2524 ("p2.renames", "%(p2.renamedfiles)12d"),
2525 2525 ("p2.time", "%(p2.time)12.3f"),
2526 2526 ("renames", "%(nbrenamedfiles)12d"),
2527 2527 ("total.time", "%(time)12.3f"),
2528 2528 ]
2529 2529 if not dotiming:
2530 2530 output_template = [
2531 2531 i
2532 2532 for i in output_template
2533 2533 if not ('time' in i[0] or 'renames' in i[0])
2534 2534 ]
2535 2535 header_names = [h for (h, v) in output_template]
2536 2536 output = ' '.join([v for (h, v) in output_template]) + '\n'
2537 2537 header = ' '.join(['%12s'] * len(header_names)) + '\n'
2538 2538 fm.plain(header % tuple(header_names))
2539 2539
2540 2540 if not revs:
2541 2541 revs = ['all()']
2542 2542 revs = scmutil.revrange(repo, revs)
2543 2543
2544 2544 if dostats:
2545 2545 alldata = {
2546 2546 'nbrevs': [],
2547 2547 'nbmissingfiles': [],
2548 2548 }
2549 2549 if dotiming:
2550 2550 alldata['parentnbrenames'] = []
2551 2551 alldata['totalnbrenames'] = []
2552 2552 alldata['parenttime'] = []
2553 2553 alldata['totaltime'] = []
2554 2554
2555 2555 roi = repo.revs('merge() and %ld', revs)
2556 2556 for r in roi:
2557 2557 ctx = repo[r]
2558 2558 p1 = ctx.p1()
2559 2559 p2 = ctx.p2()
2560 2560 bases = repo.changelog._commonancestorsheads(p1.rev(), p2.rev())
2561 2561 for b in bases:
2562 2562 b = repo[b]
2563 2563 p1missing = copies._computeforwardmissing(b, p1)
2564 2564 p2missing = copies._computeforwardmissing(b, p2)
2565 2565 data = {
2566 2566 b'base': b.hex(),
2567 2567 b'p1.node': p1.hex(),
2568 2568 b'p1.nbrevs': len(repo.revs('only(%d, %d)', p1.rev(), b.rev())),
2569 2569 b'p1.nbmissingfiles': len(p1missing),
2570 2570 b'p2.node': p2.hex(),
2571 2571 b'p2.nbrevs': len(repo.revs('only(%d, %d)', p2.rev(), b.rev())),
2572 2572 b'p2.nbmissingfiles': len(p2missing),
2573 2573 }
2574 2574 if dostats:
2575 2575 if p1missing:
2576 2576 alldata['nbrevs'].append(
2577 2577 (data['p1.nbrevs'], b.hex(), p1.hex())
2578 2578 )
2579 2579 alldata['nbmissingfiles'].append(
2580 2580 (data['p1.nbmissingfiles'], b.hex(), p1.hex())
2581 2581 )
2582 2582 if p2missing:
2583 2583 alldata['nbrevs'].append(
2584 2584 (data['p2.nbrevs'], b.hex(), p2.hex())
2585 2585 )
2586 2586 alldata['nbmissingfiles'].append(
2587 2587 (data['p2.nbmissingfiles'], b.hex(), p2.hex())
2588 2588 )
2589 2589 if dotiming:
2590 2590 begin = util.timer()
2591 2591 mergedata = copies.mergecopies(repo, p1, p2, b)
2592 2592 end = util.timer()
2593 2593 # not very stable timing since we did only one run
2594 2594 data['time'] = end - begin
2595 2595 # mergedata contains five dicts: "copy", "movewithdir",
2596 2596 # "diverge", "renamedelete" and "dirmove".
2597 2597 # The first 4 are about renamed file so lets count that.
2598 2598 renames = len(mergedata[0])
2599 2599 renames += len(mergedata[1])
2600 2600 renames += len(mergedata[2])
2601 2601 renames += len(mergedata[3])
2602 2602 data['nbrenamedfiles'] = renames
2603 2603 begin = util.timer()
2604 2604 p1renames = copies.pathcopies(b, p1)
2605 2605 end = util.timer()
2606 2606 data['p1.time'] = end - begin
2607 2607 begin = util.timer()
2608 2608 p2renames = copies.pathcopies(b, p2)
2609 2609 end = util.timer()
2610 2610 data['p2.time'] = end - begin
2611 2611 data['p1.renamedfiles'] = len(p1renames)
2612 2612 data['p2.renamedfiles'] = len(p2renames)
2613 2613
2614 2614 if dostats:
2615 2615 if p1missing:
2616 2616 alldata['parentnbrenames'].append(
2617 2617 (data['p1.renamedfiles'], b.hex(), p1.hex())
2618 2618 )
2619 2619 alldata['parenttime'].append(
2620 2620 (data['p1.time'], b.hex(), p1.hex())
2621 2621 )
2622 2622 if p2missing:
2623 2623 alldata['parentnbrenames'].append(
2624 2624 (data['p2.renamedfiles'], b.hex(), p2.hex())
2625 2625 )
2626 2626 alldata['parenttime'].append(
2627 2627 (data['p2.time'], b.hex(), p2.hex())
2628 2628 )
2629 2629 if p1missing or p2missing:
2630 2630 alldata['totalnbrenames'].append(
2631 2631 (
2632 2632 data['nbrenamedfiles'],
2633 2633 b.hex(),
2634 2634 p1.hex(),
2635 2635 p2.hex(),
2636 2636 )
2637 2637 )
2638 2638 alldata['totaltime'].append(
2639 2639 (data['time'], b.hex(), p1.hex(), p2.hex())
2640 2640 )
2641 2641 fm.startitem()
2642 2642 fm.data(**data)
2643 2643 # make node pretty for the human output
2644 2644 out = data.copy()
2645 2645 out['base'] = fm.hexfunc(b.node())
2646 2646 out['p1.node'] = fm.hexfunc(p1.node())
2647 2647 out['p2.node'] = fm.hexfunc(p2.node())
2648 2648 fm.plain(output % out)
2649 2649
2650 2650 fm.end()
2651 2651 if dostats:
2652 2652 # use a second formatter because the data are quite different, not sure
2653 2653 # how it flies with the templater.
2654 2654 entries = [
2655 2655 ('nbrevs', 'number of revision covered'),
2656 2656 ('nbmissingfiles', 'number of missing files at head'),
2657 2657 ]
2658 2658 if dotiming:
2659 2659 entries.append(
2660 2660 ('parentnbrenames', 'rename from one parent to base')
2661 2661 )
2662 2662 entries.append(('totalnbrenames', 'total number of renames'))
2663 2663 entries.append(('parenttime', 'time for one parent'))
2664 2664 entries.append(('totaltime', 'time for both parents'))
2665 2665 _displaystats(ui, opts, entries, alldata)
2666 2666
2667 2667
2668 2668 @command(
2669 2669 b'perf::helper-pathcopies|perfhelper-pathcopies',
2670 2670 formatteropts
2671 2671 + [
2672 2672 (b'r', b'revs', [], b'restrict search to these revisions'),
2673 2673 (b'', b'timing', False, b'provides extra data (costly)'),
2674 2674 (b'', b'stats', False, b'provides statistic about the measured data'),
2675 2675 ],
2676 2676 )
2677 2677 def perfhelperpathcopies(ui, repo, revs=[], **opts):
2678 2678 """find statistic about potential parameters for the `perftracecopies`
2679 2679
2680 2680 This command find source-destination pair relevant for copytracing testing.
2681 2681 It report value for some of the parameters that impact copy tracing time.
2682 2682
2683 2683 If `--timing` is set, rename detection is run and the associated timing
2684 2684 will be reported. The extra details comes at the cost of a slower command
2685 2685 execution.
2686 2686
2687 2687 Since the rename detection is only run once, other factors might easily
2688 2688 affect the precision of the timing. However it should give a good
2689 2689 approximation of which revision pairs are very costly.
2690 2690 """
2691 2691 opts = _byteskwargs(opts)
2692 2692 fm = ui.formatter(b'perf', opts)
2693 2693 dotiming = opts[b'timing']
2694 2694 dostats = opts[b'stats']
2695 2695
2696 2696 if dotiming:
2697 2697 header = '%12s %12s %12s %12s %12s %12s\n'
2698 2698 output = (
2699 2699 "%(source)12s %(destination)12s "
2700 2700 "%(nbrevs)12d %(nbmissingfiles)12d "
2701 2701 "%(nbrenamedfiles)12d %(time)18.5f\n"
2702 2702 )
2703 2703 header_names = (
2704 2704 "source",
2705 2705 "destination",
2706 2706 "nb-revs",
2707 2707 "nb-files",
2708 2708 "nb-renames",
2709 2709 "time",
2710 2710 )
2711 2711 fm.plain(header % header_names)
2712 2712 else:
2713 2713 header = '%12s %12s %12s %12s\n'
2714 2714 output = (
2715 2715 "%(source)12s %(destination)12s "
2716 2716 "%(nbrevs)12d %(nbmissingfiles)12d\n"
2717 2717 )
2718 2718 fm.plain(header % ("source", "destination", "nb-revs", "nb-files"))
2719 2719
2720 2720 if not revs:
2721 2721 revs = ['all()']
2722 2722 revs = scmutil.revrange(repo, revs)
2723 2723
2724 2724 if dostats:
2725 2725 alldata = {
2726 2726 'nbrevs': [],
2727 2727 'nbmissingfiles': [],
2728 2728 }
2729 2729 if dotiming:
2730 2730 alldata['nbrenames'] = []
2731 2731 alldata['time'] = []
2732 2732
2733 2733 roi = repo.revs('merge() and %ld', revs)
2734 2734 for r in roi:
2735 2735 ctx = repo[r]
2736 2736 p1 = ctx.p1().rev()
2737 2737 p2 = ctx.p2().rev()
2738 2738 bases = repo.changelog._commonancestorsheads(p1, p2)
2739 2739 for p in (p1, p2):
2740 2740 for b in bases:
2741 2741 base = repo[b]
2742 2742 parent = repo[p]
2743 2743 missing = copies._computeforwardmissing(base, parent)
2744 2744 if not missing:
2745 2745 continue
2746 2746 data = {
2747 2747 b'source': base.hex(),
2748 2748 b'destination': parent.hex(),
2749 2749 b'nbrevs': len(repo.revs('only(%d, %d)', p, b)),
2750 2750 b'nbmissingfiles': len(missing),
2751 2751 }
2752 2752 if dostats:
2753 2753 alldata['nbrevs'].append(
2754 2754 (
2755 2755 data['nbrevs'],
2756 2756 base.hex(),
2757 2757 parent.hex(),
2758 2758 )
2759 2759 )
2760 2760 alldata['nbmissingfiles'].append(
2761 2761 (
2762 2762 data['nbmissingfiles'],
2763 2763 base.hex(),
2764 2764 parent.hex(),
2765 2765 )
2766 2766 )
2767 2767 if dotiming:
2768 2768 begin = util.timer()
2769 2769 renames = copies.pathcopies(base, parent)
2770 2770 end = util.timer()
2771 2771 # not very stable timing since we did only one run
2772 2772 data['time'] = end - begin
2773 2773 data['nbrenamedfiles'] = len(renames)
2774 2774 if dostats:
2775 2775 alldata['time'].append(
2776 2776 (
2777 2777 data['time'],
2778 2778 base.hex(),
2779 2779 parent.hex(),
2780 2780 )
2781 2781 )
2782 2782 alldata['nbrenames'].append(
2783 2783 (
2784 2784 data['nbrenamedfiles'],
2785 2785 base.hex(),
2786 2786 parent.hex(),
2787 2787 )
2788 2788 )
2789 2789 fm.startitem()
2790 2790 fm.data(**data)
2791 2791 out = data.copy()
2792 2792 out['source'] = fm.hexfunc(base.node())
2793 2793 out['destination'] = fm.hexfunc(parent.node())
2794 2794 fm.plain(output % out)
2795 2795
2796 2796 fm.end()
2797 2797 if dostats:
2798 2798 entries = [
2799 2799 ('nbrevs', 'number of revision covered'),
2800 2800 ('nbmissingfiles', 'number of missing files at head'),
2801 2801 ]
2802 2802 if dotiming:
2803 2803 entries.append(('nbrenames', 'renamed files'))
2804 2804 entries.append(('time', 'time'))
2805 2805 _displaystats(ui, opts, entries, alldata)
2806 2806
2807 2807
2808 2808 @command(b'perf::cca|perfcca', formatteropts)
2809 2809 def perfcca(ui, repo, **opts):
2810 2810 opts = _byteskwargs(opts)
2811 2811 timer, fm = gettimer(ui, opts)
2812 2812 timer(lambda: scmutil.casecollisionauditor(ui, False, repo.dirstate))
2813 2813 fm.end()
2814 2814
2815 2815
2816 2816 @command(b'perf::fncacheload|perffncacheload', formatteropts)
2817 2817 def perffncacheload(ui, repo, **opts):
2818 2818 opts = _byteskwargs(opts)
2819 2819 timer, fm = gettimer(ui, opts)
2820 2820 s = repo.store
2821 2821
2822 2822 def d():
2823 2823 s.fncache._load()
2824 2824
2825 2825 timer(d)
2826 2826 fm.end()
2827 2827
2828 2828
2829 2829 @command(b'perf::fncachewrite|perffncachewrite', formatteropts)
2830 2830 def perffncachewrite(ui, repo, **opts):
2831 2831 opts = _byteskwargs(opts)
2832 2832 timer, fm = gettimer(ui, opts)
2833 2833 s = repo.store
2834 2834 lock = repo.lock()
2835 2835 s.fncache._load()
2836 2836 tr = repo.transaction(b'perffncachewrite')
2837 2837 tr.addbackup(b'fncache')
2838 2838
2839 2839 def d():
2840 2840 s.fncache._dirty = True
2841 2841 s.fncache.write(tr)
2842 2842
2843 2843 timer(d)
2844 2844 tr.close()
2845 2845 lock.release()
2846 2846 fm.end()
2847 2847
2848 2848
2849 2849 @command(b'perf::fncacheencode|perffncacheencode', formatteropts)
2850 2850 def perffncacheencode(ui, repo, **opts):
2851 2851 opts = _byteskwargs(opts)
2852 2852 timer, fm = gettimer(ui, opts)
2853 2853 s = repo.store
2854 2854 s.fncache._load()
2855 2855
2856 2856 def d():
2857 2857 for p in s.fncache.entries:
2858 2858 s.encode(p)
2859 2859
2860 2860 timer(d)
2861 2861 fm.end()
2862 2862
2863 2863
2864 2864 def _bdiffworker(q, blocks, xdiff, ready, done):
2865 2865 while not done.is_set():
2866 2866 pair = q.get()
2867 2867 while pair is not None:
2868 2868 if xdiff:
2869 2869 mdiff.bdiff.xdiffblocks(*pair)
2870 2870 elif blocks:
2871 2871 mdiff.bdiff.blocks(*pair)
2872 2872 else:
2873 2873 mdiff.textdiff(*pair)
2874 2874 q.task_done()
2875 2875 pair = q.get()
2876 2876 q.task_done() # for the None one
2877 2877 with ready:
2878 2878 ready.wait()
2879 2879
2880 2880
2881 2881 def _manifestrevision(repo, mnode):
2882 2882 ml = repo.manifestlog
2883 2883
2884 2884 if util.safehasattr(ml, b'getstorage'):
2885 2885 store = ml.getstorage(b'')
2886 2886 else:
2887 2887 store = ml._revlog
2888 2888
2889 2889 return store.revision(mnode)
2890 2890
2891 2891
2892 2892 @command(
2893 2893 b'perf::bdiff|perfbdiff',
2894 2894 revlogopts
2895 2895 + formatteropts
2896 2896 + [
2897 2897 (
2898 2898 b'',
2899 2899 b'count',
2900 2900 1,
2901 2901 b'number of revisions to test (when using --startrev)',
2902 2902 ),
2903 2903 (b'', b'alldata', False, b'test bdiffs for all associated revisions'),
2904 2904 (b'', b'threads', 0, b'number of thread to use (disable with 0)'),
2905 2905 (b'', b'blocks', False, b'test computing diffs into blocks'),
2906 2906 (b'', b'xdiff', False, b'use xdiff algorithm'),
2907 2907 ],
2908 2908 b'-c|-m|FILE REV',
2909 2909 )
2910 2910 def perfbdiff(ui, repo, file_, rev=None, count=None, threads=0, **opts):
2911 2911 """benchmark a bdiff between revisions
2912 2912
2913 2913 By default, benchmark a bdiff between its delta parent and itself.
2914 2914
2915 2915 With ``--count``, benchmark bdiffs between delta parents and self for N
2916 2916 revisions starting at the specified revision.
2917 2917
2918 2918 With ``--alldata``, assume the requested revision is a changeset and
2919 2919 measure bdiffs for all changes related to that changeset (manifest
2920 2920 and filelogs).
2921 2921 """
2922 2922 opts = _byteskwargs(opts)
2923 2923
2924 2924 if opts[b'xdiff'] and not opts[b'blocks']:
2925 2925 raise error.CommandError(b'perfbdiff', b'--xdiff requires --blocks')
2926 2926
2927 2927 if opts[b'alldata']:
2928 2928 opts[b'changelog'] = True
2929 2929
2930 2930 if opts.get(b'changelog') or opts.get(b'manifest'):
2931 2931 file_, rev = None, file_
2932 2932 elif rev is None:
2933 2933 raise error.CommandError(b'perfbdiff', b'invalid arguments')
2934 2934
2935 2935 blocks = opts[b'blocks']
2936 2936 xdiff = opts[b'xdiff']
2937 2937 textpairs = []
2938 2938
2939 2939 r = cmdutil.openrevlog(repo, b'perfbdiff', file_, opts)
2940 2940
2941 2941 startrev = r.rev(r.lookup(rev))
2942 2942 for rev in range(startrev, min(startrev + count, len(r) - 1)):
2943 2943 if opts[b'alldata']:
2944 2944 # Load revisions associated with changeset.
2945 2945 ctx = repo[rev]
2946 2946 mtext = _manifestrevision(repo, ctx.manifestnode())
2947 2947 for pctx in ctx.parents():
2948 2948 pman = _manifestrevision(repo, pctx.manifestnode())
2949 2949 textpairs.append((pman, mtext))
2950 2950
2951 2951 # Load filelog revisions by iterating manifest delta.
2952 2952 man = ctx.manifest()
2953 2953 pman = ctx.p1().manifest()
2954 2954 for filename, change in pman.diff(man).items():
2955 2955 fctx = repo.file(filename)
2956 2956 f1 = fctx.revision(change[0][0] or -1)
2957 2957 f2 = fctx.revision(change[1][0] or -1)
2958 2958 textpairs.append((f1, f2))
2959 2959 else:
2960 2960 dp = r.deltaparent(rev)
2961 2961 textpairs.append((r.revision(dp), r.revision(rev)))
2962 2962
2963 2963 withthreads = threads > 0
2964 2964 if not withthreads:
2965 2965
2966 2966 def d():
2967 2967 for pair in textpairs:
2968 2968 if xdiff:
2969 2969 mdiff.bdiff.xdiffblocks(*pair)
2970 2970 elif blocks:
2971 2971 mdiff.bdiff.blocks(*pair)
2972 2972 else:
2973 2973 mdiff.textdiff(*pair)
2974 2974
2975 2975 else:
2976 2976 q = queue()
2977 2977 for i in _xrange(threads):
2978 2978 q.put(None)
2979 2979 ready = threading.Condition()
2980 2980 done = threading.Event()
2981 2981 for i in _xrange(threads):
2982 2982 threading.Thread(
2983 2983 target=_bdiffworker, args=(q, blocks, xdiff, ready, done)
2984 2984 ).start()
2985 2985 q.join()
2986 2986
2987 2987 def d():
2988 2988 for pair in textpairs:
2989 2989 q.put(pair)
2990 2990 for i in _xrange(threads):
2991 2991 q.put(None)
2992 2992 with ready:
2993 2993 ready.notify_all()
2994 2994 q.join()
2995 2995
2996 2996 timer, fm = gettimer(ui, opts)
2997 2997 timer(d)
2998 2998 fm.end()
2999 2999
3000 3000 if withthreads:
3001 3001 done.set()
3002 3002 for i in _xrange(threads):
3003 3003 q.put(None)
3004 3004 with ready:
3005 3005 ready.notify_all()
3006 3006
3007 3007
3008 3008 @command(
3009 3009 b'perf::unbundle',
3010 3010 formatteropts,
3011 3011 b'BUNDLE_FILE',
3012 3012 )
3013 3013 def perf_unbundle(ui, repo, fname, **opts):
3014 3014 """benchmark application of a bundle in a repository.
3015 3015
3016 3016 This does not include the final transaction processing"""
3017 3017
3018 3018 from mercurial import exchange
3019 3019 from mercurial import bundle2
3020 3020 from mercurial import transaction
3021 3021
3022 3022 opts = _byteskwargs(opts)
3023 3023
3024 3024 ### some compatibility hotfix
3025 3025 #
3026 3026 # the data attribute is dropped in 63edc384d3b7 a changeset introducing a
3027 3027 # critical regression that break transaction rollback for files that are
3028 3028 # de-inlined.
3029 3029 method = transaction.transaction._addentry
3030 3030 pre_63edc384d3b7 = "data" in getargspec(method).args
3031 3031 # the `detailed_exit_code` attribute is introduced in 33c0c25d0b0f
3032 3032 # a changeset that is a close descendant of 18415fc918a1, the changeset
3033 3033 # that conclude the fix run for the bug introduced in 63edc384d3b7.
3034 3034 args = getargspec(error.Abort.__init__).args
3035 3035 post_18415fc918a1 = "detailed_exit_code" in args
3036 3036
3037 3037 old_max_inline = None
3038 3038 try:
3039 3039 if not (pre_63edc384d3b7 or post_18415fc918a1):
3040 3040 # disable inlining
3041 3041 old_max_inline = mercurial.revlog._maxinline
3042 3042 # large enough to never happen
3043 3043 mercurial.revlog._maxinline = 2 ** 50
3044 3044
3045 3045 with repo.lock():
3046 3046 bundle = [None, None]
3047 3047 orig_quiet = repo.ui.quiet
3048 3048 try:
3049 3049 repo.ui.quiet = True
3050 3050 with open(fname, mode="rb") as f:
3051 3051
3052 3052 def noop_report(*args, **kwargs):
3053 3053 pass
3054 3054
3055 3055 def setup():
3056 3056 gen, tr = bundle
3057 3057 if tr is not None:
3058 3058 tr.abort()
3059 3059 bundle[:] = [None, None]
3060 3060 f.seek(0)
3061 3061 bundle[0] = exchange.readbundle(ui, f, fname)
3062 3062 bundle[1] = repo.transaction(b'perf::unbundle')
3063 3063 # silence the transaction
3064 3064 bundle[1]._report = noop_report
3065 3065
3066 3066 def apply():
3067 3067 gen, tr = bundle
3068 3068 bundle2.applybundle(
3069 3069 repo,
3070 3070 gen,
3071 3071 tr,
3072 3072 source=b'perf::unbundle',
3073 3073 url=fname,
3074 3074 )
3075 3075
3076 3076 timer, fm = gettimer(ui, opts)
3077 3077 timer(apply, setup=setup)
3078 3078 fm.end()
3079 3079 finally:
3080 3080 repo.ui.quiet == orig_quiet
3081 3081 gen, tr = bundle
3082 3082 if tr is not None:
3083 3083 tr.abort()
3084 3084 finally:
3085 3085 if old_max_inline is not None:
3086 3086 mercurial.revlog._maxinline = old_max_inline
3087 3087
3088 3088
3089 3089 @command(
3090 3090 b'perf::unidiff|perfunidiff',
3091 3091 revlogopts
3092 3092 + formatteropts
3093 3093 + [
3094 3094 (
3095 3095 b'',
3096 3096 b'count',
3097 3097 1,
3098 3098 b'number of revisions to test (when using --startrev)',
3099 3099 ),
3100 3100 (b'', b'alldata', False, b'test unidiffs for all associated revisions'),
3101 3101 ],
3102 3102 b'-c|-m|FILE REV',
3103 3103 )
3104 3104 def perfunidiff(ui, repo, file_, rev=None, count=None, **opts):
3105 3105 """benchmark a unified diff between revisions
3106 3106
3107 3107 This doesn't include any copy tracing - it's just a unified diff
3108 3108 of the texts.
3109 3109
3110 3110 By default, benchmark a diff between its delta parent and itself.
3111 3111
3112 3112 With ``--count``, benchmark diffs between delta parents and self for N
3113 3113 revisions starting at the specified revision.
3114 3114
3115 3115 With ``--alldata``, assume the requested revision is a changeset and
3116 3116 measure diffs for all changes related to that changeset (manifest
3117 3117 and filelogs).
3118 3118 """
3119 3119 opts = _byteskwargs(opts)
3120 3120 if opts[b'alldata']:
3121 3121 opts[b'changelog'] = True
3122 3122
3123 3123 if opts.get(b'changelog') or opts.get(b'manifest'):
3124 3124 file_, rev = None, file_
3125 3125 elif rev is None:
3126 3126 raise error.CommandError(b'perfunidiff', b'invalid arguments')
3127 3127
3128 3128 textpairs = []
3129 3129
3130 3130 r = cmdutil.openrevlog(repo, b'perfunidiff', file_, opts)
3131 3131
3132 3132 startrev = r.rev(r.lookup(rev))
3133 3133 for rev in range(startrev, min(startrev + count, len(r) - 1)):
3134 3134 if opts[b'alldata']:
3135 3135 # Load revisions associated with changeset.
3136 3136 ctx = repo[rev]
3137 3137 mtext = _manifestrevision(repo, ctx.manifestnode())
3138 3138 for pctx in ctx.parents():
3139 3139 pman = _manifestrevision(repo, pctx.manifestnode())
3140 3140 textpairs.append((pman, mtext))
3141 3141
3142 3142 # Load filelog revisions by iterating manifest delta.
3143 3143 man = ctx.manifest()
3144 3144 pman = ctx.p1().manifest()
3145 3145 for filename, change in pman.diff(man).items():
3146 3146 fctx = repo.file(filename)
3147 3147 f1 = fctx.revision(change[0][0] or -1)
3148 3148 f2 = fctx.revision(change[1][0] or -1)
3149 3149 textpairs.append((f1, f2))
3150 3150 else:
3151 3151 dp = r.deltaparent(rev)
3152 3152 textpairs.append((r.revision(dp), r.revision(rev)))
3153 3153
3154 3154 def d():
3155 3155 for left, right in textpairs:
3156 3156 # The date strings don't matter, so we pass empty strings.
3157 3157 headerlines, hunks = mdiff.unidiff(
3158 3158 left, b'', right, b'', b'left', b'right', binary=False
3159 3159 )
3160 3160 # consume iterators in roughly the way patch.py does
3161 3161 b'\n'.join(headerlines)
3162 3162 b''.join(sum((list(hlines) for hrange, hlines in hunks), []))
3163 3163
3164 3164 timer, fm = gettimer(ui, opts)
3165 3165 timer(d)
3166 3166 fm.end()
3167 3167
3168 3168
3169 3169 @command(b'perf::diffwd|perfdiffwd', formatteropts)
3170 3170 def perfdiffwd(ui, repo, **opts):
3171 3171 """Profile diff of working directory changes"""
3172 3172 opts = _byteskwargs(opts)
3173 3173 timer, fm = gettimer(ui, opts)
3174 3174 options = {
3175 3175 'w': 'ignore_all_space',
3176 3176 'b': 'ignore_space_change',
3177 3177 'B': 'ignore_blank_lines',
3178 3178 }
3179 3179
3180 3180 for diffopt in ('', 'w', 'b', 'B', 'wB'):
3181 3181 opts = {options[c]: b'1' for c in diffopt}
3182 3182
3183 3183 def d():
3184 3184 ui.pushbuffer()
3185 3185 commands.diff(ui, repo, **opts)
3186 3186 ui.popbuffer()
3187 3187
3188 3188 diffopt = diffopt.encode('ascii')
3189 3189 title = b'diffopts: %s' % (diffopt and (b'-' + diffopt) or b'none')
3190 3190 timer(d, title=title)
3191 3191 fm.end()
3192 3192
3193 3193
3194 3194 @command(
3195 3195 b'perf::revlogindex|perfrevlogindex',
3196 3196 revlogopts + formatteropts,
3197 3197 b'-c|-m|FILE',
3198 3198 )
3199 3199 def perfrevlogindex(ui, repo, file_=None, **opts):
3200 3200 """Benchmark operations against a revlog index.
3201 3201
3202 3202 This tests constructing a revlog instance, reading index data,
3203 3203 parsing index data, and performing various operations related to
3204 3204 index data.
3205 3205 """
3206 3206
3207 3207 opts = _byteskwargs(opts)
3208 3208
3209 3209 rl = cmdutil.openrevlog(repo, b'perfrevlogindex', file_, opts)
3210 3210
3211 3211 opener = getattr(rl, 'opener') # trick linter
3212 3212 # compat with hg <= 5.8
3213 3213 radix = getattr(rl, 'radix', None)
3214 3214 indexfile = getattr(rl, '_indexfile', None)
3215 3215 if indexfile is None:
3216 3216 # compatibility with <= hg-5.8
3217 3217 indexfile = getattr(rl, 'indexfile')
3218 3218 data = opener.read(indexfile)
3219 3219
3220 3220 header = struct.unpack(b'>I', data[0:4])[0]
3221 3221 version = header & 0xFFFF
3222 3222 if version == 1:
3223 3223 inline = header & (1 << 16)
3224 3224 else:
3225 3225 raise error.Abort(b'unsupported revlog version: %d' % version)
3226 3226
3227 3227 parse_index_v1 = getattr(mercurial.revlog, 'parse_index_v1', None)
3228 3228 if parse_index_v1 is None:
3229 3229 parse_index_v1 = mercurial.revlog.revlogio().parseindex
3230 3230
3231 3231 rllen = len(rl)
3232 3232
3233 3233 node0 = rl.node(0)
3234 3234 node25 = rl.node(rllen // 4)
3235 3235 node50 = rl.node(rllen // 2)
3236 3236 node75 = rl.node(rllen // 4 * 3)
3237 3237 node100 = rl.node(rllen - 1)
3238 3238
3239 3239 allrevs = range(rllen)
3240 3240 allrevsrev = list(reversed(allrevs))
3241 3241 allnodes = [rl.node(rev) for rev in range(rllen)]
3242 3242 allnodesrev = list(reversed(allnodes))
3243 3243
3244 3244 def constructor():
3245 3245 if radix is not None:
3246 3246 revlog(opener, radix=radix)
3247 3247 else:
3248 3248 # hg <= 5.8
3249 3249 revlog(opener, indexfile=indexfile)
3250 3250
3251 3251 def read():
3252 3252 with opener(indexfile) as fh:
3253 3253 fh.read()
3254 3254
3255 3255 def parseindex():
3256 3256 parse_index_v1(data, inline)
3257 3257
3258 3258 def getentry(revornode):
3259 3259 index = parse_index_v1(data, inline)[0]
3260 3260 index[revornode]
3261 3261
3262 3262 def getentries(revs, count=1):
3263 3263 index = parse_index_v1(data, inline)[0]
3264 3264
3265 3265 for i in range(count):
3266 3266 for rev in revs:
3267 3267 index[rev]
3268 3268
3269 3269 def resolvenode(node):
3270 3270 index = parse_index_v1(data, inline)[0]
3271 3271 rev = getattr(index, 'rev', None)
3272 3272 if rev is None:
3273 3273 nodemap = getattr(parse_index_v1(data, inline)[0], 'nodemap', None)
3274 3274 # This only works for the C code.
3275 3275 if nodemap is None:
3276 3276 return
3277 3277 rev = nodemap.__getitem__
3278 3278
3279 3279 try:
3280 3280 rev(node)
3281 3281 except error.RevlogError:
3282 3282 pass
3283 3283
3284 3284 def resolvenodes(nodes, count=1):
3285 3285 index = parse_index_v1(data, inline)[0]
3286 3286 rev = getattr(index, 'rev', None)
3287 3287 if rev is None:
3288 3288 nodemap = getattr(parse_index_v1(data, inline)[0], 'nodemap', None)
3289 3289 # This only works for the C code.
3290 3290 if nodemap is None:
3291 3291 return
3292 3292 rev = nodemap.__getitem__
3293 3293
3294 3294 for i in range(count):
3295 3295 for node in nodes:
3296 3296 try:
3297 3297 rev(node)
3298 3298 except error.RevlogError:
3299 3299 pass
3300 3300
3301 3301 benches = [
3302 3302 (constructor, b'revlog constructor'),
3303 3303 (read, b'read'),
3304 3304 (parseindex, b'create index object'),
3305 3305 (lambda: getentry(0), b'retrieve index entry for rev 0'),
3306 3306 (lambda: resolvenode(b'a' * 20), b'look up missing node'),
3307 3307 (lambda: resolvenode(node0), b'look up node at rev 0'),
3308 3308 (lambda: resolvenode(node25), b'look up node at 1/4 len'),
3309 3309 (lambda: resolvenode(node50), b'look up node at 1/2 len'),
3310 3310 (lambda: resolvenode(node75), b'look up node at 3/4 len'),
3311 3311 (lambda: resolvenode(node100), b'look up node at tip'),
3312 3312 # 2x variation is to measure caching impact.
3313 3313 (lambda: resolvenodes(allnodes), b'look up all nodes (forward)'),
3314 3314 (lambda: resolvenodes(allnodes, 2), b'look up all nodes 2x (forward)'),
3315 3315 (lambda: resolvenodes(allnodesrev), b'look up all nodes (reverse)'),
3316 3316 (
3317 3317 lambda: resolvenodes(allnodesrev, 2),
3318 3318 b'look up all nodes 2x (reverse)',
3319 3319 ),
3320 3320 (lambda: getentries(allrevs), b'retrieve all index entries (forward)'),
3321 3321 (
3322 3322 lambda: getentries(allrevs, 2),
3323 3323 b'retrieve all index entries 2x (forward)',
3324 3324 ),
3325 3325 (
3326 3326 lambda: getentries(allrevsrev),
3327 3327 b'retrieve all index entries (reverse)',
3328 3328 ),
3329 3329 (
3330 3330 lambda: getentries(allrevsrev, 2),
3331 3331 b'retrieve all index entries 2x (reverse)',
3332 3332 ),
3333 3333 ]
3334 3334
3335 3335 for fn, title in benches:
3336 3336 timer, fm = gettimer(ui, opts)
3337 3337 timer(fn, title=title)
3338 3338 fm.end()
3339 3339
3340 3340
3341 3341 @command(
3342 3342 b'perf::revlogrevisions|perfrevlogrevisions',
3343 3343 revlogopts
3344 3344 + formatteropts
3345 3345 + [
3346 3346 (b'd', b'dist', 100, b'distance between the revisions'),
3347 3347 (b's', b'startrev', 0, b'revision to start reading at'),
3348 3348 (b'', b'reverse', False, b'read in reverse'),
3349 3349 ],
3350 3350 b'-c|-m|FILE',
3351 3351 )
3352 3352 def perfrevlogrevisions(
3353 3353 ui, repo, file_=None, startrev=0, reverse=False, **opts
3354 3354 ):
3355 3355 """Benchmark reading a series of revisions from a revlog.
3356 3356
3357 3357 By default, we read every ``-d/--dist`` revision from 0 to tip of
3358 3358 the specified revlog.
3359 3359
3360 3360 The start revision can be defined via ``-s/--startrev``.
3361 3361 """
3362 3362 opts = _byteskwargs(opts)
3363 3363
3364 3364 rl = cmdutil.openrevlog(repo, b'perfrevlogrevisions', file_, opts)
3365 3365 rllen = getlen(ui)(rl)
3366 3366
3367 3367 if startrev < 0:
3368 3368 startrev = rllen + startrev
3369 3369
3370 3370 def d():
3371 3371 rl.clearcaches()
3372 3372
3373 3373 beginrev = startrev
3374 3374 endrev = rllen
3375 3375 dist = opts[b'dist']
3376 3376
3377 3377 if reverse:
3378 3378 beginrev, endrev = endrev - 1, beginrev - 1
3379 3379 dist = -1 * dist
3380 3380
3381 3381 for x in _xrange(beginrev, endrev, dist):
3382 3382 # Old revisions don't support passing int.
3383 3383 n = rl.node(x)
3384 3384 rl.revision(n)
3385 3385
3386 3386 timer, fm = gettimer(ui, opts)
3387 3387 timer(d)
3388 3388 fm.end()
3389 3389
3390 3390
3391 3391 @command(
3392 3392 b'perf::revlogwrite|perfrevlogwrite',
3393 3393 revlogopts
3394 3394 + formatteropts
3395 3395 + [
3396 3396 (b's', b'startrev', 1000, b'revision to start writing at'),
3397 3397 (b'', b'stoprev', -1, b'last revision to write'),
3398 3398 (b'', b'count', 3, b'number of passes to perform'),
3399 3399 (b'', b'details', False, b'print timing for every revisions tested'),
3400 3400 (b'', b'source', b'full', b'the kind of data feed in the revlog'),
3401 3401 (b'', b'lazydeltabase', True, b'try the provided delta first'),
3402 3402 (b'', b'clear-caches', True, b'clear revlog cache between calls'),
3403 3403 ],
3404 3404 b'-c|-m|FILE',
3405 3405 )
3406 3406 def perfrevlogwrite(ui, repo, file_=None, startrev=1000, stoprev=-1, **opts):
3407 3407 """Benchmark writing a series of revisions to a revlog.
3408 3408
3409 3409 Possible source values are:
3410 3410 * `full`: add from a full text (default).
3411 3411 * `parent-1`: add from a delta to the first parent
3412 3412 * `parent-2`: add from a delta to the second parent if it exists
3413 3413 (use a delta from the first parent otherwise)
3414 3414 * `parent-smallest`: add from the smallest delta (either p1 or p2)
3415 3415 * `storage`: add from the existing precomputed deltas
3416 3416
3417 3417 Note: This performance command measures performance in a custom way. As a
3418 3418 result some of the global configuration of the 'perf' command does not
3419 3419 apply to it:
3420 3420
3421 3421 * ``pre-run``: disabled
3422 3422
3423 3423 * ``profile-benchmark``: disabled
3424 3424
3425 3425 * ``run-limits``: disabled use --count instead
3426 3426 """
3427 3427 opts = _byteskwargs(opts)
3428 3428
3429 3429 rl = cmdutil.openrevlog(repo, b'perfrevlogwrite', file_, opts)
3430 3430 rllen = getlen(ui)(rl)
3431 3431 if startrev < 0:
3432 3432 startrev = rllen + startrev
3433 3433 if stoprev < 0:
3434 3434 stoprev = rllen + stoprev
3435 3435
3436 3436 lazydeltabase = opts['lazydeltabase']
3437 3437 source = opts['source']
3438 3438 clearcaches = opts['clear_caches']
3439 3439 validsource = (
3440 3440 b'full',
3441 3441 b'parent-1',
3442 3442 b'parent-2',
3443 3443 b'parent-smallest',
3444 3444 b'storage',
3445 3445 )
3446 3446 if source not in validsource:
3447 3447 raise error.Abort('invalid source type: %s' % source)
3448 3448
3449 3449 ### actually gather results
3450 3450 count = opts['count']
3451 3451 if count <= 0:
3452 3452 raise error.Abort('invalide run count: %d' % count)
3453 3453 allresults = []
3454 3454 for c in range(count):
3455 3455 timing = _timeonewrite(
3456 3456 ui,
3457 3457 rl,
3458 3458 source,
3459 3459 startrev,
3460 3460 stoprev,
3461 3461 c + 1,
3462 3462 lazydeltabase=lazydeltabase,
3463 3463 clearcaches=clearcaches,
3464 3464 )
3465 3465 allresults.append(timing)
3466 3466
3467 3467 ### consolidate the results in a single list
3468 3468 results = []
3469 3469 for idx, (rev, t) in enumerate(allresults[0]):
3470 3470 ts = [t]
3471 3471 for other in allresults[1:]:
3472 3472 orev, ot = other[idx]
3473 3473 assert orev == rev
3474 3474 ts.append(ot)
3475 3475 results.append((rev, ts))
3476 3476 resultcount = len(results)
3477 3477
3478 3478 ### Compute and display relevant statistics
3479 3479
3480 3480 # get a formatter
3481 3481 fm = ui.formatter(b'perf', opts)
3482 3482 displayall = ui.configbool(b"perf", b"all-timing", True)
3483 3483
3484 3484 # print individual details if requested
3485 3485 if opts['details']:
3486 3486 for idx, item in enumerate(results, 1):
3487 3487 rev, data = item
3488 3488 title = 'revisions #%d of %d, rev %d' % (idx, resultcount, rev)
3489 3489 formatone(fm, data, title=title, displayall=displayall)
3490 3490
3491 3491 # sorts results by median time
3492 3492 results.sort(key=lambda x: sorted(x[1])[len(x[1]) // 2])
3493 3493 # list of (name, index) to display)
3494 3494 relevants = [
3495 3495 ("min", 0),
3496 3496 ("10%", resultcount * 10 // 100),
3497 3497 ("25%", resultcount * 25 // 100),
3498 3498 ("50%", resultcount * 70 // 100),
3499 3499 ("75%", resultcount * 75 // 100),
3500 3500 ("90%", resultcount * 90 // 100),
3501 3501 ("95%", resultcount * 95 // 100),
3502 3502 ("99%", resultcount * 99 // 100),
3503 3503 ("99.9%", resultcount * 999 // 1000),
3504 3504 ("99.99%", resultcount * 9999 // 10000),
3505 3505 ("99.999%", resultcount * 99999 // 100000),
3506 3506 ("max", -1),
3507 3507 ]
3508 3508 if not ui.quiet:
3509 3509 for name, idx in relevants:
3510 3510 data = results[idx]
3511 3511 title = '%s of %d, rev %d' % (name, resultcount, data[0])
3512 3512 formatone(fm, data[1], title=title, displayall=displayall)
3513 3513
3514 3514 # XXX summing that many float will not be very precise, we ignore this fact
3515 3515 # for now
3516 3516 totaltime = []
3517 3517 for item in allresults:
3518 3518 totaltime.append(
3519 3519 (
3520 3520 sum(x[1][0] for x in item),
3521 3521 sum(x[1][1] for x in item),
3522 3522 sum(x[1][2] for x in item),
3523 3523 )
3524 3524 )
3525 3525 formatone(
3526 3526 fm,
3527 3527 totaltime,
3528 3528 title="total time (%d revs)" % resultcount,
3529 3529 displayall=displayall,
3530 3530 )
3531 3531 fm.end()
3532 3532
3533 3533
3534 3534 class _faketr:
3535 3535 def add(s, x, y, z=None):
3536 3536 return None
3537 3537
3538 3538
3539 3539 def _timeonewrite(
3540 3540 ui,
3541 3541 orig,
3542 3542 source,
3543 3543 startrev,
3544 3544 stoprev,
3545 3545 runidx=None,
3546 3546 lazydeltabase=True,
3547 3547 clearcaches=True,
3548 3548 ):
3549 3549 timings = []
3550 3550 tr = _faketr()
3551 3551 with _temprevlog(ui, orig, startrev) as dest:
3552 dest._lazydeltabase = lazydeltabase
3552 if hasattr(dest, "delta_config"):
3553 dest.delta_config.lazy_delta_base = lazydeltabase
3554 else:
3555 dest._lazydeltabase = lazydeltabase
3553 3556 revs = list(orig.revs(startrev, stoprev))
3554 3557 total = len(revs)
3555 3558 topic = 'adding'
3556 3559 if runidx is not None:
3557 3560 topic += ' (run #%d)' % runidx
3558 3561 # Support both old and new progress API
3559 3562 if util.safehasattr(ui, 'makeprogress'):
3560 3563 progress = ui.makeprogress(topic, unit='revs', total=total)
3561 3564
3562 3565 def updateprogress(pos):
3563 3566 progress.update(pos)
3564 3567
3565 3568 def completeprogress():
3566 3569 progress.complete()
3567 3570
3568 3571 else:
3569 3572
3570 3573 def updateprogress(pos):
3571 3574 ui.progress(topic, pos, unit='revs', total=total)
3572 3575
3573 3576 def completeprogress():
3574 3577 ui.progress(topic, None, unit='revs', total=total)
3575 3578
3576 3579 for idx, rev in enumerate(revs):
3577 3580 updateprogress(idx)
3578 3581 addargs, addkwargs = _getrevisionseed(orig, rev, tr, source)
3579 3582 if clearcaches:
3580 3583 dest.index.clearcaches()
3581 3584 dest.clearcaches()
3582 3585 with timeone() as r:
3583 3586 dest.addrawrevision(*addargs, **addkwargs)
3584 3587 timings.append((rev, r[0]))
3585 3588 updateprogress(total)
3586 3589 completeprogress()
3587 3590 return timings
3588 3591
3589 3592
3590 3593 def _getrevisionseed(orig, rev, tr, source):
3591 3594 from mercurial.node import nullid
3592 3595
3593 3596 linkrev = orig.linkrev(rev)
3594 3597 node = orig.node(rev)
3595 3598 p1, p2 = orig.parents(node)
3596 3599 flags = orig.flags(rev)
3597 3600 cachedelta = None
3598 3601 text = None
3599 3602
3600 3603 if source == b'full':
3601 3604 text = orig.revision(rev)
3602 3605 elif source == b'parent-1':
3603 3606 baserev = orig.rev(p1)
3604 3607 cachedelta = (baserev, orig.revdiff(p1, rev))
3605 3608 elif source == b'parent-2':
3606 3609 parent = p2
3607 3610 if p2 == nullid:
3608 3611 parent = p1
3609 3612 baserev = orig.rev(parent)
3610 3613 cachedelta = (baserev, orig.revdiff(parent, rev))
3611 3614 elif source == b'parent-smallest':
3612 3615 p1diff = orig.revdiff(p1, rev)
3613 3616 parent = p1
3614 3617 diff = p1diff
3615 3618 if p2 != nullid:
3616 3619 p2diff = orig.revdiff(p2, rev)
3617 3620 if len(p1diff) > len(p2diff):
3618 3621 parent = p2
3619 3622 diff = p2diff
3620 3623 baserev = orig.rev(parent)
3621 3624 cachedelta = (baserev, diff)
3622 3625 elif source == b'storage':
3623 3626 baserev = orig.deltaparent(rev)
3624 3627 cachedelta = (baserev, orig.revdiff(orig.node(baserev), rev))
3625 3628
3626 3629 return (
3627 3630 (text, tr, linkrev, p1, p2),
3628 3631 {'node': node, 'flags': flags, 'cachedelta': cachedelta},
3629 3632 )
3630 3633
3631 3634
3632 3635 @contextlib.contextmanager
3633 3636 def _temprevlog(ui, orig, truncaterev):
3634 3637 from mercurial import vfs as vfsmod
3635 3638
3636 3639 if orig._inline:
3637 3640 raise error.Abort('not supporting inline revlog (yet)')
3638 3641 revlogkwargs = {}
3639 3642 k = 'upperboundcomp'
3640 3643 if util.safehasattr(orig, k):
3641 3644 revlogkwargs[k] = getattr(orig, k)
3642 3645
3643 3646 indexfile = getattr(orig, '_indexfile', None)
3644 3647 if indexfile is None:
3645 3648 # compatibility with <= hg-5.8
3646 3649 indexfile = getattr(orig, 'indexfile')
3647 3650 origindexpath = orig.opener.join(indexfile)
3648 3651
3649 3652 datafile = getattr(orig, '_datafile', getattr(orig, 'datafile'))
3650 3653 origdatapath = orig.opener.join(datafile)
3651 3654 radix = b'revlog'
3652 3655 indexname = b'revlog.i'
3653 3656 dataname = b'revlog.d'
3654 3657
3655 3658 tmpdir = tempfile.mkdtemp(prefix='tmp-hgperf-')
3656 3659 try:
3657 3660 # copy the data file in a temporary directory
3658 3661 ui.debug('copying data in %s\n' % tmpdir)
3659 3662 destindexpath = os.path.join(tmpdir, 'revlog.i')
3660 3663 destdatapath = os.path.join(tmpdir, 'revlog.d')
3661 3664 shutil.copyfile(origindexpath, destindexpath)
3662 3665 shutil.copyfile(origdatapath, destdatapath)
3663 3666
3664 3667 # remove the data we want to add again
3665 3668 ui.debug('truncating data to be rewritten\n')
3666 3669 with open(destindexpath, 'ab') as index:
3667 3670 index.seek(0)
3668 3671 index.truncate(truncaterev * orig._io.size)
3669 3672 with open(destdatapath, 'ab') as data:
3670 3673 data.seek(0)
3671 3674 data.truncate(orig.start(truncaterev))
3672 3675
3673 3676 # instantiate a new revlog from the temporary copy
3674 3677 ui.debug('truncating adding to be rewritten\n')
3675 3678 vfs = vfsmod.vfs(tmpdir)
3676 3679 vfs.options = getattr(orig.opener, 'options', None)
3677 3680
3678 3681 try:
3679 3682 dest = revlog(vfs, radix=radix, **revlogkwargs)
3680 3683 except TypeError:
3681 3684 dest = revlog(
3682 3685 vfs, indexfile=indexname, datafile=dataname, **revlogkwargs
3683 3686 )
3684 3687 if dest._inline:
3685 3688 raise error.Abort('not supporting inline revlog (yet)')
3686 3689 # make sure internals are initialized
3687 3690 dest.revision(len(dest) - 1)
3688 3691 yield dest
3689 3692 del dest, vfs
3690 3693 finally:
3691 3694 shutil.rmtree(tmpdir, True)
3692 3695
3693 3696
3694 3697 @command(
3695 3698 b'perf::revlogchunks|perfrevlogchunks',
3696 3699 revlogopts
3697 3700 + formatteropts
3698 3701 + [
3699 3702 (b'e', b'engines', b'', b'compression engines to use'),
3700 3703 (b's', b'startrev', 0, b'revision to start at'),
3701 3704 ],
3702 3705 b'-c|-m|FILE',
3703 3706 )
3704 3707 def perfrevlogchunks(ui, repo, file_=None, engines=None, startrev=0, **opts):
3705 3708 """Benchmark operations on revlog chunks.
3706 3709
3707 3710 Logically, each revlog is a collection of fulltext revisions. However,
3708 3711 stored within each revlog are "chunks" of possibly compressed data. This
3709 3712 data needs to be read and decompressed or compressed and written.
3710 3713
3711 3714 This command measures the time it takes to read+decompress and recompress
3712 3715 chunks in a revlog. It effectively isolates I/O and compression performance.
3713 3716 For measurements of higher-level operations like resolving revisions,
3714 3717 see ``perfrevlogrevisions`` and ``perfrevlogrevision``.
3715 3718 """
3716 3719 opts = _byteskwargs(opts)
3717 3720
3718 3721 rl = cmdutil.openrevlog(repo, b'perfrevlogchunks', file_, opts)
3719 3722
3720 3723 # _chunkraw was renamed to _getsegmentforrevs.
3721 3724 try:
3722 3725 segmentforrevs = rl._getsegmentforrevs
3723 3726 except AttributeError:
3724 3727 segmentforrevs = rl._chunkraw
3725 3728
3726 3729 # Verify engines argument.
3727 3730 if engines:
3728 3731 engines = {e.strip() for e in engines.split(b',')}
3729 3732 for engine in engines:
3730 3733 try:
3731 3734 util.compressionengines[engine]
3732 3735 except KeyError:
3733 3736 raise error.Abort(b'unknown compression engine: %s' % engine)
3734 3737 else:
3735 3738 engines = []
3736 3739 for e in util.compengines:
3737 3740 engine = util.compengines[e]
3738 3741 try:
3739 3742 if engine.available():
3740 3743 engine.revlogcompressor().compress(b'dummy')
3741 3744 engines.append(e)
3742 3745 except NotImplementedError:
3743 3746 pass
3744 3747
3745 3748 revs = list(rl.revs(startrev, len(rl) - 1))
3746 3749
3747 3750 @contextlib.contextmanager
3748 3751 def reading(rl):
3749 3752 if getattr(rl, 'reading', None) is not None:
3750 3753 with rl.reading():
3751 3754 yield None
3752 3755 elif rl._inline:
3753 3756 indexfile = getattr(rl, '_indexfile', None)
3754 3757 if indexfile is None:
3755 3758 # compatibility with <= hg-5.8
3756 3759 indexfile = getattr(rl, 'indexfile')
3757 3760 yield getsvfs(repo)(indexfile)
3758 3761 else:
3759 3762 datafile = getattr(rl, 'datafile', getattr(rl, 'datafile'))
3760 3763 yield getsvfs(repo)(datafile)
3761 3764
3762 3765 if getattr(rl, 'reading', None) is not None:
3763 3766
3764 3767 @contextlib.contextmanager
3765 3768 def lazy_reading(rl):
3766 3769 with rl.reading():
3767 3770 yield
3768 3771
3769 3772 else:
3770 3773
3771 3774 @contextlib.contextmanager
3772 3775 def lazy_reading(rl):
3773 3776 yield
3774 3777
3775 3778 def doread():
3776 3779 rl.clearcaches()
3777 3780 for rev in revs:
3778 3781 with lazy_reading(rl):
3779 3782 segmentforrevs(rev, rev)
3780 3783
3781 3784 def doreadcachedfh():
3782 3785 rl.clearcaches()
3783 3786 with reading(rl) as fh:
3784 3787 if fh is not None:
3785 3788 for rev in revs:
3786 3789 segmentforrevs(rev, rev, df=fh)
3787 3790 else:
3788 3791 for rev in revs:
3789 3792 segmentforrevs(rev, rev)
3790 3793
3791 3794 def doreadbatch():
3792 3795 rl.clearcaches()
3793 3796 with lazy_reading(rl):
3794 3797 segmentforrevs(revs[0], revs[-1])
3795 3798
3796 3799 def doreadbatchcachedfh():
3797 3800 rl.clearcaches()
3798 3801 with reading(rl) as fh:
3799 3802 if fh is not None:
3800 3803 segmentforrevs(revs[0], revs[-1], df=fh)
3801 3804 else:
3802 3805 segmentforrevs(revs[0], revs[-1])
3803 3806
3804 3807 def dochunk():
3805 3808 rl.clearcaches()
3806 3809 with reading(rl) as fh:
3807 3810 if fh is not None:
3808 3811 for rev in revs:
3809 3812 rl._chunk(rev, df=fh)
3810 3813 else:
3811 3814 for rev in revs:
3812 3815 rl._chunk(rev)
3813 3816
3814 3817 chunks = [None]
3815 3818
3816 3819 def dochunkbatch():
3817 3820 rl.clearcaches()
3818 3821 with reading(rl) as fh:
3819 3822 if fh is not None:
3820 3823 # Save chunks as a side-effect.
3821 3824 chunks[0] = rl._chunks(revs, df=fh)
3822 3825 else:
3823 3826 # Save chunks as a side-effect.
3824 3827 chunks[0] = rl._chunks(revs)
3825 3828
3826 3829 def docompress(compressor):
3827 3830 rl.clearcaches()
3828 3831
3829 3832 try:
3830 3833 # Swap in the requested compression engine.
3831 3834 oldcompressor = rl._compressor
3832 3835 rl._compressor = compressor
3833 3836 for chunk in chunks[0]:
3834 3837 rl.compress(chunk)
3835 3838 finally:
3836 3839 rl._compressor = oldcompressor
3837 3840
3838 3841 benches = [
3839 3842 (lambda: doread(), b'read'),
3840 3843 (lambda: doreadcachedfh(), b'read w/ reused fd'),
3841 3844 (lambda: doreadbatch(), b'read batch'),
3842 3845 (lambda: doreadbatchcachedfh(), b'read batch w/ reused fd'),
3843 3846 (lambda: dochunk(), b'chunk'),
3844 3847 (lambda: dochunkbatch(), b'chunk batch'),
3845 3848 ]
3846 3849
3847 3850 for engine in sorted(engines):
3848 3851 compressor = util.compengines[engine].revlogcompressor()
3849 3852 benches.append(
3850 3853 (
3851 3854 functools.partial(docompress, compressor),
3852 3855 b'compress w/ %s' % engine,
3853 3856 )
3854 3857 )
3855 3858
3856 3859 for fn, title in benches:
3857 3860 timer, fm = gettimer(ui, opts)
3858 3861 timer(fn, title=title)
3859 3862 fm.end()
3860 3863
3861 3864
3862 3865 @command(
3863 3866 b'perf::revlogrevision|perfrevlogrevision',
3864 3867 revlogopts
3865 3868 + formatteropts
3866 3869 + [(b'', b'cache', False, b'use caches instead of clearing')],
3867 3870 b'-c|-m|FILE REV',
3868 3871 )
3869 3872 def perfrevlogrevision(ui, repo, file_, rev=None, cache=None, **opts):
3870 3873 """Benchmark obtaining a revlog revision.
3871 3874
3872 3875 Obtaining a revlog revision consists of roughly the following steps:
3873 3876
3874 3877 1. Compute the delta chain
3875 3878 2. Slice the delta chain if applicable
3876 3879 3. Obtain the raw chunks for that delta chain
3877 3880 4. Decompress each raw chunk
3878 3881 5. Apply binary patches to obtain fulltext
3879 3882 6. Verify hash of fulltext
3880 3883
3881 3884 This command measures the time spent in each of these phases.
3882 3885 """
3883 3886 opts = _byteskwargs(opts)
3884 3887
3885 3888 if opts.get(b'changelog') or opts.get(b'manifest'):
3886 3889 file_, rev = None, file_
3887 3890 elif rev is None:
3888 3891 raise error.CommandError(b'perfrevlogrevision', b'invalid arguments')
3889 3892
3890 3893 r = cmdutil.openrevlog(repo, b'perfrevlogrevision', file_, opts)
3891 3894
3892 3895 # _chunkraw was renamed to _getsegmentforrevs.
3893 3896 try:
3894 3897 segmentforrevs = r._getsegmentforrevs
3895 3898 except AttributeError:
3896 3899 segmentforrevs = r._chunkraw
3897 3900
3898 3901 node = r.lookup(rev)
3899 3902 rev = r.rev(node)
3900 3903
3901 3904 if getattr(r, 'reading', None) is not None:
3902 3905
3903 3906 @contextlib.contextmanager
3904 3907 def lazy_reading(r):
3905 3908 with r.reading():
3906 3909 yield
3907 3910
3908 3911 else:
3909 3912
3910 3913 @contextlib.contextmanager
3911 3914 def lazy_reading(r):
3912 3915 yield
3913 3916
3914 3917 def getrawchunks(data, chain):
3915 3918 start = r.start
3916 3919 length = r.length
3917 3920 inline = r._inline
3918 3921 try:
3919 3922 iosize = r.index.entry_size
3920 3923 except AttributeError:
3921 3924 iosize = r._io.size
3922 3925 buffer = util.buffer
3923 3926
3924 3927 chunks = []
3925 3928 ladd = chunks.append
3926 3929 for idx, item in enumerate(chain):
3927 3930 offset = start(item[0])
3928 3931 bits = data[idx]
3929 3932 for rev in item:
3930 3933 chunkstart = start(rev)
3931 3934 if inline:
3932 3935 chunkstart += (rev + 1) * iosize
3933 3936 chunklength = length(rev)
3934 3937 ladd(buffer(bits, chunkstart - offset, chunklength))
3935 3938
3936 3939 return chunks
3937 3940
3938 3941 def dodeltachain(rev):
3939 3942 if not cache:
3940 3943 r.clearcaches()
3941 3944 r._deltachain(rev)
3942 3945
3943 3946 def doread(chain):
3944 3947 if not cache:
3945 3948 r.clearcaches()
3946 3949 for item in slicedchain:
3947 3950 with lazy_reading(r):
3948 3951 segmentforrevs(item[0], item[-1])
3949 3952
3950 3953 def doslice(r, chain, size):
3951 3954 for s in slicechunk(r, chain, targetsize=size):
3952 3955 pass
3953 3956
3954 3957 def dorawchunks(data, chain):
3955 3958 if not cache:
3956 3959 r.clearcaches()
3957 3960 getrawchunks(data, chain)
3958 3961
3959 3962 def dodecompress(chunks):
3960 3963 decomp = r.decompress
3961 3964 for chunk in chunks:
3962 3965 decomp(chunk)
3963 3966
3964 3967 def dopatch(text, bins):
3965 3968 if not cache:
3966 3969 r.clearcaches()
3967 3970 mdiff.patches(text, bins)
3968 3971
3969 3972 def dohash(text):
3970 3973 if not cache:
3971 3974 r.clearcaches()
3972 3975 r.checkhash(text, node, rev=rev)
3973 3976
3974 3977 def dorevision():
3975 3978 if not cache:
3976 3979 r.clearcaches()
3977 3980 r.revision(node)
3978 3981
3979 3982 try:
3980 3983 from mercurial.revlogutils.deltas import slicechunk
3981 3984 except ImportError:
3982 3985 slicechunk = getattr(revlog, '_slicechunk', None)
3983 3986
3984 3987 size = r.length(rev)
3985 3988 chain = r._deltachain(rev)[0]
3986 3989
3987 3990 with_sparse_read = False
3988 3991 if hasattr(r, 'data_config'):
3989 3992 with_sparse_read = r.data_config.with_sparse_read
3990 3993 elif hasattr(r, '_withsparseread'):
3991 3994 with_sparse_read = r._withsparseread
3992 3995 if with_sparse_read:
3993 3996 slicedchain = (chain,)
3994 3997 else:
3995 3998 slicedchain = tuple(slicechunk(r, chain, targetsize=size))
3996 3999 data = [segmentforrevs(seg[0], seg[-1])[1] for seg in slicedchain]
3997 4000 rawchunks = getrawchunks(data, slicedchain)
3998 4001 bins = r._chunks(chain)
3999 4002 text = bytes(bins[0])
4000 4003 bins = bins[1:]
4001 4004 text = mdiff.patches(text, bins)
4002 4005
4003 4006 benches = [
4004 4007 (lambda: dorevision(), b'full'),
4005 4008 (lambda: dodeltachain(rev), b'deltachain'),
4006 4009 (lambda: doread(chain), b'read'),
4007 4010 ]
4008 4011
4009 4012 if with_sparse_read:
4010 4013 slicing = (lambda: doslice(r, chain, size), b'slice-sparse-chain')
4011 4014 benches.append(slicing)
4012 4015
4013 4016 benches.extend(
4014 4017 [
4015 4018 (lambda: dorawchunks(data, slicedchain), b'rawchunks'),
4016 4019 (lambda: dodecompress(rawchunks), b'decompress'),
4017 4020 (lambda: dopatch(text, bins), b'patch'),
4018 4021 (lambda: dohash(text), b'hash'),
4019 4022 ]
4020 4023 )
4021 4024
4022 4025 timer, fm = gettimer(ui, opts)
4023 4026 for fn, title in benches:
4024 4027 timer(fn, title=title)
4025 4028 fm.end()
4026 4029
4027 4030
4028 4031 @command(
4029 4032 b'perf::revset|perfrevset',
4030 4033 [
4031 4034 (b'C', b'clear', False, b'clear volatile cache between each call.'),
4032 4035 (b'', b'contexts', False, b'obtain changectx for each revision'),
4033 4036 ]
4034 4037 + formatteropts,
4035 4038 b"REVSET",
4036 4039 )
4037 4040 def perfrevset(ui, repo, expr, clear=False, contexts=False, **opts):
4038 4041 """benchmark the execution time of a revset
4039 4042
4040 4043 Use the --clean option if need to evaluate the impact of build volatile
4041 4044 revisions set cache on the revset execution. Volatile cache hold filtered
4042 4045 and obsolete related cache."""
4043 4046 opts = _byteskwargs(opts)
4044 4047
4045 4048 timer, fm = gettimer(ui, opts)
4046 4049
4047 4050 def d():
4048 4051 if clear:
4049 4052 repo.invalidatevolatilesets()
4050 4053 if contexts:
4051 4054 for ctx in repo.set(expr):
4052 4055 pass
4053 4056 else:
4054 4057 for r in repo.revs(expr):
4055 4058 pass
4056 4059
4057 4060 timer(d)
4058 4061 fm.end()
4059 4062
4060 4063
4061 4064 @command(
4062 4065 b'perf::volatilesets|perfvolatilesets',
4063 4066 [
4064 4067 (b'', b'clear-obsstore', False, b'drop obsstore between each call.'),
4065 4068 ]
4066 4069 + formatteropts,
4067 4070 )
4068 4071 def perfvolatilesets(ui, repo, *names, **opts):
4069 4072 """benchmark the computation of various volatile set
4070 4073
4071 4074 Volatile set computes element related to filtering and obsolescence."""
4072 4075 opts = _byteskwargs(opts)
4073 4076 timer, fm = gettimer(ui, opts)
4074 4077 repo = repo.unfiltered()
4075 4078
4076 4079 def getobs(name):
4077 4080 def d():
4078 4081 repo.invalidatevolatilesets()
4079 4082 if opts[b'clear_obsstore']:
4080 4083 clearfilecache(repo, b'obsstore')
4081 4084 obsolete.getrevs(repo, name)
4082 4085
4083 4086 return d
4084 4087
4085 4088 allobs = sorted(obsolete.cachefuncs)
4086 4089 if names:
4087 4090 allobs = [n for n in allobs if n in names]
4088 4091
4089 4092 for name in allobs:
4090 4093 timer(getobs(name), title=name)
4091 4094
4092 4095 def getfiltered(name):
4093 4096 def d():
4094 4097 repo.invalidatevolatilesets()
4095 4098 if opts[b'clear_obsstore']:
4096 4099 clearfilecache(repo, b'obsstore')
4097 4100 repoview.filterrevs(repo, name)
4098 4101
4099 4102 return d
4100 4103
4101 4104 allfilter = sorted(repoview.filtertable)
4102 4105 if names:
4103 4106 allfilter = [n for n in allfilter if n in names]
4104 4107
4105 4108 for name in allfilter:
4106 4109 timer(getfiltered(name), title=name)
4107 4110 fm.end()
4108 4111
4109 4112
4110 4113 @command(
4111 4114 b'perf::branchmap|perfbranchmap',
4112 4115 [
4113 4116 (b'f', b'full', False, b'Includes build time of subset'),
4114 4117 (
4115 4118 b'',
4116 4119 b'clear-revbranch',
4117 4120 False,
4118 4121 b'purge the revbranch cache between computation',
4119 4122 ),
4120 4123 ]
4121 4124 + formatteropts,
4122 4125 )
4123 4126 def perfbranchmap(ui, repo, *filternames, **opts):
4124 4127 """benchmark the update of a branchmap
4125 4128
4126 4129 This benchmarks the full repo.branchmap() call with read and write disabled
4127 4130 """
4128 4131 opts = _byteskwargs(opts)
4129 4132 full = opts.get(b"full", False)
4130 4133 clear_revbranch = opts.get(b"clear_revbranch", False)
4131 4134 timer, fm = gettimer(ui, opts)
4132 4135
4133 4136 def getbranchmap(filtername):
4134 4137 """generate a benchmark function for the filtername"""
4135 4138 if filtername is None:
4136 4139 view = repo
4137 4140 else:
4138 4141 view = repo.filtered(filtername)
4139 4142 if util.safehasattr(view._branchcaches, '_per_filter'):
4140 4143 filtered = view._branchcaches._per_filter
4141 4144 else:
4142 4145 # older versions
4143 4146 filtered = view._branchcaches
4144 4147
4145 4148 def d():
4146 4149 if clear_revbranch:
4147 4150 repo.revbranchcache()._clear()
4148 4151 if full:
4149 4152 view._branchcaches.clear()
4150 4153 else:
4151 4154 filtered.pop(filtername, None)
4152 4155 view.branchmap()
4153 4156
4154 4157 return d
4155 4158
4156 4159 # add filter in smaller subset to bigger subset
4157 4160 possiblefilters = set(repoview.filtertable)
4158 4161 if filternames:
4159 4162 possiblefilters &= set(filternames)
4160 4163 subsettable = getbranchmapsubsettable()
4161 4164 allfilters = []
4162 4165 while possiblefilters:
4163 4166 for name in possiblefilters:
4164 4167 subset = subsettable.get(name)
4165 4168 if subset not in possiblefilters:
4166 4169 break
4167 4170 else:
4168 4171 assert False, b'subset cycle %s!' % possiblefilters
4169 4172 allfilters.append(name)
4170 4173 possiblefilters.remove(name)
4171 4174
4172 4175 # warm the cache
4173 4176 if not full:
4174 4177 for name in allfilters:
4175 4178 repo.filtered(name).branchmap()
4176 4179 if not filternames or b'unfiltered' in filternames:
4177 4180 # add unfiltered
4178 4181 allfilters.append(None)
4179 4182
4180 4183 if util.safehasattr(branchmap.branchcache, 'fromfile'):
4181 4184 branchcacheread = safeattrsetter(branchmap.branchcache, b'fromfile')
4182 4185 branchcacheread.set(classmethod(lambda *args: None))
4183 4186 else:
4184 4187 # older versions
4185 4188 branchcacheread = safeattrsetter(branchmap, b'read')
4186 4189 branchcacheread.set(lambda *args: None)
4187 4190 branchcachewrite = safeattrsetter(branchmap.branchcache, b'write')
4188 4191 branchcachewrite.set(lambda *args: None)
4189 4192 try:
4190 4193 for name in allfilters:
4191 4194 printname = name
4192 4195 if name is None:
4193 4196 printname = b'unfiltered'
4194 4197 timer(getbranchmap(name), title=printname)
4195 4198 finally:
4196 4199 branchcacheread.restore()
4197 4200 branchcachewrite.restore()
4198 4201 fm.end()
4199 4202
4200 4203
4201 4204 @command(
4202 4205 b'perf::branchmapupdate|perfbranchmapupdate',
4203 4206 [
4204 4207 (b'', b'base', [], b'subset of revision to start from'),
4205 4208 (b'', b'target', [], b'subset of revision to end with'),
4206 4209 (b'', b'clear-caches', False, b'clear cache between each runs'),
4207 4210 ]
4208 4211 + formatteropts,
4209 4212 )
4210 4213 def perfbranchmapupdate(ui, repo, base=(), target=(), **opts):
4211 4214 """benchmark branchmap update from for <base> revs to <target> revs
4212 4215
4213 4216 If `--clear-caches` is passed, the following items will be reset before
4214 4217 each update:
4215 4218 * the changelog instance and associated indexes
4216 4219 * the rev-branch-cache instance
4217 4220
4218 4221 Examples:
4219 4222
4220 4223 # update for the one last revision
4221 4224 $ hg perfbranchmapupdate --base 'not tip' --target 'tip'
4222 4225
4223 4226 $ update for change coming with a new branch
4224 4227 $ hg perfbranchmapupdate --base 'stable' --target 'default'
4225 4228 """
4226 4229 from mercurial import branchmap
4227 4230 from mercurial import repoview
4228 4231
4229 4232 opts = _byteskwargs(opts)
4230 4233 timer, fm = gettimer(ui, opts)
4231 4234 clearcaches = opts[b'clear_caches']
4232 4235 unfi = repo.unfiltered()
4233 4236 x = [None] # used to pass data between closure
4234 4237
4235 4238 # we use a `list` here to avoid possible side effect from smartset
4236 4239 baserevs = list(scmutil.revrange(repo, base))
4237 4240 targetrevs = list(scmutil.revrange(repo, target))
4238 4241 if not baserevs:
4239 4242 raise error.Abort(b'no revisions selected for --base')
4240 4243 if not targetrevs:
4241 4244 raise error.Abort(b'no revisions selected for --target')
4242 4245
4243 4246 # make sure the target branchmap also contains the one in the base
4244 4247 targetrevs = list(set(baserevs) | set(targetrevs))
4245 4248 targetrevs.sort()
4246 4249
4247 4250 cl = repo.changelog
4248 4251 allbaserevs = list(cl.ancestors(baserevs, inclusive=True))
4249 4252 allbaserevs.sort()
4250 4253 alltargetrevs = frozenset(cl.ancestors(targetrevs, inclusive=True))
4251 4254
4252 4255 newrevs = list(alltargetrevs.difference(allbaserevs))
4253 4256 newrevs.sort()
4254 4257
4255 4258 allrevs = frozenset(unfi.changelog.revs())
4256 4259 basefilterrevs = frozenset(allrevs.difference(allbaserevs))
4257 4260 targetfilterrevs = frozenset(allrevs.difference(alltargetrevs))
4258 4261
4259 4262 def basefilter(repo, visibilityexceptions=None):
4260 4263 return basefilterrevs
4261 4264
4262 4265 def targetfilter(repo, visibilityexceptions=None):
4263 4266 return targetfilterrevs
4264 4267
4265 4268 msg = b'benchmark of branchmap with %d revisions with %d new ones\n'
4266 4269 ui.status(msg % (len(allbaserevs), len(newrevs)))
4267 4270 if targetfilterrevs:
4268 4271 msg = b'(%d revisions still filtered)\n'
4269 4272 ui.status(msg % len(targetfilterrevs))
4270 4273
4271 4274 try:
4272 4275 repoview.filtertable[b'__perf_branchmap_update_base'] = basefilter
4273 4276 repoview.filtertable[b'__perf_branchmap_update_target'] = targetfilter
4274 4277
4275 4278 baserepo = repo.filtered(b'__perf_branchmap_update_base')
4276 4279 targetrepo = repo.filtered(b'__perf_branchmap_update_target')
4277 4280
4278 4281 # try to find an existing branchmap to reuse
4279 4282 subsettable = getbranchmapsubsettable()
4280 4283 candidatefilter = subsettable.get(None)
4281 4284 while candidatefilter is not None:
4282 4285 candidatebm = repo.filtered(candidatefilter).branchmap()
4283 4286 if candidatebm.validfor(baserepo):
4284 4287 filtered = repoview.filterrevs(repo, candidatefilter)
4285 4288 missing = [r for r in allbaserevs if r in filtered]
4286 4289 base = candidatebm.copy()
4287 4290 base.update(baserepo, missing)
4288 4291 break
4289 4292 candidatefilter = subsettable.get(candidatefilter)
4290 4293 else:
4291 4294 # no suitable subset where found
4292 4295 base = branchmap.branchcache()
4293 4296 base.update(baserepo, allbaserevs)
4294 4297
4295 4298 def setup():
4296 4299 x[0] = base.copy()
4297 4300 if clearcaches:
4298 4301 unfi._revbranchcache = None
4299 4302 clearchangelog(repo)
4300 4303
4301 4304 def bench():
4302 4305 x[0].update(targetrepo, newrevs)
4303 4306
4304 4307 timer(bench, setup=setup)
4305 4308 fm.end()
4306 4309 finally:
4307 4310 repoview.filtertable.pop(b'__perf_branchmap_update_base', None)
4308 4311 repoview.filtertable.pop(b'__perf_branchmap_update_target', None)
4309 4312
4310 4313
4311 4314 @command(
4312 4315 b'perf::branchmapload|perfbranchmapload',
4313 4316 [
4314 4317 (b'f', b'filter', b'', b'Specify repoview filter'),
4315 4318 (b'', b'list', False, b'List brachmap filter caches'),
4316 4319 (b'', b'clear-revlogs', False, b'refresh changelog and manifest'),
4317 4320 ]
4318 4321 + formatteropts,
4319 4322 )
4320 4323 def perfbranchmapload(ui, repo, filter=b'', list=False, **opts):
4321 4324 """benchmark reading the branchmap"""
4322 4325 opts = _byteskwargs(opts)
4323 4326 clearrevlogs = opts[b'clear_revlogs']
4324 4327
4325 4328 if list:
4326 4329 for name, kind, st in repo.cachevfs.readdir(stat=True):
4327 4330 if name.startswith(b'branch2'):
4328 4331 filtername = name.partition(b'-')[2] or b'unfiltered'
4329 4332 ui.status(
4330 4333 b'%s - %s\n' % (filtername, util.bytecount(st.st_size))
4331 4334 )
4332 4335 return
4333 4336 if not filter:
4334 4337 filter = None
4335 4338 subsettable = getbranchmapsubsettable()
4336 4339 if filter is None:
4337 4340 repo = repo.unfiltered()
4338 4341 else:
4339 4342 repo = repoview.repoview(repo, filter)
4340 4343
4341 4344 repo.branchmap() # make sure we have a relevant, up to date branchmap
4342 4345
4343 4346 try:
4344 4347 fromfile = branchmap.branchcache.fromfile
4345 4348 except AttributeError:
4346 4349 # older versions
4347 4350 fromfile = branchmap.read
4348 4351
4349 4352 currentfilter = filter
4350 4353 # try once without timer, the filter may not be cached
4351 4354 while fromfile(repo) is None:
4352 4355 currentfilter = subsettable.get(currentfilter)
4353 4356 if currentfilter is None:
4354 4357 raise error.Abort(
4355 4358 b'No branchmap cached for %s repo' % (filter or b'unfiltered')
4356 4359 )
4357 4360 repo = repo.filtered(currentfilter)
4358 4361 timer, fm = gettimer(ui, opts)
4359 4362
4360 4363 def setup():
4361 4364 if clearrevlogs:
4362 4365 clearchangelog(repo)
4363 4366
4364 4367 def bench():
4365 4368 fromfile(repo)
4366 4369
4367 4370 timer(bench, setup=setup)
4368 4371 fm.end()
4369 4372
4370 4373
4371 4374 @command(b'perf::loadmarkers|perfloadmarkers')
4372 4375 def perfloadmarkers(ui, repo):
4373 4376 """benchmark the time to parse the on-disk markers for a repo
4374 4377
4375 4378 Result is the number of markers in the repo."""
4376 4379 timer, fm = gettimer(ui)
4377 4380 svfs = getsvfs(repo)
4378 4381 timer(lambda: len(obsolete.obsstore(repo, svfs)))
4379 4382 fm.end()
4380 4383
4381 4384
4382 4385 @command(
4383 4386 b'perf::lrucachedict|perflrucachedict',
4384 4387 formatteropts
4385 4388 + [
4386 4389 (b'', b'costlimit', 0, b'maximum total cost of items in cache'),
4387 4390 (b'', b'mincost', 0, b'smallest cost of items in cache'),
4388 4391 (b'', b'maxcost', 100, b'maximum cost of items in cache'),
4389 4392 (b'', b'size', 4, b'size of cache'),
4390 4393 (b'', b'gets', 10000, b'number of key lookups'),
4391 4394 (b'', b'sets', 10000, b'number of key sets'),
4392 4395 (b'', b'mixed', 10000, b'number of mixed mode operations'),
4393 4396 (
4394 4397 b'',
4395 4398 b'mixedgetfreq',
4396 4399 50,
4397 4400 b'frequency of get vs set ops in mixed mode',
4398 4401 ),
4399 4402 ],
4400 4403 norepo=True,
4401 4404 )
4402 4405 def perflrucache(
4403 4406 ui,
4404 4407 mincost=0,
4405 4408 maxcost=100,
4406 4409 costlimit=0,
4407 4410 size=4,
4408 4411 gets=10000,
4409 4412 sets=10000,
4410 4413 mixed=10000,
4411 4414 mixedgetfreq=50,
4412 4415 **opts
4413 4416 ):
4414 4417 opts = _byteskwargs(opts)
4415 4418
4416 4419 def doinit():
4417 4420 for i in _xrange(10000):
4418 4421 util.lrucachedict(size)
4419 4422
4420 4423 costrange = list(range(mincost, maxcost + 1))
4421 4424
4422 4425 values = []
4423 4426 for i in _xrange(size):
4424 4427 values.append(random.randint(0, _maxint))
4425 4428
4426 4429 # Get mode fills the cache and tests raw lookup performance with no
4427 4430 # eviction.
4428 4431 getseq = []
4429 4432 for i in _xrange(gets):
4430 4433 getseq.append(random.choice(values))
4431 4434
4432 4435 def dogets():
4433 4436 d = util.lrucachedict(size)
4434 4437 for v in values:
4435 4438 d[v] = v
4436 4439 for key in getseq:
4437 4440 value = d[key]
4438 4441 value # silence pyflakes warning
4439 4442
4440 4443 def dogetscost():
4441 4444 d = util.lrucachedict(size, maxcost=costlimit)
4442 4445 for i, v in enumerate(values):
4443 4446 d.insert(v, v, cost=costs[i])
4444 4447 for key in getseq:
4445 4448 try:
4446 4449 value = d[key]
4447 4450 value # silence pyflakes warning
4448 4451 except KeyError:
4449 4452 pass
4450 4453
4451 4454 # Set mode tests insertion speed with cache eviction.
4452 4455 setseq = []
4453 4456 costs = []
4454 4457 for i in _xrange(sets):
4455 4458 setseq.append(random.randint(0, _maxint))
4456 4459 costs.append(random.choice(costrange))
4457 4460
4458 4461 def doinserts():
4459 4462 d = util.lrucachedict(size)
4460 4463 for v in setseq:
4461 4464 d.insert(v, v)
4462 4465
4463 4466 def doinsertscost():
4464 4467 d = util.lrucachedict(size, maxcost=costlimit)
4465 4468 for i, v in enumerate(setseq):
4466 4469 d.insert(v, v, cost=costs[i])
4467 4470
4468 4471 def dosets():
4469 4472 d = util.lrucachedict(size)
4470 4473 for v in setseq:
4471 4474 d[v] = v
4472 4475
4473 4476 # Mixed mode randomly performs gets and sets with eviction.
4474 4477 mixedops = []
4475 4478 for i in _xrange(mixed):
4476 4479 r = random.randint(0, 100)
4477 4480 if r < mixedgetfreq:
4478 4481 op = 0
4479 4482 else:
4480 4483 op = 1
4481 4484
4482 4485 mixedops.append(
4483 4486 (op, random.randint(0, size * 2), random.choice(costrange))
4484 4487 )
4485 4488
4486 4489 def domixed():
4487 4490 d = util.lrucachedict(size)
4488 4491
4489 4492 for op, v, cost in mixedops:
4490 4493 if op == 0:
4491 4494 try:
4492 4495 d[v]
4493 4496 except KeyError:
4494 4497 pass
4495 4498 else:
4496 4499 d[v] = v
4497 4500
4498 4501 def domixedcost():
4499 4502 d = util.lrucachedict(size, maxcost=costlimit)
4500 4503
4501 4504 for op, v, cost in mixedops:
4502 4505 if op == 0:
4503 4506 try:
4504 4507 d[v]
4505 4508 except KeyError:
4506 4509 pass
4507 4510 else:
4508 4511 d.insert(v, v, cost=cost)
4509 4512
4510 4513 benches = [
4511 4514 (doinit, b'init'),
4512 4515 ]
4513 4516
4514 4517 if costlimit:
4515 4518 benches.extend(
4516 4519 [
4517 4520 (dogetscost, b'gets w/ cost limit'),
4518 4521 (doinsertscost, b'inserts w/ cost limit'),
4519 4522 (domixedcost, b'mixed w/ cost limit'),
4520 4523 ]
4521 4524 )
4522 4525 else:
4523 4526 benches.extend(
4524 4527 [
4525 4528 (dogets, b'gets'),
4526 4529 (doinserts, b'inserts'),
4527 4530 (dosets, b'sets'),
4528 4531 (domixed, b'mixed'),
4529 4532 ]
4530 4533 )
4531 4534
4532 4535 for fn, title in benches:
4533 4536 timer, fm = gettimer(ui, opts)
4534 4537 timer(fn, title=title)
4535 4538 fm.end()
4536 4539
4537 4540
4538 4541 @command(
4539 4542 b'perf::write|perfwrite',
4540 4543 formatteropts
4541 4544 + [
4542 4545 (b'', b'write-method', b'write', b'ui write method'),
4543 4546 (b'', b'nlines', 100, b'number of lines'),
4544 4547 (b'', b'nitems', 100, b'number of items (per line)'),
4545 4548 (b'', b'item', b'x', b'item that is written'),
4546 4549 (b'', b'batch-line', None, b'pass whole line to write method at once'),
4547 4550 (b'', b'flush-line', None, b'flush after each line'),
4548 4551 ],
4549 4552 )
4550 4553 def perfwrite(ui, repo, **opts):
4551 4554 """microbenchmark ui.write (and others)"""
4552 4555 opts = _byteskwargs(opts)
4553 4556
4554 4557 write = getattr(ui, _sysstr(opts[b'write_method']))
4555 4558 nlines = int(opts[b'nlines'])
4556 4559 nitems = int(opts[b'nitems'])
4557 4560 item = opts[b'item']
4558 4561 batch_line = opts.get(b'batch_line')
4559 4562 flush_line = opts.get(b'flush_line')
4560 4563
4561 4564 if batch_line:
4562 4565 line = item * nitems + b'\n'
4563 4566
4564 4567 def benchmark():
4565 4568 for i in pycompat.xrange(nlines):
4566 4569 if batch_line:
4567 4570 write(line)
4568 4571 else:
4569 4572 for i in pycompat.xrange(nitems):
4570 4573 write(item)
4571 4574 write(b'\n')
4572 4575 if flush_line:
4573 4576 ui.flush()
4574 4577 ui.flush()
4575 4578
4576 4579 timer, fm = gettimer(ui, opts)
4577 4580 timer(benchmark)
4578 4581 fm.end()
4579 4582
4580 4583
4581 4584 def uisetup(ui):
4582 4585 if util.safehasattr(cmdutil, b'openrevlog') and not util.safehasattr(
4583 4586 commands, b'debugrevlogopts'
4584 4587 ):
4585 4588 # for "historical portability":
4586 4589 # In this case, Mercurial should be 1.9 (or a79fea6b3e77) -
4587 4590 # 3.7 (or 5606f7d0d063). Therefore, '--dir' option for
4588 4591 # openrevlog() should cause failure, because it has been
4589 4592 # available since 3.5 (or 49c583ca48c4).
4590 4593 def openrevlog(orig, repo, cmd, file_, opts):
4591 4594 if opts.get(b'dir') and not util.safehasattr(repo, b'dirlog'):
4592 4595 raise error.Abort(
4593 4596 b"This version doesn't support --dir option",
4594 4597 hint=b"use 3.5 or later",
4595 4598 )
4596 4599 return orig(repo, cmd, file_, opts)
4597 4600
4598 4601 name = _sysstr(b'openrevlog')
4599 4602 extensions.wrapfunction(cmdutil, name, openrevlog)
4600 4603
4601 4604
4602 4605 @command(
4603 4606 b'perf::progress|perfprogress',
4604 4607 formatteropts
4605 4608 + [
4606 4609 (b'', b'topic', b'topic', b'topic for progress messages'),
4607 4610 (b'c', b'total', 1000000, b'total value we are progressing to'),
4608 4611 ],
4609 4612 norepo=True,
4610 4613 )
4611 4614 def perfprogress(ui, topic=None, total=None, **opts):
4612 4615 """printing of progress bars"""
4613 4616 opts = _byteskwargs(opts)
4614 4617
4615 4618 timer, fm = gettimer(ui, opts)
4616 4619
4617 4620 def doprogress():
4618 4621 with ui.makeprogress(topic, total=total) as progress:
4619 4622 for i in _xrange(total):
4620 4623 progress.increment()
4621 4624
4622 4625 timer(doprogress)
4623 4626 fm.end()
@@ -1,1630 +1,1630 b''
1 1 # revlogdeltas.py - Logic around delta computation for revlog
2 2 #
3 3 # Copyright 2005-2007 Olivia Mackall <olivia@selenic.com>
4 4 # Copyright 2018 Octobus <contact@octobus.net>
5 5 #
6 6 # This software may be used and distributed according to the terms of the
7 7 # GNU General Public License version 2 or any later version.
8 8 """Helper class to compute deltas stored inside revlogs"""
9 9
10 10
11 11 import collections
12 12 import struct
13 13
14 14 # import stuff from node for others to import from revlog
15 15 from ..node import nullrev
16 16 from ..i18n import _
17 17
18 18 from .constants import (
19 19 COMP_MODE_DEFAULT,
20 20 COMP_MODE_INLINE,
21 21 COMP_MODE_PLAIN,
22 22 DELTA_BASE_REUSE_FORCE,
23 23 DELTA_BASE_REUSE_NO,
24 24 KIND_CHANGELOG,
25 25 KIND_FILELOG,
26 26 KIND_MANIFESTLOG,
27 27 REVIDX_ISCENSORED,
28 28 REVIDX_RAWTEXT_CHANGING_FLAGS,
29 29 )
30 30
31 31 from ..thirdparty import attr
32 32
33 33 from .. import (
34 34 error,
35 35 mdiff,
36 36 util,
37 37 )
38 38
39 39 from . import flagutil
40 40
41 41 # maximum <delta-chain-data>/<revision-text-length> ratio
42 42 LIMIT_DELTA2TEXT = 2
43 43
44 44
45 45 class _testrevlog:
46 46 """minimalist fake revlog to use in doctests"""
47 47
48 48 def __init__(self, data, density=0.5, mingap=0, snapshot=()):
49 49 """data is an list of revision payload boundaries"""
50 50 from .. import revlog
51 51
52 52 self._data = data
53 53 self.data_config = revlog.DataConfig()
54 54 self.data_config.sr_density_threshold = density
55 55 self.data_config.sr_min_gap_size = mingap
56 56 self.delta_config = revlog.DeltaConfig()
57 57 self.feature_config = revlog.FeatureConfig()
58 58 self._snapshot = set(snapshot)
59 59 self.index = None
60 60
61 61 def start(self, rev):
62 62 if rev == nullrev:
63 63 return 0
64 64 if rev == 0:
65 65 return 0
66 66 return self._data[rev - 1]
67 67
68 68 def end(self, rev):
69 69 if rev == nullrev:
70 70 return 0
71 71 return self._data[rev]
72 72
73 73 def length(self, rev):
74 74 return self.end(rev) - self.start(rev)
75 75
76 76 def __len__(self):
77 77 return len(self._data)
78 78
79 79 def issnapshot(self, rev):
80 80 if rev == nullrev:
81 81 return True
82 82 return rev in self._snapshot
83 83
84 84
85 85 def slicechunk(revlog, revs, targetsize=None):
86 86 """slice revs to reduce the amount of unrelated data to be read from disk.
87 87
88 88 ``revs`` is sliced into groups that should be read in one time.
89 89 Assume that revs are sorted.
90 90
91 91 The initial chunk is sliced until the overall density (payload/chunks-span
92 92 ratio) is above `revlog.data_config.sr_density_threshold`. No gap smaller
93 93 than `revlog.data_config.sr_min_gap_size` is skipped.
94 94
95 95 If `targetsize` is set, no chunk larger than `targetsize` will be yield.
96 96 For consistency with other slicing choice, this limit won't go lower than
97 97 `revlog.data_config.sr_min_gap_size`.
98 98
99 99 If individual revisions chunk are larger than this limit, they will still
100 100 be raised individually.
101 101
102 102 >>> data = [
103 103 ... 5, #00 (5)
104 104 ... 10, #01 (5)
105 105 ... 12, #02 (2)
106 106 ... 12, #03 (empty)
107 107 ... 27, #04 (15)
108 108 ... 31, #05 (4)
109 109 ... 31, #06 (empty)
110 110 ... 42, #07 (11)
111 111 ... 47, #08 (5)
112 112 ... 47, #09 (empty)
113 113 ... 48, #10 (1)
114 114 ... 51, #11 (3)
115 115 ... 74, #12 (23)
116 116 ... 85, #13 (11)
117 117 ... 86, #14 (1)
118 118 ... 91, #15 (5)
119 119 ... ]
120 120 >>> revlog = _testrevlog(data, snapshot=range(16))
121 121
122 122 >>> list(slicechunk(revlog, list(range(16))))
123 123 [[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15]]
124 124 >>> list(slicechunk(revlog, [0, 15]))
125 125 [[0], [15]]
126 126 >>> list(slicechunk(revlog, [0, 11, 15]))
127 127 [[0], [11], [15]]
128 128 >>> list(slicechunk(revlog, [0, 11, 13, 15]))
129 129 [[0], [11, 13, 15]]
130 130 >>> list(slicechunk(revlog, [1, 2, 3, 5, 8, 10, 11, 14]))
131 131 [[1, 2], [5, 8, 10, 11], [14]]
132 132
133 133 Slicing with a maximum chunk size
134 134 >>> list(slicechunk(revlog, [0, 11, 13, 15], targetsize=15))
135 135 [[0], [11], [13], [15]]
136 136 >>> list(slicechunk(revlog, [0, 11, 13, 15], targetsize=20))
137 137 [[0], [11], [13, 15]]
138 138
139 139 Slicing involving nullrev
140 140 >>> list(slicechunk(revlog, [-1, 0, 11, 13, 15], targetsize=20))
141 141 [[-1, 0], [11], [13, 15]]
142 142 >>> list(slicechunk(revlog, [-1, 13, 15], targetsize=5))
143 143 [[-1], [13], [15]]
144 144 """
145 145 if targetsize is not None:
146 146 targetsize = max(targetsize, revlog.data_config.sr_min_gap_size)
147 147 # targetsize should not be specified when evaluating delta candidates:
148 148 # * targetsize is used to ensure we stay within specification when reading,
149 149 densityslicing = getattr(revlog.index, 'slicechunktodensity', None)
150 150 if densityslicing is None:
151 151 densityslicing = lambda x, y, z: _slicechunktodensity(revlog, x, y, z)
152 152 for chunk in densityslicing(
153 153 revs,
154 154 revlog.data_config.sr_density_threshold,
155 155 revlog.data_config.sr_min_gap_size,
156 156 ):
157 157 for subchunk in _slicechunktosize(revlog, chunk, targetsize):
158 158 yield subchunk
159 159
160 160
161 161 def _slicechunktosize(revlog, revs, targetsize=None):
162 162 """slice revs to match the target size
163 163
164 164 This is intended to be used on chunk that density slicing selected by that
165 165 are still too large compared to the read garantee of revlog. This might
166 166 happens when "minimal gap size" interrupted the slicing or when chain are
167 167 built in a way that create large blocks next to each other.
168 168
169 169 >>> data = [
170 170 ... 3, #0 (3)
171 171 ... 5, #1 (2)
172 172 ... 6, #2 (1)
173 173 ... 8, #3 (2)
174 174 ... 8, #4 (empty)
175 175 ... 11, #5 (3)
176 176 ... 12, #6 (1)
177 177 ... 13, #7 (1)
178 178 ... 14, #8 (1)
179 179 ... ]
180 180
181 181 == All snapshots cases ==
182 182 >>> revlog = _testrevlog(data, snapshot=range(9))
183 183
184 184 Cases where chunk is already small enough
185 185 >>> list(_slicechunktosize(revlog, [0], 3))
186 186 [[0]]
187 187 >>> list(_slicechunktosize(revlog, [6, 7], 3))
188 188 [[6, 7]]
189 189 >>> list(_slicechunktosize(revlog, [0], None))
190 190 [[0]]
191 191 >>> list(_slicechunktosize(revlog, [6, 7], None))
192 192 [[6, 7]]
193 193
194 194 cases where we need actual slicing
195 195 >>> list(_slicechunktosize(revlog, [0, 1], 3))
196 196 [[0], [1]]
197 197 >>> list(_slicechunktosize(revlog, [1, 3], 3))
198 198 [[1], [3]]
199 199 >>> list(_slicechunktosize(revlog, [1, 2, 3], 3))
200 200 [[1, 2], [3]]
201 201 >>> list(_slicechunktosize(revlog, [3, 5], 3))
202 202 [[3], [5]]
203 203 >>> list(_slicechunktosize(revlog, [3, 4, 5], 3))
204 204 [[3], [5]]
205 205 >>> list(_slicechunktosize(revlog, [5, 6, 7, 8], 3))
206 206 [[5], [6, 7, 8]]
207 207 >>> list(_slicechunktosize(revlog, [0, 1, 2, 3, 4, 5, 6, 7, 8], 3))
208 208 [[0], [1, 2], [3], [5], [6, 7, 8]]
209 209
210 210 Case with too large individual chunk (must return valid chunk)
211 211 >>> list(_slicechunktosize(revlog, [0, 1], 2))
212 212 [[0], [1]]
213 213 >>> list(_slicechunktosize(revlog, [1, 3], 1))
214 214 [[1], [3]]
215 215 >>> list(_slicechunktosize(revlog, [3, 4, 5], 2))
216 216 [[3], [5]]
217 217
218 218 == No Snapshot cases ==
219 219 >>> revlog = _testrevlog(data)
220 220
221 221 Cases where chunk is already small enough
222 222 >>> list(_slicechunktosize(revlog, [0], 3))
223 223 [[0]]
224 224 >>> list(_slicechunktosize(revlog, [6, 7], 3))
225 225 [[6, 7]]
226 226 >>> list(_slicechunktosize(revlog, [0], None))
227 227 [[0]]
228 228 >>> list(_slicechunktosize(revlog, [6, 7], None))
229 229 [[6, 7]]
230 230
231 231 cases where we need actual slicing
232 232 >>> list(_slicechunktosize(revlog, [0, 1], 3))
233 233 [[0], [1]]
234 234 >>> list(_slicechunktosize(revlog, [1, 3], 3))
235 235 [[1], [3]]
236 236 >>> list(_slicechunktosize(revlog, [1, 2, 3], 3))
237 237 [[1], [2, 3]]
238 238 >>> list(_slicechunktosize(revlog, [3, 5], 3))
239 239 [[3], [5]]
240 240 >>> list(_slicechunktosize(revlog, [3, 4, 5], 3))
241 241 [[3], [4, 5]]
242 242 >>> list(_slicechunktosize(revlog, [5, 6, 7, 8], 3))
243 243 [[5], [6, 7, 8]]
244 244 >>> list(_slicechunktosize(revlog, [0, 1, 2, 3, 4, 5, 6, 7, 8], 3))
245 245 [[0], [1, 2], [3], [5], [6, 7, 8]]
246 246
247 247 Case with too large individual chunk (must return valid chunk)
248 248 >>> list(_slicechunktosize(revlog, [0, 1], 2))
249 249 [[0], [1]]
250 250 >>> list(_slicechunktosize(revlog, [1, 3], 1))
251 251 [[1], [3]]
252 252 >>> list(_slicechunktosize(revlog, [3, 4, 5], 2))
253 253 [[3], [5]]
254 254
255 255 == mixed case ==
256 256 >>> revlog = _testrevlog(data, snapshot=[0, 1, 2])
257 257 >>> list(_slicechunktosize(revlog, list(range(9)), 5))
258 258 [[0, 1], [2], [3, 4, 5], [6, 7, 8]]
259 259 """
260 260 assert targetsize is None or 0 <= targetsize
261 261 startdata = revlog.start(revs[0])
262 262 enddata = revlog.end(revs[-1])
263 263 fullspan = enddata - startdata
264 264 if targetsize is None or fullspan <= targetsize:
265 265 yield revs
266 266 return
267 267
268 268 startrevidx = 0
269 269 endrevidx = 1
270 270 iterrevs = enumerate(revs)
271 271 next(iterrevs) # skip first rev.
272 272 # first step: get snapshots out of the way
273 273 for idx, r in iterrevs:
274 274 span = revlog.end(r) - startdata
275 275 snapshot = revlog.issnapshot(r)
276 276 if span <= targetsize and snapshot:
277 277 endrevidx = idx + 1
278 278 else:
279 279 chunk = _trimchunk(revlog, revs, startrevidx, endrevidx)
280 280 if chunk:
281 281 yield chunk
282 282 startrevidx = idx
283 283 startdata = revlog.start(r)
284 284 endrevidx = idx + 1
285 285 if not snapshot:
286 286 break
287 287
288 288 # for the others, we use binary slicing to quickly converge toward valid
289 289 # chunks (otherwise, we might end up looking for start/end of many
290 290 # revisions). This logic is not looking for the perfect slicing point, it
291 291 # focuses on quickly converging toward valid chunks.
292 292 nbitem = len(revs)
293 293 while (enddata - startdata) > targetsize:
294 294 endrevidx = nbitem
295 295 if nbitem - startrevidx <= 1:
296 296 break # protect against individual chunk larger than limit
297 297 localenddata = revlog.end(revs[endrevidx - 1])
298 298 span = localenddata - startdata
299 299 while span > targetsize:
300 300 if endrevidx - startrevidx <= 1:
301 301 break # protect against individual chunk larger than limit
302 302 endrevidx -= (endrevidx - startrevidx) // 2
303 303 localenddata = revlog.end(revs[endrevidx - 1])
304 304 span = localenddata - startdata
305 305 chunk = _trimchunk(revlog, revs, startrevidx, endrevidx)
306 306 if chunk:
307 307 yield chunk
308 308 startrevidx = endrevidx
309 309 startdata = revlog.start(revs[startrevidx])
310 310
311 311 chunk = _trimchunk(revlog, revs, startrevidx)
312 312 if chunk:
313 313 yield chunk
314 314
315 315
316 316 def _slicechunktodensity(revlog, revs, targetdensity=0.5, mingapsize=0):
317 317 """slice revs to reduce the amount of unrelated data to be read from disk.
318 318
319 319 ``revs`` is sliced into groups that should be read in one time.
320 320 Assume that revs are sorted.
321 321
322 322 The initial chunk is sliced until the overall density (payload/chunks-span
323 323 ratio) is above `targetdensity`. No gap smaller than `mingapsize` is
324 324 skipped.
325 325
326 326 >>> revlog = _testrevlog([
327 327 ... 5, #00 (5)
328 328 ... 10, #01 (5)
329 329 ... 12, #02 (2)
330 330 ... 12, #03 (empty)
331 331 ... 27, #04 (15)
332 332 ... 31, #05 (4)
333 333 ... 31, #06 (empty)
334 334 ... 42, #07 (11)
335 335 ... 47, #08 (5)
336 336 ... 47, #09 (empty)
337 337 ... 48, #10 (1)
338 338 ... 51, #11 (3)
339 339 ... 74, #12 (23)
340 340 ... 85, #13 (11)
341 341 ... 86, #14 (1)
342 342 ... 91, #15 (5)
343 343 ... ])
344 344
345 345 >>> list(_slicechunktodensity(revlog, list(range(16))))
346 346 [[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15]]
347 347 >>> list(_slicechunktodensity(revlog, [0, 15]))
348 348 [[0], [15]]
349 349 >>> list(_slicechunktodensity(revlog, [0, 11, 15]))
350 350 [[0], [11], [15]]
351 351 >>> list(_slicechunktodensity(revlog, [0, 11, 13, 15]))
352 352 [[0], [11, 13, 15]]
353 353 >>> list(_slicechunktodensity(revlog, [1, 2, 3, 5, 8, 10, 11, 14]))
354 354 [[1, 2], [5, 8, 10, 11], [14]]
355 355 >>> list(_slicechunktodensity(revlog, [1, 2, 3, 5, 8, 10, 11, 14],
356 356 ... mingapsize=20))
357 357 [[1, 2, 3, 5, 8, 10, 11], [14]]
358 358 >>> list(_slicechunktodensity(revlog, [1, 2, 3, 5, 8, 10, 11, 14],
359 359 ... targetdensity=0.95))
360 360 [[1, 2], [5], [8, 10, 11], [14]]
361 361 >>> list(_slicechunktodensity(revlog, [1, 2, 3, 5, 8, 10, 11, 14],
362 362 ... targetdensity=0.95, mingapsize=12))
363 363 [[1, 2], [5, 8, 10, 11], [14]]
364 364 """
365 365 start = revlog.start
366 366 length = revlog.length
367 367
368 368 if len(revs) <= 1:
369 369 yield revs
370 370 return
371 371
372 372 deltachainspan = segmentspan(revlog, revs)
373 373
374 374 if deltachainspan < mingapsize:
375 375 yield revs
376 376 return
377 377
378 378 readdata = deltachainspan
379 379 chainpayload = sum(length(r) for r in revs)
380 380
381 381 if deltachainspan:
382 382 density = chainpayload / float(deltachainspan)
383 383 else:
384 384 density = 1.0
385 385
386 386 if density >= targetdensity:
387 387 yield revs
388 388 return
389 389
390 390 # Store the gaps in a heap to have them sorted by decreasing size
391 391 gaps = []
392 392 prevend = None
393 393 for i, rev in enumerate(revs):
394 394 revstart = start(rev)
395 395 revlen = length(rev)
396 396
397 397 # Skip empty revisions to form larger holes
398 398 if revlen == 0:
399 399 continue
400 400
401 401 if prevend is not None:
402 402 gapsize = revstart - prevend
403 403 # only consider holes that are large enough
404 404 if gapsize > mingapsize:
405 405 gaps.append((gapsize, i))
406 406
407 407 prevend = revstart + revlen
408 408 # sort the gaps to pop them from largest to small
409 409 gaps.sort()
410 410
411 411 # Collect the indices of the largest holes until the density is acceptable
412 412 selected = []
413 413 while gaps and density < targetdensity:
414 414 gapsize, gapidx = gaps.pop()
415 415
416 416 selected.append(gapidx)
417 417
418 418 # the gap sizes are stored as negatives to be sorted decreasingly
419 419 # by the heap
420 420 readdata -= gapsize
421 421 if readdata > 0:
422 422 density = chainpayload / float(readdata)
423 423 else:
424 424 density = 1.0
425 425 selected.sort()
426 426
427 427 # Cut the revs at collected indices
428 428 previdx = 0
429 429 for idx in selected:
430 430
431 431 chunk = _trimchunk(revlog, revs, previdx, idx)
432 432 if chunk:
433 433 yield chunk
434 434
435 435 previdx = idx
436 436
437 437 chunk = _trimchunk(revlog, revs, previdx)
438 438 if chunk:
439 439 yield chunk
440 440
441 441
442 442 def _trimchunk(revlog, revs, startidx, endidx=None):
443 443 """returns revs[startidx:endidx] without empty trailing revs
444 444
445 445 Doctest Setup
446 446 >>> revlog = _testrevlog([
447 447 ... 5, #0
448 448 ... 10, #1
449 449 ... 12, #2
450 450 ... 12, #3 (empty)
451 451 ... 17, #4
452 452 ... 21, #5
453 453 ... 21, #6 (empty)
454 454 ... ])
455 455
456 456 Contiguous cases:
457 457 >>> _trimchunk(revlog, [0, 1, 2, 3, 4, 5, 6], 0)
458 458 [0, 1, 2, 3, 4, 5]
459 459 >>> _trimchunk(revlog, [0, 1, 2, 3, 4, 5, 6], 0, 5)
460 460 [0, 1, 2, 3, 4]
461 461 >>> _trimchunk(revlog, [0, 1, 2, 3, 4, 5, 6], 0, 4)
462 462 [0, 1, 2]
463 463 >>> _trimchunk(revlog, [0, 1, 2, 3, 4, 5, 6], 2, 4)
464 464 [2]
465 465 >>> _trimchunk(revlog, [0, 1, 2, 3, 4, 5, 6], 3)
466 466 [3, 4, 5]
467 467 >>> _trimchunk(revlog, [0, 1, 2, 3, 4, 5, 6], 3, 5)
468 468 [3, 4]
469 469
470 470 Discontiguous cases:
471 471 >>> _trimchunk(revlog, [1, 3, 5, 6], 0)
472 472 [1, 3, 5]
473 473 >>> _trimchunk(revlog, [1, 3, 5, 6], 0, 2)
474 474 [1]
475 475 >>> _trimchunk(revlog, [1, 3, 5, 6], 1, 3)
476 476 [3, 5]
477 477 >>> _trimchunk(revlog, [1, 3, 5, 6], 1)
478 478 [3, 5]
479 479 """
480 480 length = revlog.length
481 481
482 482 if endidx is None:
483 483 endidx = len(revs)
484 484
485 485 # If we have a non-emtpy delta candidate, there are nothing to trim
486 486 if revs[endidx - 1] < len(revlog):
487 487 # Trim empty revs at the end, except the very first revision of a chain
488 488 while (
489 489 endidx > 1 and endidx > startidx and length(revs[endidx - 1]) == 0
490 490 ):
491 491 endidx -= 1
492 492
493 493 return revs[startidx:endidx]
494 494
495 495
496 496 def segmentspan(revlog, revs):
497 497 """Get the byte span of a segment of revisions
498 498
499 499 revs is a sorted array of revision numbers
500 500
501 501 >>> revlog = _testrevlog([
502 502 ... 5, #0
503 503 ... 10, #1
504 504 ... 12, #2
505 505 ... 12, #3 (empty)
506 506 ... 17, #4
507 507 ... ])
508 508
509 509 >>> segmentspan(revlog, [0, 1, 2, 3, 4])
510 510 17
511 511 >>> segmentspan(revlog, [0, 4])
512 512 17
513 513 >>> segmentspan(revlog, [3, 4])
514 514 5
515 515 >>> segmentspan(revlog, [1, 2, 3,])
516 516 7
517 517 >>> segmentspan(revlog, [1, 3])
518 518 7
519 519 """
520 520 if not revs:
521 521 return 0
522 522 end = revlog.end(revs[-1])
523 523 return end - revlog.start(revs[0])
524 524
525 525
526 526 def _textfromdelta(revlog, baserev, delta, p1, p2, flags, expectednode):
527 527 """build full text from a (base, delta) pair and other metadata"""
528 528 # special case deltas which replace entire base; no need to decode
529 529 # base revision. this neatly avoids censored bases, which throw when
530 530 # they're decoded.
531 531 hlen = struct.calcsize(b">lll")
532 532 if delta[:hlen] == mdiff.replacediffheader(
533 533 revlog.rawsize(baserev), len(delta) - hlen
534 534 ):
535 535 fulltext = delta[hlen:]
536 536 else:
537 537 # deltabase is rawtext before changed by flag processors, which is
538 538 # equivalent to non-raw text
539 539 basetext = revlog.revision(baserev)
540 540 fulltext = mdiff.patch(basetext, delta)
541 541
542 542 try:
543 543 validatehash = flagutil.processflagsraw(revlog, fulltext, flags)
544 544 if validatehash:
545 545 revlog.checkhash(fulltext, expectednode, p1=p1, p2=p2)
546 546 if flags & REVIDX_ISCENSORED:
547 547 raise error.StorageError(
548 548 _(b'node %s is not censored') % expectednode
549 549 )
550 550 except error.CensoredNodeError:
551 551 # must pass the censored index flag to add censored revisions
552 552 if not flags & REVIDX_ISCENSORED:
553 553 raise
554 554 return fulltext
555 555
556 556
557 557 @attr.s(slots=True, frozen=True)
558 558 class _deltainfo:
559 559 distance = attr.ib()
560 560 deltalen = attr.ib()
561 561 data = attr.ib()
562 562 base = attr.ib()
563 563 chainbase = attr.ib()
564 564 chainlen = attr.ib()
565 565 compresseddeltalen = attr.ib()
566 566 snapshotdepth = attr.ib()
567 567
568 568
569 569 def drop_u_compression(delta):
570 570 """turn into a "u" (no-compression) into no-compression without header
571 571
572 572 This is useful for revlog format that has better compression method.
573 573 """
574 574 assert delta.data[0] == b'u', delta.data[0]
575 575 return _deltainfo(
576 576 delta.distance,
577 577 delta.deltalen - 1,
578 578 (b'', delta.data[1]),
579 579 delta.base,
580 580 delta.chainbase,
581 581 delta.chainlen,
582 582 delta.compresseddeltalen,
583 583 delta.snapshotdepth,
584 584 )
585 585
586 586
587 587 def is_good_delta_info(revlog, deltainfo, revinfo):
588 588 """Returns True if the given delta is good. Good means that it is within
589 589 the disk span, disk size, and chain length bounds that we know to be
590 590 performant."""
591 591 if deltainfo is None:
592 592 return False
593 593
594 594 # the DELTA_BASE_REUSE_FORCE case should have been taken care of sooner so
595 595 # we should never end up asking such question. Adding the assert as a
596 596 # safe-guard to detect anything that would be fishy in this regard.
597 597 assert (
598 598 revinfo.cachedelta is None
599 599 or revinfo.cachedelta[2] != DELTA_BASE_REUSE_FORCE
600 600 or not revlog.delta_config.general_delta
601 601 )
602 602
603 603 # - 'deltainfo.distance' is the distance from the base revision --
604 604 # bounding it limits the amount of I/O we need to do.
605 605 # - 'deltainfo.compresseddeltalen' is the sum of the total size of
606 606 # deltas we need to apply -- bounding it limits the amount of CPU
607 607 # we consume.
608 608
609 609 textlen = revinfo.textlen
610 610 defaultmax = textlen * 4
611 611 maxdist = revlog.delta_config.max_deltachain_span
612 612 if not maxdist:
613 613 maxdist = deltainfo.distance # ensure the conditional pass
614 614 maxdist = max(maxdist, defaultmax)
615 615
616 616 # Bad delta from read span:
617 617 #
618 618 # If the span of data read is larger than the maximum allowed.
619 619 #
620 620 # In the sparse-revlog case, we rely on the associated "sparse reading"
621 621 # to avoid issue related to the span of data. In theory, it would be
622 622 # possible to build pathological revlog where delta pattern would lead
623 623 # to too many reads. However, they do not happen in practice at all. So
624 624 # we skip the span check entirely.
625 625 if not revlog.delta_config.sparse_revlog and maxdist < deltainfo.distance:
626 626 return False
627 627
628 628 # Bad delta from new delta size:
629 629 #
630 630 # If the delta size is larger than the target text, storing the
631 631 # delta will be inefficient.
632 632 if textlen < deltainfo.deltalen:
633 633 return False
634 634
635 635 # Bad delta from cumulated payload size:
636 636 #
637 637 # If the sum of delta get larger than K * target text length.
638 638 if textlen * LIMIT_DELTA2TEXT < deltainfo.compresseddeltalen:
639 639 return False
640 640
641 641 # Bad delta from chain length:
642 642 #
643 643 # If the number of delta in the chain gets too high.
644 644 if (
645 645 revlog.delta_config.max_chain_len
646 646 and revlog.delta_config.max_chain_len < deltainfo.chainlen
647 647 ):
648 648 return False
649 649
650 650 # bad delta from intermediate snapshot size limit
651 651 #
652 652 # If an intermediate snapshot size is higher than the limit. The
653 653 # limit exist to prevent endless chain of intermediate delta to be
654 654 # created.
655 655 if (
656 656 deltainfo.snapshotdepth is not None
657 657 and (textlen >> deltainfo.snapshotdepth) < deltainfo.deltalen
658 658 ):
659 659 return False
660 660
661 661 # bad delta if new intermediate snapshot is larger than the previous
662 662 # snapshot
663 663 if (
664 664 deltainfo.snapshotdepth
665 665 and revlog.length(deltainfo.base) < deltainfo.deltalen
666 666 ):
667 667 return False
668 668
669 669 return True
670 670
671 671
672 672 # If a revision's full text is that much bigger than a base candidate full
673 673 # text's, it is very unlikely that it will produce a valid delta. We no longer
674 674 # consider these candidates.
675 675 LIMIT_BASE2TEXT = 500
676 676
677 677
678 678 def _candidategroups(
679 679 revlog,
680 680 textlen,
681 681 p1,
682 682 p2,
683 683 cachedelta,
684 684 excluded_bases=None,
685 685 target_rev=None,
686 686 snapshot_cache=None,
687 687 ):
688 688 """Provides group of revision to be tested as delta base
689 689
690 690 This top level function focus on emitting groups with unique and worthwhile
691 691 content. See _raw_candidate_groups for details about the group order.
692 692 """
693 693 # should we try to build a delta?
694 694 if not (len(revlog) and revlog._storedeltachains):
695 695 yield None
696 696 return
697 697
698 698 if target_rev is None:
699 699 target_rev = len(revlog)
700 700
701 701 if not revlog.delta_config.general_delta:
702 702 # before general delta, there is only one possible delta base
703 703 yield (target_rev - 1,)
704 704 yield None
705 705 return
706 706
707 707 # the DELTA_BASE_REUSE_FORCE case should have been taken care of sooner so
708 708 # we should never end up asking such question. Adding the assert as a
709 709 # safe-guard to detect anything that would be fishy in this regard.
710 710 assert (
711 711 cachedelta is None
712 712 or cachedelta[2] != DELTA_BASE_REUSE_FORCE
713 713 or not revlog.delta_config.general_delta
714 714 )
715 715
716 716 deltalength = revlog.length
717 717 deltaparent = revlog.deltaparent
718 718 sparse = revlog.delta_config.sparse_revlog
719 719 good = None
720 720
721 721 deltas_limit = textlen * LIMIT_DELTA2TEXT
722 722 group_chunk_size = revlog.delta_config.candidate_group_chunk_size
723 723
724 724 tested = {nullrev}
725 725 candidates = _refinedgroups(
726 726 revlog,
727 727 p1,
728 728 p2,
729 729 cachedelta,
730 730 snapshot_cache=snapshot_cache,
731 731 )
732 732 while True:
733 733 temptative = candidates.send(good)
734 734 if temptative is None:
735 735 break
736 736 group = []
737 737 for rev in temptative:
738 738 # skip over empty delta (no need to include them in a chain)
739 739 while not (rev == nullrev or rev in tested or deltalength(rev)):
740 740 tested.add(rev)
741 741 rev = deltaparent(rev)
742 742 # no need to try a delta against nullrev, this will be done as a
743 743 # last resort.
744 744 if rev == nullrev:
745 745 continue
746 746 # filter out revision we tested already
747 747 if rev in tested:
748 748 continue
749 749
750 750 # an higher authority deamed the base unworthy (e.g. censored)
751 751 if excluded_bases is not None and rev in excluded_bases:
752 752 tested.add(rev)
753 753 continue
754 754 # We are in some recomputation cases and that rev is too high in
755 755 # the revlog
756 756 if target_rev is not None and rev >= target_rev:
757 757 tested.add(rev)
758 758 continue
759 759 # filter out delta base that will never produce good delta
760 760 if deltas_limit < revlog.length(rev):
761 761 tested.add(rev)
762 762 continue
763 763 if sparse and revlog.rawsize(rev) < (textlen // LIMIT_BASE2TEXT):
764 764 tested.add(rev)
765 765 continue
766 766 # no delta for rawtext-changing revs (see "candelta" for why)
767 767 if revlog.flags(rev) & REVIDX_RAWTEXT_CHANGING_FLAGS:
768 768 tested.add(rev)
769 769 continue
770 770
771 771 # If we reach here, we are about to build and test a delta.
772 772 # The delta building process will compute the chaininfo in all
773 773 # case, since that computation is cached, it is fine to access it
774 774 # here too.
775 775 chainlen, chainsize = revlog._chaininfo(rev)
776 776 # if chain will be too long, skip base
777 777 if (
778 778 revlog.delta_config.max_chain_len
779 779 and chainlen >= revlog.delta_config.max_chain_len
780 780 ):
781 781 tested.add(rev)
782 782 continue
783 783 # if chain already have too much data, skip base
784 784 if deltas_limit < chainsize:
785 785 tested.add(rev)
786 786 continue
787 787 if sparse and revlog.upperboundcomp is not None:
788 788 maxcomp = revlog.upperboundcomp
789 789 basenotsnap = (p1, p2, nullrev)
790 790 if rev not in basenotsnap and revlog.issnapshot(rev):
791 791 snapshotdepth = revlog.snapshotdepth(rev)
792 792 # If text is significantly larger than the base, we can
793 793 # expect the resulting delta to be proportional to the size
794 794 # difference
795 795 revsize = revlog.rawsize(rev)
796 796 rawsizedistance = max(textlen - revsize, 0)
797 797 # use an estimate of the compression upper bound.
798 798 lowestrealisticdeltalen = rawsizedistance // maxcomp
799 799
800 800 # check the absolute constraint on the delta size
801 801 snapshotlimit = textlen >> snapshotdepth
802 802 if snapshotlimit < lowestrealisticdeltalen:
803 803 # delta lower bound is larger than accepted upper bound
804 804 tested.add(rev)
805 805 continue
806 806
807 807 # check the relative constraint on the delta size
808 808 revlength = revlog.length(rev)
809 809 if revlength < lowestrealisticdeltalen:
810 810 # delta probable lower bound is larger than target base
811 811 tested.add(rev)
812 812 continue
813 813
814 814 group.append(rev)
815 815 if group:
816 816 # When the size of the candidate group is big, it can result in a
817 817 # quite significant performance impact. To reduce this, we can send
818 818 # them in smaller batches until the new batch does not provide any
819 819 # improvements.
820 820 #
821 821 # This might reduce the overall efficiency of the compression in
822 822 # some corner cases, but that should also prevent very pathological
823 823 # cases from being an issue. (eg. 20 000 candidates).
824 824 #
825 825 # XXX note that the ordering of the group becomes important as it
826 826 # now impacts the final result. The current order is unprocessed
827 827 # and can be improved.
828 828 if group_chunk_size == 0:
829 829 tested.update(group)
830 830 good = yield tuple(group)
831 831 else:
832 832 prev_good = good
833 833 for start in range(0, len(group), group_chunk_size):
834 834 sub_group = group[start : start + group_chunk_size]
835 835 tested.update(sub_group)
836 836 good = yield tuple(sub_group)
837 837 if prev_good == good:
838 838 break
839 839
840 840 yield None
841 841
842 842
843 843 def _refinedgroups(revlog, p1, p2, cachedelta, snapshot_cache=None):
844 844 good = None
845 845 # First we try to reuse a the delta contained in the bundle.
846 846 # (or from the source revlog)
847 847 #
848 848 # This logic only applies to general delta repositories and can be disabled
849 849 # through configuration. Disabling reuse source delta is useful when
850 850 # we want to make sure we recomputed "optimal" deltas.
851 851 debug_info = None
852 852 if cachedelta is not None and cachedelta[2] > DELTA_BASE_REUSE_NO:
853 853 # Assume what we received from the server is a good choice
854 854 # build delta will reuse the cache
855 855 if debug_info is not None:
856 856 debug_info['cached-delta.tested'] += 1
857 857 good = yield (cachedelta[0],)
858 858 if good is not None:
859 859 if debug_info is not None:
860 860 debug_info['cached-delta.accepted'] += 1
861 861 yield None
862 862 return
863 863 if snapshot_cache is None:
864 864 snapshot_cache = SnapshotCache()
865 865 groups = _rawgroups(
866 866 revlog,
867 867 p1,
868 868 p2,
869 869 cachedelta,
870 870 snapshot_cache,
871 871 )
872 872 for candidates in groups:
873 873 good = yield candidates
874 874 if good is not None:
875 875 break
876 876
877 877 # If sparse revlog is enabled, we can try to refine the available deltas
878 878 if not revlog.delta_config.sparse_revlog:
879 879 yield None
880 880 return
881 881
882 882 # if we have a refinable value, try to refine it
883 883 if good is not None and good not in (p1, p2) and revlog.issnapshot(good):
884 884 # refine snapshot down
885 885 previous = None
886 886 while previous != good:
887 887 previous = good
888 888 base = revlog.deltaparent(good)
889 889 if base == nullrev:
890 890 break
891 891 good = yield (base,)
892 892 # refine snapshot up
893 893 if not snapshot_cache.snapshots:
894 894 snapshot_cache.update(revlog, good + 1)
895 895 previous = None
896 896 while good != previous:
897 897 previous = good
898 898 children = tuple(sorted(c for c in snapshot_cache.snapshots[good]))
899 899 good = yield children
900 900
901 901 if debug_info is not None:
902 902 if good is None:
903 903 debug_info['no-solution'] += 1
904 904
905 905 yield None
906 906
907 907
908 908 def _rawgroups(revlog, p1, p2, cachedelta, snapshot_cache=None):
909 909 """Provides group of revision to be tested as delta base
910 910
911 911 This lower level function focus on emitting delta theorically interresting
912 912 without looking it any practical details.
913 913
914 914 The group order aims at providing fast or small candidates first.
915 915 """
916 916 # Why search for delta base if we cannot use a delta base ?
917 917 assert revlog.delta_config.general_delta
918 918 # also see issue6056
919 919 sparse = revlog.delta_config.sparse_revlog
920 920 curr = len(revlog)
921 921 prev = curr - 1
922 922 deltachain = lambda rev: revlog._deltachain(rev)[0]
923 923
924 924 # exclude already lazy tested base if any
925 925 parents = [p for p in (p1, p2) if p != nullrev]
926 926
927 927 if not revlog.delta_config.delta_both_parents and len(parents) == 2:
928 928 parents.sort()
929 929 # To minimize the chance of having to build a fulltext,
930 930 # pick first whichever parent is closest to us (max rev)
931 931 yield (parents[1],)
932 932 # then the other one (min rev) if the first did not fit
933 933 yield (parents[0],)
934 934 elif len(parents) > 0:
935 935 # Test all parents (1 or 2), and keep the best candidate
936 936 yield parents
937 937
938 938 if sparse and parents:
939 939 if snapshot_cache is None:
940 940 # map: base-rev: [snapshot-revs]
941 941 snapshot_cache = SnapshotCache()
942 942 # See if we can use an existing snapshot in the parent chains to use as
943 943 # a base for a new intermediate-snapshot
944 944 #
945 945 # search for snapshot in parents delta chain
946 946 # map: snapshot-level: snapshot-rev
947 947 parents_snaps = collections.defaultdict(set)
948 948 candidate_chains = [deltachain(p) for p in parents]
949 949 for chain in candidate_chains:
950 950 for idx, s in enumerate(chain):
951 951 if not revlog.issnapshot(s):
952 952 break
953 953 parents_snaps[idx].add(s)
954 954 snapfloor = min(parents_snaps[0]) + 1
955 955 snapshot_cache.update(revlog, snapfloor)
956 956 # search for the highest "unrelated" revision
957 957 #
958 958 # Adding snapshots used by "unrelated" revision increase the odd we
959 959 # reuse an independant, yet better snapshot chain.
960 960 #
961 961 # XXX instead of building a set of revisions, we could lazily enumerate
962 962 # over the chains. That would be more efficient, however we stick to
963 963 # simple code for now.
964 964 all_revs = set()
965 965 for chain in candidate_chains:
966 966 all_revs.update(chain)
967 967 other = None
968 968 for r in revlog.revs(prev, snapfloor):
969 969 if r not in all_revs:
970 970 other = r
971 971 break
972 972 if other is not None:
973 973 # To avoid unfair competition, we won't use unrelated intermediate
974 974 # snapshot that are deeper than the ones from the parent delta
975 975 # chain.
976 976 max_depth = max(parents_snaps.keys())
977 977 chain = deltachain(other)
978 978 for depth, s in enumerate(chain):
979 979 if s < snapfloor:
980 980 continue
981 981 if max_depth < depth:
982 982 break
983 983 if not revlog.issnapshot(s):
984 984 break
985 985 parents_snaps[depth].add(s)
986 986 # Test them as possible intermediate snapshot base
987 987 # We test them from highest to lowest level. High level one are more
988 988 # likely to result in small delta
989 989 floor = None
990 990 for idx, snaps in sorted(parents_snaps.items(), reverse=True):
991 991 siblings = set()
992 992 for s in snaps:
993 993 siblings.update(snapshot_cache.snapshots[s])
994 994 # Before considering making a new intermediate snapshot, we check
995 995 # if an existing snapshot, children of base we consider, would be
996 996 # suitable.
997 997 #
998 998 # It give a change to reuse a delta chain "unrelated" to the
999 999 # current revision instead of starting our own. Without such
1000 1000 # re-use, topological branches would keep reopening new chains.
1001 1001 # Creating more and more snapshot as the repository grow.
1002 1002
1003 1003 if floor is not None:
1004 1004 # We only do this for siblings created after the one in our
1005 1005 # parent's delta chain. Those created before has less chances
1006 1006 # to be valid base since our ancestors had to create a new
1007 1007 # snapshot.
1008 1008 siblings = [r for r in siblings if floor < r]
1009 1009 yield tuple(sorted(siblings))
1010 1010 # then test the base from our parent's delta chain.
1011 1011 yield tuple(sorted(snaps))
1012 1012 floor = min(snaps)
1013 1013 # No suitable base found in the parent chain, search if any full
1014 1014 # snapshots emitted since parent's base would be a suitable base for an
1015 1015 # intermediate snapshot.
1016 1016 #
1017 1017 # It give a chance to reuse a delta chain unrelated to the current
1018 1018 # revisions instead of starting our own. Without such re-use,
1019 1019 # topological branches would keep reopening new full chains. Creating
1020 1020 # more and more snapshot as the repository grow.
1021 1021 full = [r for r in snapshot_cache.snapshots[nullrev] if snapfloor <= r]
1022 1022 yield tuple(sorted(full))
1023 1023
1024 1024 if not sparse:
1025 1025 # other approach failed try against prev to hopefully save us a
1026 1026 # fulltext.
1027 1027 yield (prev,)
1028 1028
1029 1029
1030 1030 class SnapshotCache:
1031 1031 __slots__ = ('snapshots', '_start_rev', '_end_rev')
1032 1032
1033 1033 def __init__(self):
1034 1034 self.snapshots = collections.defaultdict(set)
1035 1035 self._start_rev = None
1036 1036 self._end_rev = None
1037 1037
1038 1038 def update(self, revlog, start_rev=0):
1039 1039 """find snapshots from start_rev to tip"""
1040 1040 nb_revs = len(revlog)
1041 1041 end_rev = nb_revs - 1
1042 1042 if start_rev > end_rev:
1043 1043 return # range is empty
1044 1044
1045 1045 if self._start_rev is None:
1046 1046 assert self._end_rev is None
1047 1047 self._update(revlog, start_rev, end_rev)
1048 1048 elif not (self._start_rev <= start_rev and end_rev <= self._end_rev):
1049 1049 if start_rev < self._start_rev:
1050 1050 self._update(revlog, start_rev, self._start_rev - 1)
1051 1051 if self._end_rev < end_rev:
1052 1052 self._update(revlog, self._end_rev + 1, end_rev)
1053 1053
1054 1054 if self._start_rev is None:
1055 1055 assert self._end_rev is None
1056 1056 self._end_rev = end_rev
1057 1057 self._start_rev = start_rev
1058 1058 else:
1059 1059 self._start_rev = min(self._start_rev, start_rev)
1060 1060 self._end_rev = max(self._end_rev, end_rev)
1061 1061 assert self._start_rev <= self._end_rev, (
1062 1062 self._start_rev,
1063 1063 self._end_rev,
1064 1064 )
1065 1065
1066 1066 def _update(self, revlog, start_rev, end_rev):
1067 1067 """internal method that actually do update content"""
1068 1068 assert self._start_rev is None or (
1069 1069 start_rev < self._start_rev or start_rev > self._end_rev
1070 1070 ), (self._start_rev, self._end_rev, start_rev, end_rev)
1071 1071 assert self._start_rev is None or (
1072 1072 end_rev < self._start_rev or end_rev > self._end_rev
1073 1073 ), (self._start_rev, self._end_rev, start_rev, end_rev)
1074 1074 cache = self.snapshots
1075 1075 if hasattr(revlog.index, 'findsnapshots'):
1076 1076 revlog.index.findsnapshots(cache, start_rev, end_rev)
1077 1077 else:
1078 1078 deltaparent = revlog.deltaparent
1079 1079 issnapshot = revlog.issnapshot
1080 1080 for rev in revlog.revs(start_rev, end_rev):
1081 1081 if issnapshot(rev):
1082 1082 cache[deltaparent(rev)].add(rev)
1083 1083
1084 1084
1085 1085 class deltacomputer:
1086 1086 def __init__(
1087 1087 self,
1088 1088 revlog,
1089 1089 write_debug=None,
1090 1090 debug_search=False,
1091 1091 debug_info=None,
1092 1092 ):
1093 1093 self.revlog = revlog
1094 1094 self._write_debug = write_debug
1095 1095 if write_debug is None:
1096 1096 self._debug_search = False
1097 1097 else:
1098 1098 self._debug_search = debug_search
1099 1099 self._debug_info = debug_info
1100 1100 self._snapshot_cache = SnapshotCache()
1101 1101
1102 1102 @property
1103 1103 def _gather_debug(self):
1104 1104 return self._write_debug is not None or self._debug_info is not None
1105 1105
1106 1106 def buildtext(self, revinfo):
1107 1107 """Builds a fulltext version of a revision
1108 1108
1109 1109 revinfo: revisioninfo instance that contains all needed info
1110 1110 """
1111 1111 btext = revinfo.btext
1112 1112 if btext[0] is not None:
1113 1113 return btext[0]
1114 1114
1115 1115 revlog = self.revlog
1116 1116 cachedelta = revinfo.cachedelta
1117 1117 baserev = cachedelta[0]
1118 1118 delta = cachedelta[1]
1119 1119
1120 1120 fulltext = btext[0] = _textfromdelta(
1121 1121 revlog,
1122 1122 baserev,
1123 1123 delta,
1124 1124 revinfo.p1,
1125 1125 revinfo.p2,
1126 1126 revinfo.flags,
1127 1127 revinfo.node,
1128 1128 )
1129 1129 return fulltext
1130 1130
1131 1131 def _builddeltadiff(self, base, revinfo):
1132 1132 revlog = self.revlog
1133 1133 t = self.buildtext(revinfo)
1134 1134 if revlog.iscensored(base):
1135 1135 # deltas based on a censored revision must replace the
1136 1136 # full content in one patch, so delta works everywhere
1137 1137 header = mdiff.replacediffheader(revlog.rawsize(base), len(t))
1138 1138 delta = header + t
1139 1139 else:
1140 1140 ptext = revlog.rawdata(base)
1141 1141 delta = mdiff.textdiff(ptext, t)
1142 1142
1143 1143 return delta
1144 1144
1145 1145 def _builddeltainfo(self, revinfo, base, target_rev=None):
1146 1146 # can we use the cached delta?
1147 1147 revlog = self.revlog
1148 1148 chainbase = revlog.chainbase(base)
1149 1149 if revlog.delta_config.general_delta:
1150 1150 deltabase = base
1151 1151 else:
1152 1152 if target_rev is not None and base != target_rev - 1:
1153 1153 msg = (
1154 1154 b'general delta cannot use delta for something else '
1155 1155 b'than `prev`: %d<-%d'
1156 1156 )
1157 1157 msg %= (base, target_rev)
1158 1158 raise error.ProgrammingError(msg)
1159 1159 deltabase = chainbase
1160 1160 snapshotdepth = None
1161 1161 if revlog.delta_config.sparse_revlog and deltabase == nullrev:
1162 1162 snapshotdepth = 0
1163 1163 elif revlog.delta_config.sparse_revlog and revlog.issnapshot(deltabase):
1164 1164 # A delta chain should always be one full snapshot,
1165 1165 # zero or more semi-snapshots, and zero or more deltas
1166 1166 p1, p2 = revlog.rev(revinfo.p1), revlog.rev(revinfo.p2)
1167 1167 if deltabase not in (p1, p2) and revlog.issnapshot(deltabase):
1168 1168 snapshotdepth = len(revlog._deltachain(deltabase)[0])
1169 1169 delta = None
1170 1170 if revinfo.cachedelta:
1171 1171 cachebase = revinfo.cachedelta[0]
1172 1172 # check if the diff still apply
1173 1173 currentbase = cachebase
1174 1174 while (
1175 1175 currentbase != nullrev
1176 1176 and currentbase != base
1177 1177 and self.revlog.length(currentbase) == 0
1178 1178 ):
1179 1179 currentbase = self.revlog.deltaparent(currentbase)
1180 if self.revlog._lazydelta and currentbase == base:
1180 if self.revlog.delta_config.lazy_delta and currentbase == base:
1181 1181 delta = revinfo.cachedelta[1]
1182 1182 if delta is None:
1183 1183 delta = self._builddeltadiff(base, revinfo)
1184 1184 if self._debug_search:
1185 1185 msg = b"DBG-DELTAS-SEARCH: uncompressed-delta-size=%d\n"
1186 1186 msg %= len(delta)
1187 1187 self._write_debug(msg)
1188 1188 # snapshotdept need to be neither None nor 0 level snapshot
1189 1189 if revlog.upperboundcomp is not None and snapshotdepth:
1190 1190 lowestrealisticdeltalen = len(delta) // revlog.upperboundcomp
1191 1191 snapshotlimit = revinfo.textlen >> snapshotdepth
1192 1192 if self._debug_search:
1193 1193 msg = b"DBG-DELTAS-SEARCH: projected-lower-size=%d\n"
1194 1194 msg %= lowestrealisticdeltalen
1195 1195 self._write_debug(msg)
1196 1196 if snapshotlimit < lowestrealisticdeltalen:
1197 1197 if self._debug_search:
1198 1198 msg = b"DBG-DELTAS-SEARCH: DISCARDED (snapshot limit)\n"
1199 1199 self._write_debug(msg)
1200 1200 return None
1201 1201 if revlog.length(base) < lowestrealisticdeltalen:
1202 1202 if self._debug_search:
1203 1203 msg = b"DBG-DELTAS-SEARCH: DISCARDED (prev size)\n"
1204 1204 self._write_debug(msg)
1205 1205 return None
1206 1206 header, data = revlog.compress(delta)
1207 1207 deltalen = len(header) + len(data)
1208 1208 offset = revlog.end(len(revlog) - 1)
1209 1209 dist = deltalen + offset - revlog.start(chainbase)
1210 1210 chainlen, compresseddeltalen = revlog._chaininfo(base)
1211 1211 chainlen += 1
1212 1212 compresseddeltalen += deltalen
1213 1213
1214 1214 return _deltainfo(
1215 1215 dist,
1216 1216 deltalen,
1217 1217 (header, data),
1218 1218 deltabase,
1219 1219 chainbase,
1220 1220 chainlen,
1221 1221 compresseddeltalen,
1222 1222 snapshotdepth,
1223 1223 )
1224 1224
1225 1225 def _fullsnapshotinfo(self, revinfo, curr):
1226 1226 rawtext = self.buildtext(revinfo)
1227 1227 data = self.revlog.compress(rawtext)
1228 1228 compresseddeltalen = deltalen = dist = len(data[1]) + len(data[0])
1229 1229 deltabase = chainbase = curr
1230 1230 snapshotdepth = 0
1231 1231 chainlen = 1
1232 1232
1233 1233 return _deltainfo(
1234 1234 dist,
1235 1235 deltalen,
1236 1236 data,
1237 1237 deltabase,
1238 1238 chainbase,
1239 1239 chainlen,
1240 1240 compresseddeltalen,
1241 1241 snapshotdepth,
1242 1242 )
1243 1243
1244 1244 def finddeltainfo(self, revinfo, excluded_bases=None, target_rev=None):
1245 1245 """Find an acceptable delta against a candidate revision
1246 1246
1247 1247 revinfo: information about the revision (instance of _revisioninfo)
1248 1248
1249 1249 Returns the first acceptable candidate revision, as ordered by
1250 1250 _candidategroups
1251 1251
1252 1252 If no suitable deltabase is found, we return delta info for a full
1253 1253 snapshot.
1254 1254
1255 1255 `excluded_bases` is an optional set of revision that cannot be used as
1256 1256 a delta base. Use this to recompute delta suitable in censor or strip
1257 1257 context.
1258 1258 """
1259 1259 if target_rev is None:
1260 1260 target_rev = len(self.revlog)
1261 1261
1262 1262 gather_debug = self._gather_debug
1263 1263 cachedelta = revinfo.cachedelta
1264 1264 revlog = self.revlog
1265 1265 p1r = p2r = None
1266 1266
1267 1267 if excluded_bases is None:
1268 1268 excluded_bases = set()
1269 1269
1270 1270 if gather_debug:
1271 1271 start = util.timer()
1272 1272 dbg = self._one_dbg_data()
1273 1273 dbg['revision'] = target_rev
1274 1274 target_revlog = b"UNKNOWN"
1275 1275 target_type = self.revlog.target[0]
1276 1276 target_key = self.revlog.target[1]
1277 1277 if target_type == KIND_CHANGELOG:
1278 1278 target_revlog = b'CHANGELOG:'
1279 1279 elif target_type == KIND_MANIFESTLOG:
1280 1280 target_revlog = b'MANIFESTLOG:'
1281 1281 if target_key:
1282 1282 target_revlog += b'%s:' % target_key
1283 1283 elif target_type == KIND_FILELOG:
1284 1284 target_revlog = b'FILELOG:'
1285 1285 if target_key:
1286 1286 target_revlog += b'%s:' % target_key
1287 1287 dbg['target-revlog'] = target_revlog
1288 1288 p1r = revlog.rev(revinfo.p1)
1289 1289 p2r = revlog.rev(revinfo.p2)
1290 1290 if p1r != nullrev:
1291 1291 p1_chain_len = revlog._chaininfo(p1r)[0]
1292 1292 else:
1293 1293 p1_chain_len = -1
1294 1294 if p2r != nullrev:
1295 1295 p2_chain_len = revlog._chaininfo(p2r)[0]
1296 1296 else:
1297 1297 p2_chain_len = -1
1298 1298 dbg['p1-chain-len'] = p1_chain_len
1299 1299 dbg['p2-chain-len'] = p2_chain_len
1300 1300
1301 1301 # 1) if the revision is empty, no amount of delta can beat it
1302 1302 #
1303 1303 # 2) no delta for flag processor revision (see "candelta" for why)
1304 1304 # not calling candelta since only one revision needs test, also to
1305 1305 # avoid overhead fetching flags again.
1306 1306 if not revinfo.textlen or revinfo.flags & REVIDX_RAWTEXT_CHANGING_FLAGS:
1307 1307 deltainfo = self._fullsnapshotinfo(revinfo, target_rev)
1308 1308 if gather_debug:
1309 1309 end = util.timer()
1310 1310 dbg['duration'] = end - start
1311 1311 dbg[
1312 1312 'delta-base'
1313 1313 ] = deltainfo.base # pytype: disable=attribute-error
1314 1314 dbg['search_round_count'] = 0
1315 1315 dbg['using-cached-base'] = False
1316 1316 dbg['delta_try_count'] = 0
1317 1317 dbg['type'] = b"full"
1318 1318 dbg['snapshot-depth'] = 0
1319 1319 self._dbg_process_data(dbg)
1320 1320 return deltainfo
1321 1321
1322 1322 deltainfo = None
1323 1323
1324 1324 # If this source delta are to be forcibly reuse, let us comply early.
1325 1325 if (
1326 1326 revlog.delta_config.general_delta
1327 1327 and revinfo.cachedelta is not None
1328 1328 and revinfo.cachedelta[2] == DELTA_BASE_REUSE_FORCE
1329 1329 ):
1330 1330 base = revinfo.cachedelta[0]
1331 1331 if base == nullrev:
1332 1332 dbg_type = b"full"
1333 1333 deltainfo = self._fullsnapshotinfo(revinfo, target_rev)
1334 1334 if gather_debug:
1335 1335 snapshotdepth = 0
1336 1336 elif base not in excluded_bases:
1337 1337 delta = revinfo.cachedelta[1]
1338 1338 header, data = revlog.compress(delta)
1339 1339 deltalen = len(header) + len(data)
1340 1340 if gather_debug:
1341 1341 offset = revlog.end(len(revlog) - 1)
1342 1342 chainbase = revlog.chainbase(base)
1343 1343 distance = deltalen + offset - revlog.start(chainbase)
1344 1344 chainlen, compresseddeltalen = revlog._chaininfo(base)
1345 1345 chainlen += 1
1346 1346 compresseddeltalen += deltalen
1347 1347 if base == p1r or base == p2r:
1348 1348 dbg_type = b"delta"
1349 1349 snapshotdepth = None
1350 1350 elif not revlog.issnapshot(base):
1351 1351 snapshotdepth = None
1352 1352 else:
1353 1353 dbg_type = b"snapshot"
1354 1354 snapshotdepth = revlog.snapshotdepth(base) + 1
1355 1355 else:
1356 1356 distance = None
1357 1357 chainbase = None
1358 1358 chainlen = None
1359 1359 compresseddeltalen = None
1360 1360 snapshotdepth = None
1361 1361 deltainfo = _deltainfo(
1362 1362 distance=distance,
1363 1363 deltalen=deltalen,
1364 1364 data=(header, data),
1365 1365 base=base,
1366 1366 chainbase=chainbase,
1367 1367 chainlen=chainlen,
1368 1368 compresseddeltalen=compresseddeltalen,
1369 1369 snapshotdepth=snapshotdepth,
1370 1370 )
1371 1371
1372 1372 if deltainfo is not None:
1373 1373 if gather_debug:
1374 1374 end = util.timer()
1375 1375 dbg['duration'] = end - start
1376 1376 dbg[
1377 1377 'delta-base'
1378 1378 ] = deltainfo.base # pytype: disable=attribute-error
1379 1379 dbg['search_round_count'] = 0
1380 1380 dbg['using-cached-base'] = True
1381 1381 dbg['delta_try_count'] = 0
1382 1382 dbg['type'] = b"full"
1383 1383 if snapshotdepth is None:
1384 1384 dbg['snapshot-depth'] = 0
1385 1385 else:
1386 1386 dbg['snapshot-depth'] = snapshotdepth
1387 1387 self._dbg_process_data(dbg)
1388 1388 return deltainfo
1389 1389
1390 1390 # count the number of different delta we tried (for debug purpose)
1391 1391 dbg_try_count = 0
1392 1392 # count the number of "search round" we did. (for debug purpose)
1393 1393 dbg_try_rounds = 0
1394 1394 dbg_type = b'unknown'
1395 1395
1396 1396 if p1r is None:
1397 1397 p1r = revlog.rev(revinfo.p1)
1398 1398 p2r = revlog.rev(revinfo.p2)
1399 1399
1400 1400 if self._debug_search:
1401 1401 msg = b"DBG-DELTAS-SEARCH: SEARCH rev=%d\n"
1402 1402 msg %= target_rev
1403 1403 self._write_debug(msg)
1404 1404
1405 1405 groups = _candidategroups(
1406 1406 self.revlog,
1407 1407 revinfo.textlen,
1408 1408 p1r,
1409 1409 p2r,
1410 1410 cachedelta,
1411 1411 excluded_bases,
1412 1412 target_rev,
1413 1413 snapshot_cache=self._snapshot_cache,
1414 1414 )
1415 1415 candidaterevs = next(groups)
1416 1416 while candidaterevs is not None:
1417 1417 dbg_try_rounds += 1
1418 1418 if self._debug_search:
1419 1419 prev = None
1420 1420 if deltainfo is not None:
1421 1421 prev = deltainfo.base
1422 1422
1423 1423 if (
1424 1424 cachedelta is not None
1425 1425 and len(candidaterevs) == 1
1426 1426 and cachedelta[0] in candidaterevs
1427 1427 ):
1428 1428 round_type = b"cached-delta"
1429 1429 elif p1r in candidaterevs or p2r in candidaterevs:
1430 1430 round_type = b"parents"
1431 1431 elif prev is not None and all(c < prev for c in candidaterevs):
1432 1432 round_type = b"refine-down"
1433 1433 elif prev is not None and all(c > prev for c in candidaterevs):
1434 1434 round_type = b"refine-up"
1435 1435 else:
1436 1436 round_type = b"search-down"
1437 1437 msg = b"DBG-DELTAS-SEARCH: ROUND #%d - %d candidates - %s\n"
1438 1438 msg %= (dbg_try_rounds, len(candidaterevs), round_type)
1439 1439 self._write_debug(msg)
1440 1440 nominateddeltas = []
1441 1441 if deltainfo is not None:
1442 1442 if self._debug_search:
1443 1443 msg = (
1444 1444 b"DBG-DELTAS-SEARCH: CONTENDER: rev=%d - length=%d\n"
1445 1445 )
1446 1446 msg %= (deltainfo.base, deltainfo.deltalen)
1447 1447 self._write_debug(msg)
1448 1448 # if we already found a good delta,
1449 1449 # challenge it against refined candidates
1450 1450 nominateddeltas.append(deltainfo)
1451 1451 for candidaterev in candidaterevs:
1452 1452 if self._debug_search:
1453 1453 msg = b"DBG-DELTAS-SEARCH: CANDIDATE: rev=%d\n"
1454 1454 msg %= candidaterev
1455 1455 self._write_debug(msg)
1456 1456 candidate_type = None
1457 1457 if candidaterev == p1r:
1458 1458 candidate_type = b"p1"
1459 1459 elif candidaterev == p2r:
1460 1460 candidate_type = b"p2"
1461 1461 elif self.revlog.issnapshot(candidaterev):
1462 1462 candidate_type = b"snapshot-%d"
1463 1463 candidate_type %= self.revlog.snapshotdepth(
1464 1464 candidaterev
1465 1465 )
1466 1466
1467 1467 if candidate_type is not None:
1468 1468 msg = b"DBG-DELTAS-SEARCH: type=%s\n"
1469 1469 msg %= candidate_type
1470 1470 self._write_debug(msg)
1471 1471 msg = b"DBG-DELTAS-SEARCH: size=%d\n"
1472 1472 msg %= self.revlog.length(candidaterev)
1473 1473 self._write_debug(msg)
1474 1474 msg = b"DBG-DELTAS-SEARCH: base=%d\n"
1475 1475 msg %= self.revlog.deltaparent(candidaterev)
1476 1476 self._write_debug(msg)
1477 1477
1478 1478 dbg_try_count += 1
1479 1479
1480 1480 if self._debug_search:
1481 1481 delta_start = util.timer()
1482 1482 candidatedelta = self._builddeltainfo(
1483 1483 revinfo,
1484 1484 candidaterev,
1485 1485 target_rev=target_rev,
1486 1486 )
1487 1487 if self._debug_search:
1488 1488 delta_end = util.timer()
1489 1489 msg = b"DBG-DELTAS-SEARCH: delta-search-time=%f\n"
1490 1490 msg %= delta_end - delta_start
1491 1491 self._write_debug(msg)
1492 1492 if candidatedelta is not None:
1493 1493 if is_good_delta_info(self.revlog, candidatedelta, revinfo):
1494 1494 if self._debug_search:
1495 1495 msg = b"DBG-DELTAS-SEARCH: DELTA: length=%d (GOOD)\n"
1496 1496 msg %= candidatedelta.deltalen
1497 1497 self._write_debug(msg)
1498 1498 nominateddeltas.append(candidatedelta)
1499 1499 elif self._debug_search:
1500 1500 msg = b"DBG-DELTAS-SEARCH: DELTA: length=%d (BAD)\n"
1501 1501 msg %= candidatedelta.deltalen
1502 1502 self._write_debug(msg)
1503 1503 elif self._debug_search:
1504 1504 msg = b"DBG-DELTAS-SEARCH: NO-DELTA\n"
1505 1505 self._write_debug(msg)
1506 1506 if nominateddeltas:
1507 1507 deltainfo = min(nominateddeltas, key=lambda x: x.deltalen)
1508 1508 if deltainfo is not None:
1509 1509 candidaterevs = groups.send(deltainfo.base)
1510 1510 else:
1511 1511 candidaterevs = next(groups)
1512 1512
1513 1513 if deltainfo is None:
1514 1514 dbg_type = b"full"
1515 1515 deltainfo = self._fullsnapshotinfo(revinfo, target_rev)
1516 1516 elif deltainfo.snapshotdepth: # pytype: disable=attribute-error
1517 1517 dbg_type = b"snapshot"
1518 1518 else:
1519 1519 dbg_type = b"delta"
1520 1520
1521 1521 if gather_debug:
1522 1522 end = util.timer()
1523 1523 if dbg_type == b'full':
1524 1524 used_cached = (
1525 1525 cachedelta is not None
1526 1526 and dbg_try_rounds == 0
1527 1527 and dbg_try_count == 0
1528 1528 and cachedelta[0] == nullrev
1529 1529 )
1530 1530 else:
1531 1531 used_cached = (
1532 1532 cachedelta is not None
1533 1533 and dbg_try_rounds == 1
1534 1534 and dbg_try_count == 1
1535 1535 and deltainfo.base == cachedelta[0]
1536 1536 )
1537 1537 dbg['duration'] = end - start
1538 1538 dbg[
1539 1539 'delta-base'
1540 1540 ] = deltainfo.base # pytype: disable=attribute-error
1541 1541 dbg['search_round_count'] = dbg_try_rounds
1542 1542 dbg['using-cached-base'] = used_cached
1543 1543 dbg['delta_try_count'] = dbg_try_count
1544 1544 dbg['type'] = dbg_type
1545 1545 if (
1546 1546 deltainfo.snapshotdepth # pytype: disable=attribute-error
1547 1547 is not None
1548 1548 ):
1549 1549 dbg[
1550 1550 'snapshot-depth'
1551 1551 ] = deltainfo.snapshotdepth # pytype: disable=attribute-error
1552 1552 else:
1553 1553 dbg['snapshot-depth'] = 0
1554 1554 self._dbg_process_data(dbg)
1555 1555 return deltainfo
1556 1556
1557 1557 def _one_dbg_data(self):
1558 1558 return {
1559 1559 'duration': None,
1560 1560 'revision': None,
1561 1561 'delta-base': None,
1562 1562 'search_round_count': None,
1563 1563 'using-cached-base': None,
1564 1564 'delta_try_count': None,
1565 1565 'type': None,
1566 1566 'p1-chain-len': None,
1567 1567 'p2-chain-len': None,
1568 1568 'snapshot-depth': None,
1569 1569 'target-revlog': None,
1570 1570 }
1571 1571
1572 1572 def _dbg_process_data(self, dbg):
1573 1573 if self._debug_info is not None:
1574 1574 self._debug_info.append(dbg)
1575 1575
1576 1576 if self._write_debug is not None:
1577 1577 msg = (
1578 1578 b"DBG-DELTAS:"
1579 1579 b" %-12s"
1580 1580 b" rev=%d:"
1581 1581 b" delta-base=%d"
1582 1582 b" is-cached=%d"
1583 1583 b" - search-rounds=%d"
1584 1584 b" try-count=%d"
1585 1585 b" - delta-type=%-6s"
1586 1586 b" snap-depth=%d"
1587 1587 b" - p1-chain-length=%d"
1588 1588 b" p2-chain-length=%d"
1589 1589 b" - duration=%f"
1590 1590 b"\n"
1591 1591 )
1592 1592 msg %= (
1593 1593 dbg["target-revlog"],
1594 1594 dbg["revision"],
1595 1595 dbg["delta-base"],
1596 1596 dbg["using-cached-base"],
1597 1597 dbg["search_round_count"],
1598 1598 dbg["delta_try_count"],
1599 1599 dbg["type"],
1600 1600 dbg["snapshot-depth"],
1601 1601 dbg["p1-chain-len"],
1602 1602 dbg["p2-chain-len"],
1603 1603 dbg["duration"],
1604 1604 )
1605 1605 self._write_debug(msg)
1606 1606
1607 1607
1608 1608 def delta_compression(default_compression_header, deltainfo):
1609 1609 """return (COMPRESSION_MODE, deltainfo)
1610 1610
1611 1611 used by revlog v2+ format to dispatch between PLAIN and DEFAULT
1612 1612 compression.
1613 1613 """
1614 1614 h, d = deltainfo.data
1615 1615 compression_mode = COMP_MODE_INLINE
1616 1616 if not h and not d:
1617 1617 # not data to store at all... declare them uncompressed
1618 1618 compression_mode = COMP_MODE_PLAIN
1619 1619 elif not h:
1620 1620 t = d[0:1]
1621 1621 if t == b'\0':
1622 1622 compression_mode = COMP_MODE_PLAIN
1623 1623 elif t == default_compression_header:
1624 1624 compression_mode = COMP_MODE_DEFAULT
1625 1625 elif h == b'u':
1626 1626 # we have a more efficient way to declare uncompressed
1627 1627 h = b''
1628 1628 compression_mode = COMP_MODE_PLAIN
1629 1629 deltainfo = drop_u_compression(deltainfo)
1630 1630 return compression_mode, deltainfo
General Comments 0
You need to be logged in to leave comments. Login now