##// END OF EJS Templates
infinitepush: drop the default value of config options which are registered...
Pulkit Goyal -
r37250:912f4f64 default
parent child Browse files
Show More
@@ -1,1186 +1,1186
1 1 # Infinite push
2 2 #
3 3 # Copyright 2016 Facebook, Inc.
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7 """ store some pushes in a remote blob store on the server (EXPERIMENTAL)
8 8
9 9 [infinitepush]
10 10 # Server-side and client-side option. Pattern of the infinitepush bookmark
11 11 branchpattern = PATTERN
12 12
13 13 # Server or client
14 14 server = False
15 15
16 16 # Server-side option. Possible values: 'disk' or 'sql'. Fails if not set
17 17 indextype = disk
18 18
19 19 # Server-side option. Used only if indextype=sql.
20 20 # Format: 'IP:PORT:DB_NAME:USER:PASSWORD'
21 21 sqlhost = IP:PORT:DB_NAME:USER:PASSWORD
22 22
23 23 # Server-side option. Used only if indextype=disk.
24 24 # Filesystem path to the index store
25 25 indexpath = PATH
26 26
27 27 # Server-side option. Possible values: 'disk' or 'external'
28 28 # Fails if not set
29 29 storetype = disk
30 30
31 31 # Server-side option.
32 32 # Path to the binary that will save bundle to the bundlestore
33 33 # Formatted cmd line will be passed to it (see `put_args`)
34 34 put_binary = put
35 35
36 36 # Serser-side option. Used only if storetype=external.
37 37 # Format cmd-line string for put binary. Placeholder: {filename}
38 38 put_args = {filename}
39 39
40 40 # Server-side option.
41 41 # Path to the binary that get bundle from the bundlestore.
42 42 # Formatted cmd line will be passed to it (see `get_args`)
43 43 get_binary = get
44 44
45 45 # Serser-side option. Used only if storetype=external.
46 46 # Format cmd-line string for get binary. Placeholders: {filename} {handle}
47 47 get_args = {filename} {handle}
48 48
49 49 # Server-side option
50 50 logfile = FIlE
51 51
52 52 # Server-side option
53 53 loglevel = DEBUG
54 54
55 55 # Server-side option. Used only if indextype=sql.
56 56 # Sets mysql wait_timeout option.
57 57 waittimeout = 300
58 58
59 59 # Server-side option. Used only if indextype=sql.
60 60 # Sets mysql innodb_lock_wait_timeout option.
61 61 locktimeout = 120
62 62
63 63 # Server-side option. Used only if indextype=sql.
64 64 # Name of the repository
65 65 reponame = ''
66 66
67 67 # Client-side option. Used by --list-remote option. List of remote scratch
68 68 # patterns to list if no patterns are specified.
69 69 defaultremotepatterns = ['*']
70 70
71 71 # Instructs infinitepush to forward all received bundle2 parts to the
72 72 # bundle for storage. Defaults to False.
73 73 storeallparts = True
74 74
75 75 # routes each incoming push to the bundlestore. defaults to False
76 76 pushtobundlestore = True
77 77
78 78 [remotenames]
79 79 # Client-side option
80 80 # This option should be set only if remotenames extension is enabled.
81 81 # Whether remote bookmarks are tracked by remotenames extension.
82 82 bookmarks = True
83 83 """
84 84
85 85 from __future__ import absolute_import
86 86
87 87 import collections
88 88 import contextlib
89 89 import errno
90 90 import functools
91 91 import logging
92 92 import os
93 93 import random
94 94 import re
95 95 import socket
96 96 import subprocess
97 97 import tempfile
98 98 import time
99 99
100 100 from mercurial.node import (
101 101 bin,
102 102 hex,
103 103 )
104 104
105 105 from mercurial.i18n import _
106 106
107 107 from mercurial.utils import (
108 108 procutil,
109 109 stringutil,
110 110 )
111 111
112 112 from mercurial import (
113 113 bundle2,
114 114 changegroup,
115 115 commands,
116 116 discovery,
117 117 encoding,
118 118 error,
119 119 exchange,
120 120 extensions,
121 121 hg,
122 122 localrepo,
123 123 peer,
124 124 phases,
125 125 pushkey,
126 126 registrar,
127 127 util,
128 128 wireproto,
129 129 )
130 130
131 131 from . import (
132 132 bundleparts,
133 133 common,
134 134 )
135 135
136 136 # Note for extension authors: ONLY specify testedwith = 'ships-with-hg-core' for
137 137 # extensions which SHIP WITH MERCURIAL. Non-mainline extensions should
138 138 # be specifying the version(s) of Mercurial they are tested with, or
139 139 # leave the attribute unspecified.
140 140 testedwith = 'ships-with-hg-core'
141 141
142 142 configtable = {}
143 143 configitem = registrar.configitem(configtable)
144 144
145 145 configitem('infinitepush', 'server',
146 146 default=False,
147 147 )
148 148 configitem('infinitepush', 'storetype',
149 149 default='',
150 150 )
151 151 configitem('infinitepush', 'indextype',
152 152 default='',
153 153 )
154 154 configitem('infinitepush', 'indexpath',
155 155 default='',
156 156 )
157 157 configitem('infinitepush', 'storeallparts',
158 158 default=False,
159 159 )
160 160 configitem('infinitepush', 'reponame',
161 161 default='',
162 162 )
163 163 configitem('scratchbranch', 'storepath',
164 164 default='',
165 165 )
166 166 configitem('infinitepush', 'branchpattern',
167 167 default='',
168 168 )
169 169 configitem('infinitepush', 'pushtobundlestore',
170 170 default=False,
171 171 )
172 172 configitem('experimental', 'server-bundlestore-bookmark',
173 173 default='',
174 174 )
175 175 configitem('experimental', 'infinitepush-scratchpush',
176 176 default=False,
177 177 )
178 178
179 179 experimental = 'experimental'
180 180 configbookmark = 'server-bundlestore-bookmark'
181 181 configscratchpush = 'infinitepush-scratchpush'
182 182
183 183 scratchbranchparttype = bundleparts.scratchbranchparttype
184 184 revsetpredicate = registrar.revsetpredicate()
185 185 templatekeyword = registrar.templatekeyword()
186 186 _scratchbranchmatcher = lambda x: False
187 187 _maybehash = re.compile(r'^[a-f0-9]+$').search
188 188
189 189 def _buildexternalbundlestore(ui):
190 190 put_args = ui.configlist('infinitepush', 'put_args', [])
191 191 put_binary = ui.config('infinitepush', 'put_binary')
192 192 if not put_binary:
193 193 raise error.Abort('put binary is not specified')
194 194 get_args = ui.configlist('infinitepush', 'get_args', [])
195 195 get_binary = ui.config('infinitepush', 'get_binary')
196 196 if not get_binary:
197 197 raise error.Abort('get binary is not specified')
198 198 from . import store
199 199 return store.externalbundlestore(put_binary, put_args, get_binary, get_args)
200 200
201 201 def _buildsqlindex(ui):
202 202 sqlhost = ui.config('infinitepush', 'sqlhost')
203 203 if not sqlhost:
204 204 raise error.Abort(_('please set infinitepush.sqlhost'))
205 205 host, port, db, user, password = sqlhost.split(':')
206 206 reponame = ui.config('infinitepush', 'reponame')
207 207 if not reponame:
208 208 raise error.Abort(_('please set infinitepush.reponame'))
209 209
210 210 logfile = ui.config('infinitepush', 'logfile', '')
211 211 waittimeout = ui.configint('infinitepush', 'waittimeout', 300)
212 212 locktimeout = ui.configint('infinitepush', 'locktimeout', 120)
213 213 from . import sqlindexapi
214 214 return sqlindexapi.sqlindexapi(
215 215 reponame, host, port, db, user, password,
216 216 logfile, _getloglevel(ui), waittimeout=waittimeout,
217 217 locktimeout=locktimeout)
218 218
219 219 def _getloglevel(ui):
220 220 loglevel = ui.config('infinitepush', 'loglevel', 'DEBUG')
221 221 numeric_loglevel = getattr(logging, loglevel.upper(), None)
222 222 if not isinstance(numeric_loglevel, int):
223 223 raise error.Abort(_('invalid log level %s') % loglevel)
224 224 return numeric_loglevel
225 225
226 226 def _tryhoist(ui, remotebookmark):
227 227 '''returns a bookmarks with hoisted part removed
228 228
229 229 Remotenames extension has a 'hoist' config that allows to use remote
230 230 bookmarks without specifying remote path. For example, 'hg update master'
231 231 works as well as 'hg update remote/master'. We want to allow the same in
232 232 infinitepush.
233 233 '''
234 234
235 235 if common.isremotebooksenabled(ui):
236 236 hoist = ui.config('remotenames', 'hoistedpeer') + '/'
237 237 if remotebookmark.startswith(hoist):
238 238 return remotebookmark[len(hoist):]
239 239 return remotebookmark
240 240
241 241 class bundlestore(object):
242 242 def __init__(self, repo):
243 243 self._repo = repo
244 storetype = self._repo.ui.config('infinitepush', 'storetype', '')
244 storetype = self._repo.ui.config('infinitepush', 'storetype')
245 245 if storetype == 'disk':
246 246 from . import store
247 247 self.store = store.filebundlestore(self._repo.ui, self._repo)
248 248 elif storetype == 'external':
249 249 self.store = _buildexternalbundlestore(self._repo.ui)
250 250 else:
251 251 raise error.Abort(
252 252 _('unknown infinitepush store type specified %s') % storetype)
253 253
254 indextype = self._repo.ui.config('infinitepush', 'indextype', '')
254 indextype = self._repo.ui.config('infinitepush', 'indextype')
255 255 if indextype == 'disk':
256 256 from . import fileindexapi
257 257 self.index = fileindexapi.fileindexapi(self._repo)
258 258 elif indextype == 'sql':
259 259 self.index = _buildsqlindex(self._repo.ui)
260 260 else:
261 261 raise error.Abort(
262 262 _('unknown infinitepush index type specified %s') % indextype)
263 263
264 264 def _isserver(ui):
265 265 return ui.configbool('infinitepush', 'server')
266 266
267 267 def reposetup(ui, repo):
268 268 if _isserver(ui) and repo.local():
269 269 repo.bundlestore = bundlestore(repo)
270 270
271 271 def extsetup(ui):
272 272 commonsetup(ui)
273 273 if _isserver(ui):
274 274 serverextsetup(ui)
275 275 else:
276 276 clientextsetup(ui)
277 277
278 278 def commonsetup(ui):
279 279 wireproto.commands['listkeyspatterns'] = (
280 280 wireprotolistkeyspatterns, 'namespace patterns')
281 281 scratchbranchpat = ui.config('infinitepush', 'branchpattern')
282 282 if scratchbranchpat:
283 283 global _scratchbranchmatcher
284 284 kind, pat, _scratchbranchmatcher = \
285 285 stringutil.stringmatcher(scratchbranchpat)
286 286
287 287 def serverextsetup(ui):
288 288 origpushkeyhandler = bundle2.parthandlermapping['pushkey']
289 289
290 290 def newpushkeyhandler(*args, **kwargs):
291 291 bundle2pushkey(origpushkeyhandler, *args, **kwargs)
292 292 newpushkeyhandler.params = origpushkeyhandler.params
293 293 bundle2.parthandlermapping['pushkey'] = newpushkeyhandler
294 294
295 295 orighandlephasehandler = bundle2.parthandlermapping['phase-heads']
296 296 newphaseheadshandler = lambda *args, **kwargs: \
297 297 bundle2handlephases(orighandlephasehandler, *args, **kwargs)
298 298 newphaseheadshandler.params = orighandlephasehandler.params
299 299 bundle2.parthandlermapping['phase-heads'] = newphaseheadshandler
300 300
301 301 extensions.wrapfunction(localrepo.localrepository, 'listkeys',
302 302 localrepolistkeys)
303 303 wireproto.commands['lookup'] = (
304 304 _lookupwrap(wireproto.commands['lookup'][0]), 'key')
305 305 extensions.wrapfunction(exchange, 'getbundlechunks', getbundlechunks)
306 306
307 307 extensions.wrapfunction(bundle2, 'processparts', processparts)
308 308
309 309 def clientextsetup(ui):
310 310 entry = extensions.wrapcommand(commands.table, 'push', _push)
311 311
312 312 entry[1].append(
313 313 ('', 'bundle-store', None,
314 314 _('force push to go to bundle store (EXPERIMENTAL)')))
315 315
316 316 extensions.wrapcommand(commands.table, 'pull', _pull)
317 317
318 318 extensions.wrapfunction(discovery, 'checkheads', _checkheads)
319 319
320 320 wireproto.wirepeer.listkeyspatterns = listkeyspatterns
321 321
322 322 partorder = exchange.b2partsgenorder
323 323 index = partorder.index('changeset')
324 324 partorder.insert(
325 325 index, partorder.pop(partorder.index(scratchbranchparttype)))
326 326
327 327 def _checkheads(orig, pushop):
328 328 if pushop.ui.configbool(experimental, configscratchpush, False):
329 329 return
330 330 return orig(pushop)
331 331
332 332 def wireprotolistkeyspatterns(repo, proto, namespace, patterns):
333 333 patterns = wireproto.decodelist(patterns)
334 334 d = repo.listkeys(encoding.tolocal(namespace), patterns).iteritems()
335 335 return pushkey.encodekeys(d)
336 336
337 337 def localrepolistkeys(orig, self, namespace, patterns=None):
338 338 if namespace == 'bookmarks' and patterns:
339 339 index = self.bundlestore.index
340 340 results = {}
341 341 bookmarks = orig(self, namespace)
342 342 for pattern in patterns:
343 343 results.update(index.getbookmarks(pattern))
344 344 if pattern.endswith('*'):
345 345 pattern = 're:^' + pattern[:-1] + '.*'
346 346 kind, pat, matcher = stringutil.stringmatcher(pattern)
347 347 for bookmark, node in bookmarks.iteritems():
348 348 if matcher(bookmark):
349 349 results[bookmark] = node
350 350 return results
351 351 else:
352 352 return orig(self, namespace)
353 353
354 354 @peer.batchable
355 355 def listkeyspatterns(self, namespace, patterns):
356 356 if not self.capable('pushkey'):
357 357 yield {}, None
358 358 f = peer.future()
359 359 self.ui.debug('preparing listkeys for "%s" with pattern "%s"\n' %
360 360 (namespace, patterns))
361 361 yield {
362 362 'namespace': encoding.fromlocal(namespace),
363 363 'patterns': wireproto.encodelist(patterns)
364 364 }, f
365 365 d = f.value
366 366 self.ui.debug('received listkey for "%s": %i bytes\n'
367 367 % (namespace, len(d)))
368 368 yield pushkey.decodekeys(d)
369 369
370 370 def _readbundlerevs(bundlerepo):
371 371 return list(bundlerepo.revs('bundle()'))
372 372
373 373 def _includefilelogstobundle(bundlecaps, bundlerepo, bundlerevs, ui):
374 374 '''Tells remotefilelog to include all changed files to the changegroup
375 375
376 376 By default remotefilelog doesn't include file content to the changegroup.
377 377 But we need to include it if we are fetching from bundlestore.
378 378 '''
379 379 changedfiles = set()
380 380 cl = bundlerepo.changelog
381 381 for r in bundlerevs:
382 382 # [3] means changed files
383 383 changedfiles.update(cl.read(r)[3])
384 384 if not changedfiles:
385 385 return bundlecaps
386 386
387 387 changedfiles = '\0'.join(changedfiles)
388 388 newcaps = []
389 389 appended = False
390 390 for cap in (bundlecaps or []):
391 391 if cap.startswith('excludepattern='):
392 392 newcaps.append('\0'.join((cap, changedfiles)))
393 393 appended = True
394 394 else:
395 395 newcaps.append(cap)
396 396 if not appended:
397 397 # Not found excludepattern cap. Just append it
398 398 newcaps.append('excludepattern=' + changedfiles)
399 399
400 400 return newcaps
401 401
402 402 def _rebundle(bundlerepo, bundleroots, unknownhead):
403 403 '''
404 404 Bundle may include more revision then user requested. For example,
405 405 if user asks for revision but bundle also consists its descendants.
406 406 This function will filter out all revision that user is not requested.
407 407 '''
408 408 parts = []
409 409
410 410 version = '02'
411 411 outgoing = discovery.outgoing(bundlerepo, commonheads=bundleroots,
412 412 missingheads=[unknownhead])
413 413 cgstream = changegroup.makestream(bundlerepo, outgoing, version, 'pull')
414 414 cgstream = util.chunkbuffer(cgstream).read()
415 415 cgpart = bundle2.bundlepart('changegroup', data=cgstream)
416 416 cgpart.addparam('version', version)
417 417 parts.append(cgpart)
418 418
419 419 return parts
420 420
421 421 def _getbundleroots(oldrepo, bundlerepo, bundlerevs):
422 422 cl = bundlerepo.changelog
423 423 bundleroots = []
424 424 for rev in bundlerevs:
425 425 node = cl.node(rev)
426 426 parents = cl.parents(node)
427 427 for parent in parents:
428 428 # include all revs that exist in the main repo
429 429 # to make sure that bundle may apply client-side
430 430 if parent in oldrepo:
431 431 bundleroots.append(parent)
432 432 return bundleroots
433 433
434 434 def _needsrebundling(head, bundlerepo):
435 435 bundleheads = list(bundlerepo.revs('heads(bundle())'))
436 436 return not (len(bundleheads) == 1 and
437 437 bundlerepo[bundleheads[0]].node() == head)
438 438
439 439 def _generateoutputparts(head, bundlerepo, bundleroots, bundlefile):
440 440 '''generates bundle that will be send to the user
441 441
442 442 returns tuple with raw bundle string and bundle type
443 443 '''
444 444 parts = []
445 445 if not _needsrebundling(head, bundlerepo):
446 446 with util.posixfile(bundlefile, "rb") as f:
447 447 unbundler = exchange.readbundle(bundlerepo.ui, f, bundlefile)
448 448 if isinstance(unbundler, changegroup.cg1unpacker):
449 449 part = bundle2.bundlepart('changegroup',
450 450 data=unbundler._stream.read())
451 451 part.addparam('version', '01')
452 452 parts.append(part)
453 453 elif isinstance(unbundler, bundle2.unbundle20):
454 454 haschangegroup = False
455 455 for part in unbundler.iterparts():
456 456 if part.type == 'changegroup':
457 457 haschangegroup = True
458 458 newpart = bundle2.bundlepart(part.type, data=part.read())
459 459 for key, value in part.params.iteritems():
460 460 newpart.addparam(key, value)
461 461 parts.append(newpart)
462 462
463 463 if not haschangegroup:
464 464 raise error.Abort(
465 465 'unexpected bundle without changegroup part, ' +
466 466 'head: %s' % hex(head),
467 467 hint='report to administrator')
468 468 else:
469 469 raise error.Abort('unknown bundle type')
470 470 else:
471 471 parts = _rebundle(bundlerepo, bundleroots, head)
472 472
473 473 return parts
474 474
475 475 def getbundlechunks(orig, repo, source, heads=None, bundlecaps=None, **kwargs):
476 476 heads = heads or []
477 477 # newheads are parents of roots of scratch bundles that were requested
478 478 newphases = {}
479 479 scratchbundles = []
480 480 newheads = []
481 481 scratchheads = []
482 482 nodestobundle = {}
483 483 allbundlestocleanup = []
484 484 try:
485 485 for head in heads:
486 486 if head not in repo.changelog.nodemap:
487 487 if head not in nodestobundle:
488 488 newbundlefile = common.downloadbundle(repo, head)
489 489 bundlepath = "bundle:%s+%s" % (repo.root, newbundlefile)
490 490 bundlerepo = hg.repository(repo.ui, bundlepath)
491 491
492 492 allbundlestocleanup.append((bundlerepo, newbundlefile))
493 493 bundlerevs = set(_readbundlerevs(bundlerepo))
494 494 bundlecaps = _includefilelogstobundle(
495 495 bundlecaps, bundlerepo, bundlerevs, repo.ui)
496 496 cl = bundlerepo.changelog
497 497 bundleroots = _getbundleroots(repo, bundlerepo, bundlerevs)
498 498 for rev in bundlerevs:
499 499 node = cl.node(rev)
500 500 newphases[hex(node)] = str(phases.draft)
501 501 nodestobundle[node] = (bundlerepo, bundleroots,
502 502 newbundlefile)
503 503
504 504 scratchbundles.append(
505 505 _generateoutputparts(head, *nodestobundle[head]))
506 506 newheads.extend(bundleroots)
507 507 scratchheads.append(head)
508 508 finally:
509 509 for bundlerepo, bundlefile in allbundlestocleanup:
510 510 bundlerepo.close()
511 511 try:
512 512 os.unlink(bundlefile)
513 513 except (IOError, OSError):
514 514 # if we can't cleanup the file then just ignore the error,
515 515 # no need to fail
516 516 pass
517 517
518 518 pullfrombundlestore = bool(scratchbundles)
519 519 wrappedchangegrouppart = False
520 520 wrappedlistkeys = False
521 521 oldchangegrouppart = exchange.getbundle2partsmapping['changegroup']
522 522 try:
523 523 def _changegrouppart(bundler, *args, **kwargs):
524 524 # Order is important here. First add non-scratch part
525 525 # and only then add parts with scratch bundles because
526 526 # non-scratch part contains parents of roots of scratch bundles.
527 527 result = oldchangegrouppart(bundler, *args, **kwargs)
528 528 for bundle in scratchbundles:
529 529 for part in bundle:
530 530 bundler.addpart(part)
531 531 return result
532 532
533 533 exchange.getbundle2partsmapping['changegroup'] = _changegrouppart
534 534 wrappedchangegrouppart = True
535 535
536 536 def _listkeys(orig, self, namespace):
537 537 origvalues = orig(self, namespace)
538 538 if namespace == 'phases' and pullfrombundlestore:
539 539 if origvalues.get('publishing') == 'True':
540 540 # Make repo non-publishing to preserve draft phase
541 541 del origvalues['publishing']
542 542 origvalues.update(newphases)
543 543 return origvalues
544 544
545 545 extensions.wrapfunction(localrepo.localrepository, 'listkeys',
546 546 _listkeys)
547 547 wrappedlistkeys = True
548 548 heads = list((set(newheads) | set(heads)) - set(scratchheads))
549 549 result = orig(repo, source, heads=heads,
550 550 bundlecaps=bundlecaps, **kwargs)
551 551 finally:
552 552 if wrappedchangegrouppart:
553 553 exchange.getbundle2partsmapping['changegroup'] = oldchangegrouppart
554 554 if wrappedlistkeys:
555 555 extensions.unwrapfunction(localrepo.localrepository, 'listkeys',
556 556 _listkeys)
557 557 return result
558 558
559 559 def _lookupwrap(orig):
560 560 def _lookup(repo, proto, key):
561 561 localkey = encoding.tolocal(key)
562 562
563 563 if isinstance(localkey, str) and _scratchbranchmatcher(localkey):
564 564 scratchnode = repo.bundlestore.index.getnode(localkey)
565 565 if scratchnode:
566 566 return "%s %s\n" % (1, scratchnode)
567 567 else:
568 568 return "%s %s\n" % (0, 'scratch branch %s not found' % localkey)
569 569 else:
570 570 try:
571 571 r = hex(repo.lookup(localkey))
572 572 return "%s %s\n" % (1, r)
573 573 except Exception as inst:
574 574 if repo.bundlestore.index.getbundle(localkey):
575 575 return "%s %s\n" % (1, localkey)
576 576 else:
577 577 r = str(inst)
578 578 return "%s %s\n" % (0, r)
579 579 return _lookup
580 580
581 581 def _pull(orig, ui, repo, source="default", **opts):
582 582 # Copy paste from `pull` command
583 583 source, branches = hg.parseurl(ui.expandpath(source), opts.get('branch'))
584 584
585 585 scratchbookmarks = {}
586 586 unfi = repo.unfiltered()
587 587 unknownnodes = []
588 588 for rev in opts.get('rev', []):
589 589 if rev not in unfi:
590 590 unknownnodes.append(rev)
591 591 if opts.get('bookmark'):
592 592 bookmarks = []
593 593 revs = opts.get('rev') or []
594 594 for bookmark in opts.get('bookmark'):
595 595 if _scratchbranchmatcher(bookmark):
596 596 # rev is not known yet
597 597 # it will be fetched with listkeyspatterns next
598 598 scratchbookmarks[bookmark] = 'REVTOFETCH'
599 599 else:
600 600 bookmarks.append(bookmark)
601 601
602 602 if scratchbookmarks:
603 603 other = hg.peer(repo, opts, source)
604 604 fetchedbookmarks = other.listkeyspatterns(
605 605 'bookmarks', patterns=scratchbookmarks)
606 606 for bookmark in scratchbookmarks:
607 607 if bookmark not in fetchedbookmarks:
608 608 raise error.Abort('remote bookmark %s not found!' %
609 609 bookmark)
610 610 scratchbookmarks[bookmark] = fetchedbookmarks[bookmark]
611 611 revs.append(fetchedbookmarks[bookmark])
612 612 opts['bookmark'] = bookmarks
613 613 opts['rev'] = revs
614 614
615 615 if scratchbookmarks or unknownnodes:
616 616 # Set anyincoming to True
617 617 extensions.wrapfunction(discovery, 'findcommonincoming',
618 618 _findcommonincoming)
619 619 try:
620 620 # Remote scratch bookmarks will be deleted because remotenames doesn't
621 621 # know about them. Let's save it before pull and restore after
622 622 remotescratchbookmarks = _readscratchremotebookmarks(ui, repo, source)
623 623 result = orig(ui, repo, source, **opts)
624 624 # TODO(stash): race condition is possible
625 625 # if scratch bookmarks was updated right after orig.
626 626 # But that's unlikely and shouldn't be harmful.
627 627 if common.isremotebooksenabled(ui):
628 628 remotescratchbookmarks.update(scratchbookmarks)
629 629 _saveremotebookmarks(repo, remotescratchbookmarks, source)
630 630 else:
631 631 _savelocalbookmarks(repo, scratchbookmarks)
632 632 return result
633 633 finally:
634 634 if scratchbookmarks:
635 635 extensions.unwrapfunction(discovery, 'findcommonincoming')
636 636
637 637 def _readscratchremotebookmarks(ui, repo, other):
638 638 if common.isremotebooksenabled(ui):
639 639 remotenamesext = extensions.find('remotenames')
640 640 remotepath = remotenamesext.activepath(repo.ui, other)
641 641 result = {}
642 642 # Let's refresh remotenames to make sure we have it up to date
643 643 # Seems that `repo.names['remotebookmarks']` may return stale bookmarks
644 644 # and it results in deleting scratch bookmarks. Our best guess how to
645 645 # fix it is to use `clearnames()`
646 646 repo._remotenames.clearnames()
647 647 for remotebookmark in repo.names['remotebookmarks'].listnames(repo):
648 648 path, bookname = remotenamesext.splitremotename(remotebookmark)
649 649 if path == remotepath and _scratchbranchmatcher(bookname):
650 650 nodes = repo.names['remotebookmarks'].nodes(repo,
651 651 remotebookmark)
652 652 if nodes:
653 653 result[bookname] = hex(nodes[0])
654 654 return result
655 655 else:
656 656 return {}
657 657
658 658 def _saveremotebookmarks(repo, newbookmarks, remote):
659 659 remotenamesext = extensions.find('remotenames')
660 660 remotepath = remotenamesext.activepath(repo.ui, remote)
661 661 branches = collections.defaultdict(list)
662 662 bookmarks = {}
663 663 remotenames = remotenamesext.readremotenames(repo)
664 664 for hexnode, nametype, remote, rname in remotenames:
665 665 if remote != remotepath:
666 666 continue
667 667 if nametype == 'bookmarks':
668 668 if rname in newbookmarks:
669 669 # It's possible if we have a normal bookmark that matches
670 670 # scratch branch pattern. In this case just use the current
671 671 # bookmark node
672 672 del newbookmarks[rname]
673 673 bookmarks[rname] = hexnode
674 674 elif nametype == 'branches':
675 675 # saveremotenames expects 20 byte binary nodes for branches
676 676 branches[rname].append(bin(hexnode))
677 677
678 678 for bookmark, hexnode in newbookmarks.iteritems():
679 679 bookmarks[bookmark] = hexnode
680 680 remotenamesext.saveremotenames(repo, remotepath, branches, bookmarks)
681 681
682 682 def _savelocalbookmarks(repo, bookmarks):
683 683 if not bookmarks:
684 684 return
685 685 with repo.wlock(), repo.lock(), repo.transaction('bookmark') as tr:
686 686 changes = []
687 687 for scratchbook, node in bookmarks.iteritems():
688 688 changectx = repo[node]
689 689 changes.append((scratchbook, changectx.node()))
690 690 repo._bookmarks.applychanges(repo, tr, changes)
691 691
692 692 def _findcommonincoming(orig, *args, **kwargs):
693 693 common, inc, remoteheads = orig(*args, **kwargs)
694 694 return common, True, remoteheads
695 695
696 696 def _push(orig, ui, repo, dest=None, *args, **opts):
697 697
698 698 bookmark = opts.get('bookmark')
699 699 # we only support pushing one infinitepush bookmark at once
700 700 if len(bookmark) == 1:
701 701 bookmark = bookmark[0]
702 702 else:
703 703 bookmark = ''
704 704
705 705 oldphasemove = None
706 706 overrides = {(experimental, configbookmark): bookmark}
707 707
708 708 with ui.configoverride(overrides, 'infinitepush'):
709 709 scratchpush = opts.get('bundle_store')
710 710 if _scratchbranchmatcher(bookmark):
711 711 scratchpush = True
712 712 # bundle2 can be sent back after push (for example, bundle2
713 713 # containing `pushkey` part to update bookmarks)
714 714 ui.setconfig(experimental, 'bundle2.pushback', True)
715 715
716 716 if scratchpush:
717 717 # this is an infinitepush, we don't want the bookmark to be applied
718 718 # rather that should be stored in the bundlestore
719 719 opts['bookmark'] = []
720 720 ui.setconfig(experimental, configscratchpush, True)
721 721 oldphasemove = extensions.wrapfunction(exchange,
722 722 '_localphasemove',
723 723 _phasemove)
724 724 # Copy-paste from `push` command
725 725 path = ui.paths.getpath(dest, default=('default-push', 'default'))
726 726 if not path:
727 727 raise error.Abort(_('default repository not configured!'),
728 728 hint=_("see 'hg help config.paths'"))
729 729 destpath = path.pushloc or path.loc
730 730 # Remote scratch bookmarks will be deleted because remotenames doesn't
731 731 # know about them. Let's save it before push and restore after
732 732 remotescratchbookmarks = _readscratchremotebookmarks(ui, repo, destpath)
733 733 result = orig(ui, repo, dest, *args, **opts)
734 734 if common.isremotebooksenabled(ui):
735 735 if bookmark and scratchpush:
736 736 other = hg.peer(repo, opts, destpath)
737 737 fetchedbookmarks = other.listkeyspatterns('bookmarks',
738 738 patterns=[bookmark])
739 739 remotescratchbookmarks.update(fetchedbookmarks)
740 740 _saveremotebookmarks(repo, remotescratchbookmarks, destpath)
741 741 if oldphasemove:
742 742 exchange._localphasemove = oldphasemove
743 743 return result
744 744
745 745 def _deleteinfinitepushbookmarks(ui, repo, path, names):
746 746 """Prune remote names by removing the bookmarks we don't want anymore,
747 747 then writing the result back to disk
748 748 """
749 749 remotenamesext = extensions.find('remotenames')
750 750
751 751 # remotename format is:
752 752 # (node, nametype ("branches" or "bookmarks"), remote, name)
753 753 nametype_idx = 1
754 754 remote_idx = 2
755 755 name_idx = 3
756 756 remotenames = [remotename for remotename in \
757 757 remotenamesext.readremotenames(repo) \
758 758 if remotename[remote_idx] == path]
759 759 remote_bm_names = [remotename[name_idx] for remotename in \
760 760 remotenames if remotename[nametype_idx] == "bookmarks"]
761 761
762 762 for name in names:
763 763 if name not in remote_bm_names:
764 764 raise error.Abort(_("infinitepush bookmark '{}' does not exist "
765 765 "in path '{}'").format(name, path))
766 766
767 767 bookmarks = {}
768 768 branches = collections.defaultdict(list)
769 769 for node, nametype, remote, name in remotenames:
770 770 if nametype == "bookmarks" and name not in names:
771 771 bookmarks[name] = node
772 772 elif nametype == "branches":
773 773 # saveremotenames wants binary nodes for branches
774 774 branches[name].append(bin(node))
775 775
776 776 remotenamesext.saveremotenames(repo, path, branches, bookmarks)
777 777
778 778 def _phasemove(orig, pushop, nodes, phase=phases.public):
779 779 """prevent commits from being marked public
780 780
781 781 Since these are going to a scratch branch, they aren't really being
782 782 published."""
783 783
784 784 if phase != phases.public:
785 785 orig(pushop, nodes, phase)
786 786
787 787 @exchange.b2partsgenerator(scratchbranchparttype)
788 788 def partgen(pushop, bundler):
789 789 bookmark = pushop.ui.config(experimental, configbookmark)
790 790 scratchpush = pushop.ui.configbool(experimental, configscratchpush)
791 791 if 'changesets' in pushop.stepsdone or not scratchpush:
792 792 return
793 793
794 794 if scratchbranchparttype not in bundle2.bundle2caps(pushop.remote):
795 795 return
796 796
797 797 pushop.stepsdone.add('changesets')
798 798 if not pushop.outgoing.missing:
799 799 pushop.ui.status(_('no changes found\n'))
800 800 pushop.cgresult = 0
801 801 return
802 802
803 803 # This parameter tells the server that the following bundle is an
804 804 # infinitepush. This let's it switch the part processing to our infinitepush
805 805 # code path.
806 806 bundler.addparam("infinitepush", "True")
807 807
808 808 scratchparts = bundleparts.getscratchbranchparts(pushop.repo,
809 809 pushop.remote,
810 810 pushop.outgoing,
811 811 pushop.ui,
812 812 bookmark)
813 813
814 814 for scratchpart in scratchparts:
815 815 bundler.addpart(scratchpart)
816 816
817 817 def handlereply(op):
818 818 # server either succeeds or aborts; no code to read
819 819 pushop.cgresult = 1
820 820
821 821 return handlereply
822 822
823 823 bundle2.capabilities[bundleparts.scratchbranchparttype] = ()
824 824
825 825 def _getrevs(bundle, oldnode, force, bookmark):
826 826 'extracts and validates the revs to be imported'
827 827 revs = [bundle[r] for r in bundle.revs('sort(bundle())')]
828 828
829 829 # new bookmark
830 830 if oldnode is None:
831 831 return revs
832 832
833 833 # Fast forward update
834 834 if oldnode in bundle and list(bundle.set('bundle() & %s::', oldnode)):
835 835 return revs
836 836
837 837 return revs
838 838
839 839 @contextlib.contextmanager
840 840 def logservicecall(logger, service, **kwargs):
841 841 start = time.time()
842 842 logger(service, eventtype='start', **kwargs)
843 843 try:
844 844 yield
845 845 logger(service, eventtype='success',
846 846 elapsedms=(time.time() - start) * 1000, **kwargs)
847 847 except Exception as e:
848 848 logger(service, eventtype='failure',
849 849 elapsedms=(time.time() - start) * 1000, errormsg=str(e),
850 850 **kwargs)
851 851 raise
852 852
853 853 def _getorcreateinfinitepushlogger(op):
854 854 logger = op.records['infinitepushlogger']
855 855 if not logger:
856 856 ui = op.repo.ui
857 857 try:
858 858 username = procutil.getuser()
859 859 except Exception:
860 860 username = 'unknown'
861 861 # Generate random request id to be able to find all logged entries
862 862 # for the same request. Since requestid is pseudo-generated it may
863 863 # not be unique, but we assume that (hostname, username, requestid)
864 864 # is unique.
865 865 random.seed()
866 866 requestid = random.randint(0, 2000000000)
867 867 hostname = socket.gethostname()
868 868 logger = functools.partial(ui.log, 'infinitepush', user=username,
869 869 requestid=requestid, hostname=hostname,
870 870 reponame=ui.config('infinitepush',
871 871 'reponame'))
872 872 op.records.add('infinitepushlogger', logger)
873 873 else:
874 874 logger = logger[0]
875 875 return logger
876 876
877 877 def storetobundlestore(orig, repo, op, unbundler):
878 878 """stores the incoming bundle coming from push command to the bundlestore
879 879 instead of applying on the revlogs"""
880 880
881 881 repo.ui.status(_("storing changesets on the bundlestore\n"))
882 882 bundler = bundle2.bundle20(repo.ui)
883 883
884 884 # processing each part and storing it in bundler
885 885 with bundle2.partiterator(repo, op, unbundler) as parts:
886 886 for part in parts:
887 887 bundlepart = None
888 888 if part.type == 'replycaps':
889 889 # This configures the current operation to allow reply parts.
890 890 bundle2._processpart(op, part)
891 891 else:
892 892 bundlepart = bundle2.bundlepart(part.type, data=part.read())
893 893 for key, value in part.params.iteritems():
894 894 bundlepart.addparam(key, value)
895 895
896 896 # Certain parts require a response
897 897 if part.type in ('pushkey', 'changegroup'):
898 898 if op.reply is not None:
899 899 rpart = op.reply.newpart('reply:%s' % part.type)
900 900 rpart.addparam('in-reply-to', str(part.id),
901 901 mandatory=False)
902 902 rpart.addparam('return', '1', mandatory=False)
903 903
904 904 op.records.add(part.type, {
905 905 'return': 1,
906 906 })
907 907 if bundlepart:
908 908 bundler.addpart(bundlepart)
909 909
910 910 # storing the bundle in the bundlestore
911 911 buf = util.chunkbuffer(bundler.getchunks())
912 912 fd, bundlefile = tempfile.mkstemp()
913 913 try:
914 914 try:
915 915 fp = os.fdopen(fd, 'wb')
916 916 fp.write(buf.read())
917 917 finally:
918 918 fp.close()
919 919 storebundle(op, {}, bundlefile)
920 920 finally:
921 921 try:
922 922 os.unlink(bundlefile)
923 923 except Exception:
924 924 # we would rather see the original exception
925 925 pass
926 926
927 927 def processparts(orig, repo, op, unbundler):
928 928
929 929 # make sure we don't wrap processparts in case of `hg unbundle`
930 930 tr = repo.currenttransaction()
931 931 if tr:
932 932 if tr.names[0].startswith('unbundle'):
933 933 return orig(repo, op, unbundler)
934 934
935 935 # this server routes each push to bundle store
936 936 if repo.ui.configbool('infinitepush', 'pushtobundlestore'):
937 937 return storetobundlestore(orig, repo, op, unbundler)
938 938
939 939 if unbundler.params.get('infinitepush') != 'True':
940 940 return orig(repo, op, unbundler)
941 941
942 942 handleallparts = repo.ui.configbool('infinitepush', 'storeallparts')
943 943
944 944 bundler = bundle2.bundle20(repo.ui)
945 945 cgparams = None
946 946 with bundle2.partiterator(repo, op, unbundler) as parts:
947 947 for part in parts:
948 948 bundlepart = None
949 949 if part.type == 'replycaps':
950 950 # This configures the current operation to allow reply parts.
951 951 bundle2._processpart(op, part)
952 952 elif part.type == bundleparts.scratchbranchparttype:
953 953 # Scratch branch parts need to be converted to normal
954 954 # changegroup parts, and the extra parameters stored for later
955 955 # when we upload to the store. Eventually those parameters will
956 956 # be put on the actual bundle instead of this part, then we can
957 957 # send a vanilla changegroup instead of the scratchbranch part.
958 958 cgversion = part.params.get('cgversion', '01')
959 959 bundlepart = bundle2.bundlepart('changegroup', data=part.read())
960 960 bundlepart.addparam('version', cgversion)
961 961 cgparams = part.params
962 962
963 963 # If we're not dumping all parts into the new bundle, we need to
964 964 # alert the future pushkey and phase-heads handler to skip
965 965 # the part.
966 966 if not handleallparts:
967 967 op.records.add(scratchbranchparttype + '_skippushkey', True)
968 968 op.records.add(scratchbranchparttype + '_skipphaseheads',
969 969 True)
970 970 else:
971 971 if handleallparts:
972 972 # Ideally we would not process any parts, and instead just
973 973 # forward them to the bundle for storage, but since this
974 974 # differs from previous behavior, we need to put it behind a
975 975 # config flag for incremental rollout.
976 976 bundlepart = bundle2.bundlepart(part.type, data=part.read())
977 977 for key, value in part.params.iteritems():
978 978 bundlepart.addparam(key, value)
979 979
980 980 # Certain parts require a response
981 981 if part.type == 'pushkey':
982 982 if op.reply is not None:
983 983 rpart = op.reply.newpart('reply:pushkey')
984 984 rpart.addparam('in-reply-to', str(part.id),
985 985 mandatory=False)
986 986 rpart.addparam('return', '1', mandatory=False)
987 987 else:
988 988 bundle2._processpart(op, part)
989 989
990 990 if handleallparts:
991 991 op.records.add(part.type, {
992 992 'return': 1,
993 993 })
994 994 if bundlepart:
995 995 bundler.addpart(bundlepart)
996 996
997 997 # If commits were sent, store them
998 998 if cgparams:
999 999 buf = util.chunkbuffer(bundler.getchunks())
1000 1000 fd, bundlefile = tempfile.mkstemp()
1001 1001 try:
1002 1002 try:
1003 1003 fp = os.fdopen(fd, 'wb')
1004 1004 fp.write(buf.read())
1005 1005 finally:
1006 1006 fp.close()
1007 1007 storebundle(op, cgparams, bundlefile)
1008 1008 finally:
1009 1009 try:
1010 1010 os.unlink(bundlefile)
1011 1011 except Exception:
1012 1012 # we would rather see the original exception
1013 1013 pass
1014 1014
1015 1015 def storebundle(op, params, bundlefile):
1016 1016 log = _getorcreateinfinitepushlogger(op)
1017 1017 parthandlerstart = time.time()
1018 1018 log(scratchbranchparttype, eventtype='start')
1019 1019 index = op.repo.bundlestore.index
1020 1020 store = op.repo.bundlestore.store
1021 1021 op.records.add(scratchbranchparttype + '_skippushkey', True)
1022 1022
1023 1023 bundle = None
1024 1024 try: # guards bundle
1025 1025 bundlepath = "bundle:%s+%s" % (op.repo.root, bundlefile)
1026 1026 bundle = hg.repository(op.repo.ui, bundlepath)
1027 1027
1028 1028 bookmark = params.get('bookmark')
1029 1029 bookprevnode = params.get('bookprevnode', '')
1030 1030 force = params.get('force')
1031 1031
1032 1032 if bookmark:
1033 1033 oldnode = index.getnode(bookmark)
1034 1034 else:
1035 1035 oldnode = None
1036 1036 bundleheads = bundle.revs('heads(bundle())')
1037 1037 if bookmark and len(bundleheads) > 1:
1038 1038 raise error.Abort(
1039 1039 _('cannot push more than one head to a scratch branch'))
1040 1040
1041 1041 revs = _getrevs(bundle, oldnode, force, bookmark)
1042 1042
1043 1043 # Notify the user of what is being pushed
1044 1044 plural = 's' if len(revs) > 1 else ''
1045 1045 op.repo.ui.warn(_("pushing %s commit%s:\n") % (len(revs), plural))
1046 1046 maxoutput = 10
1047 1047 for i in range(0, min(len(revs), maxoutput)):
1048 1048 firstline = bundle[revs[i]].description().split('\n')[0][:50]
1049 1049 op.repo.ui.warn((" %s %s\n") % (revs[i], firstline))
1050 1050
1051 1051 if len(revs) > maxoutput + 1:
1052 1052 op.repo.ui.warn((" ...\n"))
1053 1053 firstline = bundle[revs[-1]].description().split('\n')[0][:50]
1054 1054 op.repo.ui.warn((" %s %s\n") % (revs[-1], firstline))
1055 1055
1056 1056 nodesctx = [bundle[rev] for rev in revs]
1057 1057 inindex = lambda rev: bool(index.getbundle(bundle[rev].hex()))
1058 1058 if bundleheads:
1059 1059 newheadscount = sum(not inindex(rev) for rev in bundleheads)
1060 1060 else:
1061 1061 newheadscount = 0
1062 1062 # If there's a bookmark specified, there should be only one head,
1063 1063 # so we choose the last node, which will be that head.
1064 1064 # If a bug or malicious client allows there to be a bookmark
1065 1065 # with multiple heads, we will place the bookmark on the last head.
1066 1066 bookmarknode = nodesctx[-1].hex() if nodesctx else None
1067 1067 key = None
1068 1068 if newheadscount:
1069 1069 with open(bundlefile, 'r') as f:
1070 1070 bundledata = f.read()
1071 1071 with logservicecall(log, 'bundlestore',
1072 1072 bundlesize=len(bundledata)):
1073 1073 bundlesizelimit = 100 * 1024 * 1024 # 100 MB
1074 1074 if len(bundledata) > bundlesizelimit:
1075 1075 error_msg = ('bundle is too big: %d bytes. ' +
1076 1076 'max allowed size is 100 MB')
1077 1077 raise error.Abort(error_msg % (len(bundledata),))
1078 1078 key = store.write(bundledata)
1079 1079
1080 1080 with logservicecall(log, 'index', newheadscount=newheadscount), index:
1081 1081 if key:
1082 1082 index.addbundle(key, nodesctx)
1083 1083 if bookmark:
1084 1084 index.addbookmark(bookmark, bookmarknode)
1085 1085 _maybeaddpushbackpart(op, bookmark, bookmarknode,
1086 1086 bookprevnode, params)
1087 1087 log(scratchbranchparttype, eventtype='success',
1088 1088 elapsedms=(time.time() - parthandlerstart) * 1000)
1089 1089
1090 1090 except Exception as e:
1091 1091 log(scratchbranchparttype, eventtype='failure',
1092 1092 elapsedms=(time.time() - parthandlerstart) * 1000,
1093 1093 errormsg=str(e))
1094 1094 raise
1095 1095 finally:
1096 1096 if bundle:
1097 1097 bundle.close()
1098 1098
1099 1099 @bundle2.parthandler(scratchbranchparttype,
1100 1100 ('bookmark', 'bookprevnode', 'force',
1101 1101 'pushbackbookmarks', 'cgversion'))
1102 1102 def bundle2scratchbranch(op, part):
1103 1103 '''unbundle a bundle2 part containing a changegroup to store'''
1104 1104
1105 1105 bundler = bundle2.bundle20(op.repo.ui)
1106 1106 cgversion = part.params.get('cgversion', '01')
1107 1107 cgpart = bundle2.bundlepart('changegroup', data=part.read())
1108 1108 cgpart.addparam('version', cgversion)
1109 1109 bundler.addpart(cgpart)
1110 1110 buf = util.chunkbuffer(bundler.getchunks())
1111 1111
1112 1112 fd, bundlefile = tempfile.mkstemp()
1113 1113 try:
1114 1114 try:
1115 1115 fp = os.fdopen(fd, 'wb')
1116 1116 fp.write(buf.read())
1117 1117 finally:
1118 1118 fp.close()
1119 1119 storebundle(op, part.params, bundlefile)
1120 1120 finally:
1121 1121 try:
1122 1122 os.unlink(bundlefile)
1123 1123 except OSError as e:
1124 1124 if e.errno != errno.ENOENT:
1125 1125 raise
1126 1126
1127 1127 return 1
1128 1128
1129 1129 def _maybeaddpushbackpart(op, bookmark, newnode, oldnode, params):
1130 1130 if params.get('pushbackbookmarks'):
1131 1131 if op.reply and 'pushback' in op.reply.capabilities:
1132 1132 params = {
1133 1133 'namespace': 'bookmarks',
1134 1134 'key': bookmark,
1135 1135 'new': newnode,
1136 1136 'old': oldnode,
1137 1137 }
1138 1138 op.reply.newpart('pushkey', mandatoryparams=params.iteritems())
1139 1139
1140 1140 def bundle2pushkey(orig, op, part):
1141 1141 '''Wrapper of bundle2.handlepushkey()
1142 1142
1143 1143 The only goal is to skip calling the original function if flag is set.
1144 1144 It's set if infinitepush push is happening.
1145 1145 '''
1146 1146 if op.records[scratchbranchparttype + '_skippushkey']:
1147 1147 if op.reply is not None:
1148 1148 rpart = op.reply.newpart('reply:pushkey')
1149 1149 rpart.addparam('in-reply-to', str(part.id), mandatory=False)
1150 1150 rpart.addparam('return', '1', mandatory=False)
1151 1151 return 1
1152 1152
1153 1153 return orig(op, part)
1154 1154
1155 1155 def bundle2handlephases(orig, op, part):
1156 1156 '''Wrapper of bundle2.handlephases()
1157 1157
1158 1158 The only goal is to skip calling the original function if flag is set.
1159 1159 It's set if infinitepush push is happening.
1160 1160 '''
1161 1161
1162 1162 if op.records[scratchbranchparttype + '_skipphaseheads']:
1163 1163 return
1164 1164
1165 1165 return orig(op, part)
1166 1166
1167 1167 def _asyncsavemetadata(root, nodes):
1168 1168 '''starts a separate process that fills metadata for the nodes
1169 1169
1170 1170 This function creates a separate process and doesn't wait for it's
1171 1171 completion. This was done to avoid slowing down pushes
1172 1172 '''
1173 1173
1174 1174 maxnodes = 50
1175 1175 if len(nodes) > maxnodes:
1176 1176 return
1177 1177 nodesargs = []
1178 1178 for node in nodes:
1179 1179 nodesargs.append('--node')
1180 1180 nodesargs.append(node)
1181 1181 with open(os.devnull, 'w+b') as devnull:
1182 1182 cmdline = [util.hgexecutable(), 'debugfillinfinitepushmetadata',
1183 1183 '-R', root] + nodesargs
1184 1184 # Process will run in background. We don't care about the return code
1185 1185 subprocess.Popen(cmdline, close_fds=True, shell=False,
1186 1186 stdin=devnull, stdout=devnull, stderr=devnull)
General Comments 0
You need to be logged in to leave comments. Login now