##// END OF EJS Templates
py3: use pycompat.{strkwargs|byteskwargs} in infinitepush...
Pulkit Goyal -
r37595:e7eea858 default
parent child Browse files
Show More
@@ -1,1184 +1,1186 b''
1 # Infinite push
1 # Infinite push
2 #
2 #
3 # Copyright 2016 Facebook, Inc.
3 # Copyright 2016 Facebook, Inc.
4 #
4 #
5 # This software may be used and distributed according to the terms of the
5 # This software may be used and distributed according to the terms of the
6 # GNU General Public License version 2 or any later version.
6 # GNU General Public License version 2 or any later version.
7 """ store some pushes in a remote blob store on the server (EXPERIMENTAL)
7 """ store some pushes in a remote blob store on the server (EXPERIMENTAL)
8
8
9 [infinitepush]
9 [infinitepush]
10 # Server-side and client-side option. Pattern of the infinitepush bookmark
10 # Server-side and client-side option. Pattern of the infinitepush bookmark
11 branchpattern = PATTERN
11 branchpattern = PATTERN
12
12
13 # Server or client
13 # Server or client
14 server = False
14 server = False
15
15
16 # Server-side option. Possible values: 'disk' or 'sql'. Fails if not set
16 # Server-side option. Possible values: 'disk' or 'sql'. Fails if not set
17 indextype = disk
17 indextype = disk
18
18
19 # Server-side option. Used only if indextype=sql.
19 # Server-side option. Used only if indextype=sql.
20 # Format: 'IP:PORT:DB_NAME:USER:PASSWORD'
20 # Format: 'IP:PORT:DB_NAME:USER:PASSWORD'
21 sqlhost = IP:PORT:DB_NAME:USER:PASSWORD
21 sqlhost = IP:PORT:DB_NAME:USER:PASSWORD
22
22
23 # Server-side option. Used only if indextype=disk.
23 # Server-side option. Used only if indextype=disk.
24 # Filesystem path to the index store
24 # Filesystem path to the index store
25 indexpath = PATH
25 indexpath = PATH
26
26
27 # Server-side option. Possible values: 'disk' or 'external'
27 # Server-side option. Possible values: 'disk' or 'external'
28 # Fails if not set
28 # Fails if not set
29 storetype = disk
29 storetype = disk
30
30
31 # Server-side option.
31 # Server-side option.
32 # Path to the binary that will save bundle to the bundlestore
32 # Path to the binary that will save bundle to the bundlestore
33 # Formatted cmd line will be passed to it (see `put_args`)
33 # Formatted cmd line will be passed to it (see `put_args`)
34 put_binary = put
34 put_binary = put
35
35
36 # Serser-side option. Used only if storetype=external.
36 # Serser-side option. Used only if storetype=external.
37 # Format cmd-line string for put binary. Placeholder: {filename}
37 # Format cmd-line string for put binary. Placeholder: {filename}
38 put_args = {filename}
38 put_args = {filename}
39
39
40 # Server-side option.
40 # Server-side option.
41 # Path to the binary that get bundle from the bundlestore.
41 # Path to the binary that get bundle from the bundlestore.
42 # Formatted cmd line will be passed to it (see `get_args`)
42 # Formatted cmd line will be passed to it (see `get_args`)
43 get_binary = get
43 get_binary = get
44
44
45 # Serser-side option. Used only if storetype=external.
45 # Serser-side option. Used only if storetype=external.
46 # Format cmd-line string for get binary. Placeholders: {filename} {handle}
46 # Format cmd-line string for get binary. Placeholders: {filename} {handle}
47 get_args = {filename} {handle}
47 get_args = {filename} {handle}
48
48
49 # Server-side option
49 # Server-side option
50 logfile = FIlE
50 logfile = FIlE
51
51
52 # Server-side option
52 # Server-side option
53 loglevel = DEBUG
53 loglevel = DEBUG
54
54
55 # Server-side option. Used only if indextype=sql.
55 # Server-side option. Used only if indextype=sql.
56 # Sets mysql wait_timeout option.
56 # Sets mysql wait_timeout option.
57 waittimeout = 300
57 waittimeout = 300
58
58
59 # Server-side option. Used only if indextype=sql.
59 # Server-side option. Used only if indextype=sql.
60 # Sets mysql innodb_lock_wait_timeout option.
60 # Sets mysql innodb_lock_wait_timeout option.
61 locktimeout = 120
61 locktimeout = 120
62
62
63 # Server-side option. Used only if indextype=sql.
63 # Server-side option. Used only if indextype=sql.
64 # Name of the repository
64 # Name of the repository
65 reponame = ''
65 reponame = ''
66
66
67 # Client-side option. Used by --list-remote option. List of remote scratch
67 # Client-side option. Used by --list-remote option. List of remote scratch
68 # patterns to list if no patterns are specified.
68 # patterns to list if no patterns are specified.
69 defaultremotepatterns = ['*']
69 defaultremotepatterns = ['*']
70
70
71 # Instructs infinitepush to forward all received bundle2 parts to the
71 # Instructs infinitepush to forward all received bundle2 parts to the
72 # bundle for storage. Defaults to False.
72 # bundle for storage. Defaults to False.
73 storeallparts = True
73 storeallparts = True
74
74
75 # routes each incoming push to the bundlestore. defaults to False
75 # routes each incoming push to the bundlestore. defaults to False
76 pushtobundlestore = True
76 pushtobundlestore = True
77
77
78 [remotenames]
78 [remotenames]
79 # Client-side option
79 # Client-side option
80 # This option should be set only if remotenames extension is enabled.
80 # This option should be set only if remotenames extension is enabled.
81 # Whether remote bookmarks are tracked by remotenames extension.
81 # Whether remote bookmarks are tracked by remotenames extension.
82 bookmarks = True
82 bookmarks = True
83 """
83 """
84
84
85 from __future__ import absolute_import
85 from __future__ import absolute_import
86
86
87 import collections
87 import collections
88 import contextlib
88 import contextlib
89 import errno
89 import errno
90 import functools
90 import functools
91 import logging
91 import logging
92 import os
92 import os
93 import random
93 import random
94 import re
94 import re
95 import socket
95 import socket
96 import subprocess
96 import subprocess
97 import tempfile
97 import tempfile
98 import time
98 import time
99
99
100 from mercurial.node import (
100 from mercurial.node import (
101 bin,
101 bin,
102 hex,
102 hex,
103 )
103 )
104
104
105 from mercurial.i18n import _
105 from mercurial.i18n import _
106
106
107 from mercurial.utils import (
107 from mercurial.utils import (
108 procutil,
108 procutil,
109 stringutil,
109 stringutil,
110 )
110 )
111
111
112 from mercurial import (
112 from mercurial import (
113 bundle2,
113 bundle2,
114 changegroup,
114 changegroup,
115 commands,
115 commands,
116 discovery,
116 discovery,
117 encoding,
117 encoding,
118 error,
118 error,
119 exchange,
119 exchange,
120 extensions,
120 extensions,
121 hg,
121 hg,
122 localrepo,
122 localrepo,
123 peer,
123 peer,
124 phases,
124 phases,
125 pushkey,
125 pushkey,
126 pycompat,
126 registrar,
127 registrar,
127 util,
128 util,
128 wireproto,
129 wireproto,
129 )
130 )
130
131
131 from . import (
132 from . import (
132 bundleparts,
133 bundleparts,
133 common,
134 common,
134 )
135 )
135
136
136 # Note for extension authors: ONLY specify testedwith = 'ships-with-hg-core' for
137 # Note for extension authors: ONLY specify testedwith = 'ships-with-hg-core' for
137 # extensions which SHIP WITH MERCURIAL. Non-mainline extensions should
138 # extensions which SHIP WITH MERCURIAL. Non-mainline extensions should
138 # be specifying the version(s) of Mercurial they are tested with, or
139 # be specifying the version(s) of Mercurial they are tested with, or
139 # leave the attribute unspecified.
140 # leave the attribute unspecified.
140 testedwith = 'ships-with-hg-core'
141 testedwith = 'ships-with-hg-core'
141
142
142 configtable = {}
143 configtable = {}
143 configitem = registrar.configitem(configtable)
144 configitem = registrar.configitem(configtable)
144
145
145 configitem('infinitepush', 'server',
146 configitem('infinitepush', 'server',
146 default=False,
147 default=False,
147 )
148 )
148 configitem('infinitepush', 'storetype',
149 configitem('infinitepush', 'storetype',
149 default='',
150 default='',
150 )
151 )
151 configitem('infinitepush', 'indextype',
152 configitem('infinitepush', 'indextype',
152 default='',
153 default='',
153 )
154 )
154 configitem('infinitepush', 'indexpath',
155 configitem('infinitepush', 'indexpath',
155 default='',
156 default='',
156 )
157 )
157 configitem('infinitepush', 'storeallparts',
158 configitem('infinitepush', 'storeallparts',
158 default=False,
159 default=False,
159 )
160 )
160 configitem('infinitepush', 'reponame',
161 configitem('infinitepush', 'reponame',
161 default='',
162 default='',
162 )
163 )
163 configitem('scratchbranch', 'storepath',
164 configitem('scratchbranch', 'storepath',
164 default='',
165 default='',
165 )
166 )
166 configitem('infinitepush', 'branchpattern',
167 configitem('infinitepush', 'branchpattern',
167 default='',
168 default='',
168 )
169 )
169 configitem('infinitepush', 'pushtobundlestore',
170 configitem('infinitepush', 'pushtobundlestore',
170 default=False,
171 default=False,
171 )
172 )
172 configitem('experimental', 'server-bundlestore-bookmark',
173 configitem('experimental', 'server-bundlestore-bookmark',
173 default='',
174 default='',
174 )
175 )
175 configitem('experimental', 'infinitepush-scratchpush',
176 configitem('experimental', 'infinitepush-scratchpush',
176 default=False,
177 default=False,
177 )
178 )
178
179
179 experimental = 'experimental'
180 experimental = 'experimental'
180 configbookmark = 'server-bundlestore-bookmark'
181 configbookmark = 'server-bundlestore-bookmark'
181 configscratchpush = 'infinitepush-scratchpush'
182 configscratchpush = 'infinitepush-scratchpush'
182
183
183 scratchbranchparttype = bundleparts.scratchbranchparttype
184 scratchbranchparttype = bundleparts.scratchbranchparttype
184 revsetpredicate = registrar.revsetpredicate()
185 revsetpredicate = registrar.revsetpredicate()
185 templatekeyword = registrar.templatekeyword()
186 templatekeyword = registrar.templatekeyword()
186 _scratchbranchmatcher = lambda x: False
187 _scratchbranchmatcher = lambda x: False
187 _maybehash = re.compile(r'^[a-f0-9]+$').search
188 _maybehash = re.compile(r'^[a-f0-9]+$').search
188
189
189 def _buildexternalbundlestore(ui):
190 def _buildexternalbundlestore(ui):
190 put_args = ui.configlist('infinitepush', 'put_args', [])
191 put_args = ui.configlist('infinitepush', 'put_args', [])
191 put_binary = ui.config('infinitepush', 'put_binary')
192 put_binary = ui.config('infinitepush', 'put_binary')
192 if not put_binary:
193 if not put_binary:
193 raise error.Abort('put binary is not specified')
194 raise error.Abort('put binary is not specified')
194 get_args = ui.configlist('infinitepush', 'get_args', [])
195 get_args = ui.configlist('infinitepush', 'get_args', [])
195 get_binary = ui.config('infinitepush', 'get_binary')
196 get_binary = ui.config('infinitepush', 'get_binary')
196 if not get_binary:
197 if not get_binary:
197 raise error.Abort('get binary is not specified')
198 raise error.Abort('get binary is not specified')
198 from . import store
199 from . import store
199 return store.externalbundlestore(put_binary, put_args, get_binary, get_args)
200 return store.externalbundlestore(put_binary, put_args, get_binary, get_args)
200
201
201 def _buildsqlindex(ui):
202 def _buildsqlindex(ui):
202 sqlhost = ui.config('infinitepush', 'sqlhost')
203 sqlhost = ui.config('infinitepush', 'sqlhost')
203 if not sqlhost:
204 if not sqlhost:
204 raise error.Abort(_('please set infinitepush.sqlhost'))
205 raise error.Abort(_('please set infinitepush.sqlhost'))
205 host, port, db, user, password = sqlhost.split(':')
206 host, port, db, user, password = sqlhost.split(':')
206 reponame = ui.config('infinitepush', 'reponame')
207 reponame = ui.config('infinitepush', 'reponame')
207 if not reponame:
208 if not reponame:
208 raise error.Abort(_('please set infinitepush.reponame'))
209 raise error.Abort(_('please set infinitepush.reponame'))
209
210
210 logfile = ui.config('infinitepush', 'logfile', '')
211 logfile = ui.config('infinitepush', 'logfile', '')
211 waittimeout = ui.configint('infinitepush', 'waittimeout', 300)
212 waittimeout = ui.configint('infinitepush', 'waittimeout', 300)
212 locktimeout = ui.configint('infinitepush', 'locktimeout', 120)
213 locktimeout = ui.configint('infinitepush', 'locktimeout', 120)
213 from . import sqlindexapi
214 from . import sqlindexapi
214 return sqlindexapi.sqlindexapi(
215 return sqlindexapi.sqlindexapi(
215 reponame, host, port, db, user, password,
216 reponame, host, port, db, user, password,
216 logfile, _getloglevel(ui), waittimeout=waittimeout,
217 logfile, _getloglevel(ui), waittimeout=waittimeout,
217 locktimeout=locktimeout)
218 locktimeout=locktimeout)
218
219
219 def _getloglevel(ui):
220 def _getloglevel(ui):
220 loglevel = ui.config('infinitepush', 'loglevel', 'DEBUG')
221 loglevel = ui.config('infinitepush', 'loglevel', 'DEBUG')
221 numeric_loglevel = getattr(logging, loglevel.upper(), None)
222 numeric_loglevel = getattr(logging, loglevel.upper(), None)
222 if not isinstance(numeric_loglevel, int):
223 if not isinstance(numeric_loglevel, int):
223 raise error.Abort(_('invalid log level %s') % loglevel)
224 raise error.Abort(_('invalid log level %s') % loglevel)
224 return numeric_loglevel
225 return numeric_loglevel
225
226
226 def _tryhoist(ui, remotebookmark):
227 def _tryhoist(ui, remotebookmark):
227 '''returns a bookmarks with hoisted part removed
228 '''returns a bookmarks with hoisted part removed
228
229
229 Remotenames extension has a 'hoist' config that allows to use remote
230 Remotenames extension has a 'hoist' config that allows to use remote
230 bookmarks without specifying remote path. For example, 'hg update master'
231 bookmarks without specifying remote path. For example, 'hg update master'
231 works as well as 'hg update remote/master'. We want to allow the same in
232 works as well as 'hg update remote/master'. We want to allow the same in
232 infinitepush.
233 infinitepush.
233 '''
234 '''
234
235
235 if common.isremotebooksenabled(ui):
236 if common.isremotebooksenabled(ui):
236 hoist = ui.config('remotenames', 'hoistedpeer') + '/'
237 hoist = ui.config('remotenames', 'hoistedpeer') + '/'
237 if remotebookmark.startswith(hoist):
238 if remotebookmark.startswith(hoist):
238 return remotebookmark[len(hoist):]
239 return remotebookmark[len(hoist):]
239 return remotebookmark
240 return remotebookmark
240
241
241 class bundlestore(object):
242 class bundlestore(object):
242 def __init__(self, repo):
243 def __init__(self, repo):
243 self._repo = repo
244 self._repo = repo
244 storetype = self._repo.ui.config('infinitepush', 'storetype')
245 storetype = self._repo.ui.config('infinitepush', 'storetype')
245 if storetype == 'disk':
246 if storetype == 'disk':
246 from . import store
247 from . import store
247 self.store = store.filebundlestore(self._repo.ui, self._repo)
248 self.store = store.filebundlestore(self._repo.ui, self._repo)
248 elif storetype == 'external':
249 elif storetype == 'external':
249 self.store = _buildexternalbundlestore(self._repo.ui)
250 self.store = _buildexternalbundlestore(self._repo.ui)
250 else:
251 else:
251 raise error.Abort(
252 raise error.Abort(
252 _('unknown infinitepush store type specified %s') % storetype)
253 _('unknown infinitepush store type specified %s') % storetype)
253
254
254 indextype = self._repo.ui.config('infinitepush', 'indextype')
255 indextype = self._repo.ui.config('infinitepush', 'indextype')
255 if indextype == 'disk':
256 if indextype == 'disk':
256 from . import fileindexapi
257 from . import fileindexapi
257 self.index = fileindexapi.fileindexapi(self._repo)
258 self.index = fileindexapi.fileindexapi(self._repo)
258 elif indextype == 'sql':
259 elif indextype == 'sql':
259 self.index = _buildsqlindex(self._repo.ui)
260 self.index = _buildsqlindex(self._repo.ui)
260 else:
261 else:
261 raise error.Abort(
262 raise error.Abort(
262 _('unknown infinitepush index type specified %s') % indextype)
263 _('unknown infinitepush index type specified %s') % indextype)
263
264
264 def _isserver(ui):
265 def _isserver(ui):
265 return ui.configbool('infinitepush', 'server')
266 return ui.configbool('infinitepush', 'server')
266
267
267 def reposetup(ui, repo):
268 def reposetup(ui, repo):
268 if _isserver(ui) and repo.local():
269 if _isserver(ui) and repo.local():
269 repo.bundlestore = bundlestore(repo)
270 repo.bundlestore = bundlestore(repo)
270
271
271 def extsetup(ui):
272 def extsetup(ui):
272 commonsetup(ui)
273 commonsetup(ui)
273 if _isserver(ui):
274 if _isserver(ui):
274 serverextsetup(ui)
275 serverextsetup(ui)
275 else:
276 else:
276 clientextsetup(ui)
277 clientextsetup(ui)
277
278
278 def commonsetup(ui):
279 def commonsetup(ui):
279 wireproto.commands['listkeyspatterns'] = (
280 wireproto.commands['listkeyspatterns'] = (
280 wireprotolistkeyspatterns, 'namespace patterns')
281 wireprotolistkeyspatterns, 'namespace patterns')
281 scratchbranchpat = ui.config('infinitepush', 'branchpattern')
282 scratchbranchpat = ui.config('infinitepush', 'branchpattern')
282 if scratchbranchpat:
283 if scratchbranchpat:
283 global _scratchbranchmatcher
284 global _scratchbranchmatcher
284 kind, pat, _scratchbranchmatcher = \
285 kind, pat, _scratchbranchmatcher = \
285 stringutil.stringmatcher(scratchbranchpat)
286 stringutil.stringmatcher(scratchbranchpat)
286
287
287 def serverextsetup(ui):
288 def serverextsetup(ui):
288 origpushkeyhandler = bundle2.parthandlermapping['pushkey']
289 origpushkeyhandler = bundle2.parthandlermapping['pushkey']
289
290
290 def newpushkeyhandler(*args, **kwargs):
291 def newpushkeyhandler(*args, **kwargs):
291 bundle2pushkey(origpushkeyhandler, *args, **kwargs)
292 bundle2pushkey(origpushkeyhandler, *args, **kwargs)
292 newpushkeyhandler.params = origpushkeyhandler.params
293 newpushkeyhandler.params = origpushkeyhandler.params
293 bundle2.parthandlermapping['pushkey'] = newpushkeyhandler
294 bundle2.parthandlermapping['pushkey'] = newpushkeyhandler
294
295
295 orighandlephasehandler = bundle2.parthandlermapping['phase-heads']
296 orighandlephasehandler = bundle2.parthandlermapping['phase-heads']
296 newphaseheadshandler = lambda *args, **kwargs: \
297 newphaseheadshandler = lambda *args, **kwargs: \
297 bundle2handlephases(orighandlephasehandler, *args, **kwargs)
298 bundle2handlephases(orighandlephasehandler, *args, **kwargs)
298 newphaseheadshandler.params = orighandlephasehandler.params
299 newphaseheadshandler.params = orighandlephasehandler.params
299 bundle2.parthandlermapping['phase-heads'] = newphaseheadshandler
300 bundle2.parthandlermapping['phase-heads'] = newphaseheadshandler
300
301
301 extensions.wrapfunction(localrepo.localrepository, 'listkeys',
302 extensions.wrapfunction(localrepo.localrepository, 'listkeys',
302 localrepolistkeys)
303 localrepolistkeys)
303 wireproto.commands['lookup'] = (
304 wireproto.commands['lookup'] = (
304 _lookupwrap(wireproto.commands['lookup'][0]), 'key')
305 _lookupwrap(wireproto.commands['lookup'][0]), 'key')
305 extensions.wrapfunction(exchange, 'getbundlechunks', getbundlechunks)
306 extensions.wrapfunction(exchange, 'getbundlechunks', getbundlechunks)
306
307
307 extensions.wrapfunction(bundle2, 'processparts', processparts)
308 extensions.wrapfunction(bundle2, 'processparts', processparts)
308
309
309 def clientextsetup(ui):
310 def clientextsetup(ui):
310 entry = extensions.wrapcommand(commands.table, 'push', _push)
311 entry = extensions.wrapcommand(commands.table, 'push', _push)
311
312
312 entry[1].append(
313 entry[1].append(
313 ('', 'bundle-store', None,
314 ('', 'bundle-store', None,
314 _('force push to go to bundle store (EXPERIMENTAL)')))
315 _('force push to go to bundle store (EXPERIMENTAL)')))
315
316
316 extensions.wrapcommand(commands.table, 'pull', _pull)
317 extensions.wrapcommand(commands.table, 'pull', _pull)
317
318
318 extensions.wrapfunction(discovery, 'checkheads', _checkheads)
319 extensions.wrapfunction(discovery, 'checkheads', _checkheads)
319
320
320 wireproto.wirepeer.listkeyspatterns = listkeyspatterns
321 wireproto.wirepeer.listkeyspatterns = listkeyspatterns
321
322
322 partorder = exchange.b2partsgenorder
323 partorder = exchange.b2partsgenorder
323 index = partorder.index('changeset')
324 index = partorder.index('changeset')
324 partorder.insert(
325 partorder.insert(
325 index, partorder.pop(partorder.index(scratchbranchparttype)))
326 index, partorder.pop(partorder.index(scratchbranchparttype)))
326
327
327 def _checkheads(orig, pushop):
328 def _checkheads(orig, pushop):
328 if pushop.ui.configbool(experimental, configscratchpush, False):
329 if pushop.ui.configbool(experimental, configscratchpush, False):
329 return
330 return
330 return orig(pushop)
331 return orig(pushop)
331
332
332 def wireprotolistkeyspatterns(repo, proto, namespace, patterns):
333 def wireprotolistkeyspatterns(repo, proto, namespace, patterns):
333 patterns = wireproto.decodelist(patterns)
334 patterns = wireproto.decodelist(patterns)
334 d = repo.listkeys(encoding.tolocal(namespace), patterns).iteritems()
335 d = repo.listkeys(encoding.tolocal(namespace), patterns).iteritems()
335 return pushkey.encodekeys(d)
336 return pushkey.encodekeys(d)
336
337
337 def localrepolistkeys(orig, self, namespace, patterns=None):
338 def localrepolistkeys(orig, self, namespace, patterns=None):
338 if namespace == 'bookmarks' and patterns:
339 if namespace == 'bookmarks' and patterns:
339 index = self.bundlestore.index
340 index = self.bundlestore.index
340 results = {}
341 results = {}
341 bookmarks = orig(self, namespace)
342 bookmarks = orig(self, namespace)
342 for pattern in patterns:
343 for pattern in patterns:
343 results.update(index.getbookmarks(pattern))
344 results.update(index.getbookmarks(pattern))
344 if pattern.endswith('*'):
345 if pattern.endswith('*'):
345 pattern = 're:^' + pattern[:-1] + '.*'
346 pattern = 're:^' + pattern[:-1] + '.*'
346 kind, pat, matcher = stringutil.stringmatcher(pattern)
347 kind, pat, matcher = stringutil.stringmatcher(pattern)
347 for bookmark, node in bookmarks.iteritems():
348 for bookmark, node in bookmarks.iteritems():
348 if matcher(bookmark):
349 if matcher(bookmark):
349 results[bookmark] = node
350 results[bookmark] = node
350 return results
351 return results
351 else:
352 else:
352 return orig(self, namespace)
353 return orig(self, namespace)
353
354
354 @peer.batchable
355 @peer.batchable
355 def listkeyspatterns(self, namespace, patterns):
356 def listkeyspatterns(self, namespace, patterns):
356 if not self.capable('pushkey'):
357 if not self.capable('pushkey'):
357 yield {}, None
358 yield {}, None
358 f = peer.future()
359 f = peer.future()
359 self.ui.debug('preparing listkeys for "%s" with pattern "%s"\n' %
360 self.ui.debug('preparing listkeys for "%s" with pattern "%s"\n' %
360 (namespace, patterns))
361 (namespace, patterns))
361 yield {
362 yield {
362 'namespace': encoding.fromlocal(namespace),
363 'namespace': encoding.fromlocal(namespace),
363 'patterns': wireproto.encodelist(patterns)
364 'patterns': wireproto.encodelist(patterns)
364 }, f
365 }, f
365 d = f.value
366 d = f.value
366 self.ui.debug('received listkey for "%s": %i bytes\n'
367 self.ui.debug('received listkey for "%s": %i bytes\n'
367 % (namespace, len(d)))
368 % (namespace, len(d)))
368 yield pushkey.decodekeys(d)
369 yield pushkey.decodekeys(d)
369
370
370 def _readbundlerevs(bundlerepo):
371 def _readbundlerevs(bundlerepo):
371 return list(bundlerepo.revs('bundle()'))
372 return list(bundlerepo.revs('bundle()'))
372
373
373 def _includefilelogstobundle(bundlecaps, bundlerepo, bundlerevs, ui):
374 def _includefilelogstobundle(bundlecaps, bundlerepo, bundlerevs, ui):
374 '''Tells remotefilelog to include all changed files to the changegroup
375 '''Tells remotefilelog to include all changed files to the changegroup
375
376
376 By default remotefilelog doesn't include file content to the changegroup.
377 By default remotefilelog doesn't include file content to the changegroup.
377 But we need to include it if we are fetching from bundlestore.
378 But we need to include it if we are fetching from bundlestore.
378 '''
379 '''
379 changedfiles = set()
380 changedfiles = set()
380 cl = bundlerepo.changelog
381 cl = bundlerepo.changelog
381 for r in bundlerevs:
382 for r in bundlerevs:
382 # [3] means changed files
383 # [3] means changed files
383 changedfiles.update(cl.read(r)[3])
384 changedfiles.update(cl.read(r)[3])
384 if not changedfiles:
385 if not changedfiles:
385 return bundlecaps
386 return bundlecaps
386
387
387 changedfiles = '\0'.join(changedfiles)
388 changedfiles = '\0'.join(changedfiles)
388 newcaps = []
389 newcaps = []
389 appended = False
390 appended = False
390 for cap in (bundlecaps or []):
391 for cap in (bundlecaps or []):
391 if cap.startswith('excludepattern='):
392 if cap.startswith('excludepattern='):
392 newcaps.append('\0'.join((cap, changedfiles)))
393 newcaps.append('\0'.join((cap, changedfiles)))
393 appended = True
394 appended = True
394 else:
395 else:
395 newcaps.append(cap)
396 newcaps.append(cap)
396 if not appended:
397 if not appended:
397 # Not found excludepattern cap. Just append it
398 # Not found excludepattern cap. Just append it
398 newcaps.append('excludepattern=' + changedfiles)
399 newcaps.append('excludepattern=' + changedfiles)
399
400
400 return newcaps
401 return newcaps
401
402
402 def _rebundle(bundlerepo, bundleroots, unknownhead):
403 def _rebundle(bundlerepo, bundleroots, unknownhead):
403 '''
404 '''
404 Bundle may include more revision then user requested. For example,
405 Bundle may include more revision then user requested. For example,
405 if user asks for revision but bundle also consists its descendants.
406 if user asks for revision but bundle also consists its descendants.
406 This function will filter out all revision that user is not requested.
407 This function will filter out all revision that user is not requested.
407 '''
408 '''
408 parts = []
409 parts = []
409
410
410 version = '02'
411 version = '02'
411 outgoing = discovery.outgoing(bundlerepo, commonheads=bundleroots,
412 outgoing = discovery.outgoing(bundlerepo, commonheads=bundleroots,
412 missingheads=[unknownhead])
413 missingheads=[unknownhead])
413 cgstream = changegroup.makestream(bundlerepo, outgoing, version, 'pull')
414 cgstream = changegroup.makestream(bundlerepo, outgoing, version, 'pull')
414 cgstream = util.chunkbuffer(cgstream).read()
415 cgstream = util.chunkbuffer(cgstream).read()
415 cgpart = bundle2.bundlepart('changegroup', data=cgstream)
416 cgpart = bundle2.bundlepart('changegroup', data=cgstream)
416 cgpart.addparam('version', version)
417 cgpart.addparam('version', version)
417 parts.append(cgpart)
418 parts.append(cgpart)
418
419
419 return parts
420 return parts
420
421
421 def _getbundleroots(oldrepo, bundlerepo, bundlerevs):
422 def _getbundleroots(oldrepo, bundlerepo, bundlerevs):
422 cl = bundlerepo.changelog
423 cl = bundlerepo.changelog
423 bundleroots = []
424 bundleroots = []
424 for rev in bundlerevs:
425 for rev in bundlerevs:
425 node = cl.node(rev)
426 node = cl.node(rev)
426 parents = cl.parents(node)
427 parents = cl.parents(node)
427 for parent in parents:
428 for parent in parents:
428 # include all revs that exist in the main repo
429 # include all revs that exist in the main repo
429 # to make sure that bundle may apply client-side
430 # to make sure that bundle may apply client-side
430 if parent in oldrepo:
431 if parent in oldrepo:
431 bundleroots.append(parent)
432 bundleroots.append(parent)
432 return bundleroots
433 return bundleroots
433
434
434 def _needsrebundling(head, bundlerepo):
435 def _needsrebundling(head, bundlerepo):
435 bundleheads = list(bundlerepo.revs('heads(bundle())'))
436 bundleheads = list(bundlerepo.revs('heads(bundle())'))
436 return not (len(bundleheads) == 1 and
437 return not (len(bundleheads) == 1 and
437 bundlerepo[bundleheads[0]].node() == head)
438 bundlerepo[bundleheads[0]].node() == head)
438
439
439 def _generateoutputparts(head, bundlerepo, bundleroots, bundlefile):
440 def _generateoutputparts(head, bundlerepo, bundleroots, bundlefile):
440 '''generates bundle that will be send to the user
441 '''generates bundle that will be send to the user
441
442
442 returns tuple with raw bundle string and bundle type
443 returns tuple with raw bundle string and bundle type
443 '''
444 '''
444 parts = []
445 parts = []
445 if not _needsrebundling(head, bundlerepo):
446 if not _needsrebundling(head, bundlerepo):
446 with util.posixfile(bundlefile, "rb") as f:
447 with util.posixfile(bundlefile, "rb") as f:
447 unbundler = exchange.readbundle(bundlerepo.ui, f, bundlefile)
448 unbundler = exchange.readbundle(bundlerepo.ui, f, bundlefile)
448 if isinstance(unbundler, changegroup.cg1unpacker):
449 if isinstance(unbundler, changegroup.cg1unpacker):
449 part = bundle2.bundlepart('changegroup',
450 part = bundle2.bundlepart('changegroup',
450 data=unbundler._stream.read())
451 data=unbundler._stream.read())
451 part.addparam('version', '01')
452 part.addparam('version', '01')
452 parts.append(part)
453 parts.append(part)
453 elif isinstance(unbundler, bundle2.unbundle20):
454 elif isinstance(unbundler, bundle2.unbundle20):
454 haschangegroup = False
455 haschangegroup = False
455 for part in unbundler.iterparts():
456 for part in unbundler.iterparts():
456 if part.type == 'changegroup':
457 if part.type == 'changegroup':
457 haschangegroup = True
458 haschangegroup = True
458 newpart = bundle2.bundlepart(part.type, data=part.read())
459 newpart = bundle2.bundlepart(part.type, data=part.read())
459 for key, value in part.params.iteritems():
460 for key, value in part.params.iteritems():
460 newpart.addparam(key, value)
461 newpart.addparam(key, value)
461 parts.append(newpart)
462 parts.append(newpart)
462
463
463 if not haschangegroup:
464 if not haschangegroup:
464 raise error.Abort(
465 raise error.Abort(
465 'unexpected bundle without changegroup part, ' +
466 'unexpected bundle without changegroup part, ' +
466 'head: %s' % hex(head),
467 'head: %s' % hex(head),
467 hint='report to administrator')
468 hint='report to administrator')
468 else:
469 else:
469 raise error.Abort('unknown bundle type')
470 raise error.Abort('unknown bundle type')
470 else:
471 else:
471 parts = _rebundle(bundlerepo, bundleroots, head)
472 parts = _rebundle(bundlerepo, bundleroots, head)
472
473
473 return parts
474 return parts
474
475
475 def getbundlechunks(orig, repo, source, heads=None, bundlecaps=None, **kwargs):
476 def getbundlechunks(orig, repo, source, heads=None, bundlecaps=None, **kwargs):
476 heads = heads or []
477 heads = heads or []
477 # newheads are parents of roots of scratch bundles that were requested
478 # newheads are parents of roots of scratch bundles that were requested
478 newphases = {}
479 newphases = {}
479 scratchbundles = []
480 scratchbundles = []
480 newheads = []
481 newheads = []
481 scratchheads = []
482 scratchheads = []
482 nodestobundle = {}
483 nodestobundle = {}
483 allbundlestocleanup = []
484 allbundlestocleanup = []
484 try:
485 try:
485 for head in heads:
486 for head in heads:
486 if head not in repo.changelog.nodemap:
487 if head not in repo.changelog.nodemap:
487 if head not in nodestobundle:
488 if head not in nodestobundle:
488 newbundlefile = common.downloadbundle(repo, head)
489 newbundlefile = common.downloadbundle(repo, head)
489 bundlepath = "bundle:%s+%s" % (repo.root, newbundlefile)
490 bundlepath = "bundle:%s+%s" % (repo.root, newbundlefile)
490 bundlerepo = hg.repository(repo.ui, bundlepath)
491 bundlerepo = hg.repository(repo.ui, bundlepath)
491
492
492 allbundlestocleanup.append((bundlerepo, newbundlefile))
493 allbundlestocleanup.append((bundlerepo, newbundlefile))
493 bundlerevs = set(_readbundlerevs(bundlerepo))
494 bundlerevs = set(_readbundlerevs(bundlerepo))
494 bundlecaps = _includefilelogstobundle(
495 bundlecaps = _includefilelogstobundle(
495 bundlecaps, bundlerepo, bundlerevs, repo.ui)
496 bundlecaps, bundlerepo, bundlerevs, repo.ui)
496 cl = bundlerepo.changelog
497 cl = bundlerepo.changelog
497 bundleroots = _getbundleroots(repo, bundlerepo, bundlerevs)
498 bundleroots = _getbundleroots(repo, bundlerepo, bundlerevs)
498 for rev in bundlerevs:
499 for rev in bundlerevs:
499 node = cl.node(rev)
500 node = cl.node(rev)
500 newphases[hex(node)] = str(phases.draft)
501 newphases[hex(node)] = str(phases.draft)
501 nodestobundle[node] = (bundlerepo, bundleroots,
502 nodestobundle[node] = (bundlerepo, bundleroots,
502 newbundlefile)
503 newbundlefile)
503
504
504 scratchbundles.append(
505 scratchbundles.append(
505 _generateoutputparts(head, *nodestobundle[head]))
506 _generateoutputparts(head, *nodestobundle[head]))
506 newheads.extend(bundleroots)
507 newheads.extend(bundleroots)
507 scratchheads.append(head)
508 scratchheads.append(head)
508 finally:
509 finally:
509 for bundlerepo, bundlefile in allbundlestocleanup:
510 for bundlerepo, bundlefile in allbundlestocleanup:
510 bundlerepo.close()
511 bundlerepo.close()
511 try:
512 try:
512 os.unlink(bundlefile)
513 os.unlink(bundlefile)
513 except (IOError, OSError):
514 except (IOError, OSError):
514 # if we can't cleanup the file then just ignore the error,
515 # if we can't cleanup the file then just ignore the error,
515 # no need to fail
516 # no need to fail
516 pass
517 pass
517
518
518 pullfrombundlestore = bool(scratchbundles)
519 pullfrombundlestore = bool(scratchbundles)
519 wrappedchangegrouppart = False
520 wrappedchangegrouppart = False
520 wrappedlistkeys = False
521 wrappedlistkeys = False
521 oldchangegrouppart = exchange.getbundle2partsmapping['changegroup']
522 oldchangegrouppart = exchange.getbundle2partsmapping['changegroup']
522 try:
523 try:
523 def _changegrouppart(bundler, *args, **kwargs):
524 def _changegrouppart(bundler, *args, **kwargs):
524 # Order is important here. First add non-scratch part
525 # Order is important here. First add non-scratch part
525 # and only then add parts with scratch bundles because
526 # and only then add parts with scratch bundles because
526 # non-scratch part contains parents of roots of scratch bundles.
527 # non-scratch part contains parents of roots of scratch bundles.
527 result = oldchangegrouppart(bundler, *args, **kwargs)
528 result = oldchangegrouppart(bundler, *args, **kwargs)
528 for bundle in scratchbundles:
529 for bundle in scratchbundles:
529 for part in bundle:
530 for part in bundle:
530 bundler.addpart(part)
531 bundler.addpart(part)
531 return result
532 return result
532
533
533 exchange.getbundle2partsmapping['changegroup'] = _changegrouppart
534 exchange.getbundle2partsmapping['changegroup'] = _changegrouppart
534 wrappedchangegrouppart = True
535 wrappedchangegrouppart = True
535
536
536 def _listkeys(orig, self, namespace):
537 def _listkeys(orig, self, namespace):
537 origvalues = orig(self, namespace)
538 origvalues = orig(self, namespace)
538 if namespace == 'phases' and pullfrombundlestore:
539 if namespace == 'phases' and pullfrombundlestore:
539 if origvalues.get('publishing') == 'True':
540 if origvalues.get('publishing') == 'True':
540 # Make repo non-publishing to preserve draft phase
541 # Make repo non-publishing to preserve draft phase
541 del origvalues['publishing']
542 del origvalues['publishing']
542 origvalues.update(newphases)
543 origvalues.update(newphases)
543 return origvalues
544 return origvalues
544
545
545 extensions.wrapfunction(localrepo.localrepository, 'listkeys',
546 extensions.wrapfunction(localrepo.localrepository, 'listkeys',
546 _listkeys)
547 _listkeys)
547 wrappedlistkeys = True
548 wrappedlistkeys = True
548 heads = list((set(newheads) | set(heads)) - set(scratchheads))
549 heads = list((set(newheads) | set(heads)) - set(scratchheads))
549 result = orig(repo, source, heads=heads,
550 result = orig(repo, source, heads=heads,
550 bundlecaps=bundlecaps, **kwargs)
551 bundlecaps=bundlecaps, **kwargs)
551 finally:
552 finally:
552 if wrappedchangegrouppart:
553 if wrappedchangegrouppart:
553 exchange.getbundle2partsmapping['changegroup'] = oldchangegrouppart
554 exchange.getbundle2partsmapping['changegroup'] = oldchangegrouppart
554 if wrappedlistkeys:
555 if wrappedlistkeys:
555 extensions.unwrapfunction(localrepo.localrepository, 'listkeys',
556 extensions.unwrapfunction(localrepo.localrepository, 'listkeys',
556 _listkeys)
557 _listkeys)
557 return result
558 return result
558
559
559 def _lookupwrap(orig):
560 def _lookupwrap(orig):
560 def _lookup(repo, proto, key):
561 def _lookup(repo, proto, key):
561 localkey = encoding.tolocal(key)
562 localkey = encoding.tolocal(key)
562
563
563 if isinstance(localkey, str) and _scratchbranchmatcher(localkey):
564 if isinstance(localkey, str) and _scratchbranchmatcher(localkey):
564 scratchnode = repo.bundlestore.index.getnode(localkey)
565 scratchnode = repo.bundlestore.index.getnode(localkey)
565 if scratchnode:
566 if scratchnode:
566 return "%s %s\n" % (1, scratchnode)
567 return "%s %s\n" % (1, scratchnode)
567 else:
568 else:
568 return "%s %s\n" % (0, 'scratch branch %s not found' % localkey)
569 return "%s %s\n" % (0, 'scratch branch %s not found' % localkey)
569 else:
570 else:
570 try:
571 try:
571 r = hex(repo.lookup(localkey))
572 r = hex(repo.lookup(localkey))
572 return "%s %s\n" % (1, r)
573 return "%s %s\n" % (1, r)
573 except Exception as inst:
574 except Exception as inst:
574 if repo.bundlestore.index.getbundle(localkey):
575 if repo.bundlestore.index.getbundle(localkey):
575 return "%s %s\n" % (1, localkey)
576 return "%s %s\n" % (1, localkey)
576 else:
577 else:
577 r = str(inst)
578 r = str(inst)
578 return "%s %s\n" % (0, r)
579 return "%s %s\n" % (0, r)
579 return _lookup
580 return _lookup
580
581
581 def _pull(orig, ui, repo, source="default", **opts):
582 def _pull(orig, ui, repo, source="default", **opts):
583 opts = pycompat.byteskwargs(opts)
582 # Copy paste from `pull` command
584 # Copy paste from `pull` command
583 source, branches = hg.parseurl(ui.expandpath(source), opts.get('branch'))
585 source, branches = hg.parseurl(ui.expandpath(source), opts.get('branch'))
584
586
585 scratchbookmarks = {}
587 scratchbookmarks = {}
586 unfi = repo.unfiltered()
588 unfi = repo.unfiltered()
587 unknownnodes = []
589 unknownnodes = []
588 for rev in opts.get('rev', []):
590 for rev in opts.get('rev', []):
589 if rev not in unfi:
591 if rev not in unfi:
590 unknownnodes.append(rev)
592 unknownnodes.append(rev)
591 if opts.get('bookmark'):
593 if opts.get('bookmark'):
592 bookmarks = []
594 bookmarks = []
593 revs = opts.get('rev') or []
595 revs = opts.get('rev') or []
594 for bookmark in opts.get('bookmark'):
596 for bookmark in opts.get('bookmark'):
595 if _scratchbranchmatcher(bookmark):
597 if _scratchbranchmatcher(bookmark):
596 # rev is not known yet
598 # rev is not known yet
597 # it will be fetched with listkeyspatterns next
599 # it will be fetched with listkeyspatterns next
598 scratchbookmarks[bookmark] = 'REVTOFETCH'
600 scratchbookmarks[bookmark] = 'REVTOFETCH'
599 else:
601 else:
600 bookmarks.append(bookmark)
602 bookmarks.append(bookmark)
601
603
602 if scratchbookmarks:
604 if scratchbookmarks:
603 other = hg.peer(repo, opts, source)
605 other = hg.peer(repo, opts, source)
604 fetchedbookmarks = other.listkeyspatterns(
606 fetchedbookmarks = other.listkeyspatterns(
605 'bookmarks', patterns=scratchbookmarks)
607 'bookmarks', patterns=scratchbookmarks)
606 for bookmark in scratchbookmarks:
608 for bookmark in scratchbookmarks:
607 if bookmark not in fetchedbookmarks:
609 if bookmark not in fetchedbookmarks:
608 raise error.Abort('remote bookmark %s not found!' %
610 raise error.Abort('remote bookmark %s not found!' %
609 bookmark)
611 bookmark)
610 scratchbookmarks[bookmark] = fetchedbookmarks[bookmark]
612 scratchbookmarks[bookmark] = fetchedbookmarks[bookmark]
611 revs.append(fetchedbookmarks[bookmark])
613 revs.append(fetchedbookmarks[bookmark])
612 opts['bookmark'] = bookmarks
614 opts['bookmark'] = bookmarks
613 opts['rev'] = revs
615 opts['rev'] = revs
614
616
615 if scratchbookmarks or unknownnodes:
617 if scratchbookmarks or unknownnodes:
616 # Set anyincoming to True
618 # Set anyincoming to True
617 extensions.wrapfunction(discovery, 'findcommonincoming',
619 extensions.wrapfunction(discovery, 'findcommonincoming',
618 _findcommonincoming)
620 _findcommonincoming)
619 try:
621 try:
620 # Remote scratch bookmarks will be deleted because remotenames doesn't
622 # Remote scratch bookmarks will be deleted because remotenames doesn't
621 # know about them. Let's save it before pull and restore after
623 # know about them. Let's save it before pull and restore after
622 remotescratchbookmarks = _readscratchremotebookmarks(ui, repo, source)
624 remotescratchbookmarks = _readscratchremotebookmarks(ui, repo, source)
623 result = orig(ui, repo, source, **opts)
625 result = orig(ui, repo, source, **pycompat.strkwargs(opts))
624 # TODO(stash): race condition is possible
626 # TODO(stash): race condition is possible
625 # if scratch bookmarks was updated right after orig.
627 # if scratch bookmarks was updated right after orig.
626 # But that's unlikely and shouldn't be harmful.
628 # But that's unlikely and shouldn't be harmful.
627 if common.isremotebooksenabled(ui):
629 if common.isremotebooksenabled(ui):
628 remotescratchbookmarks.update(scratchbookmarks)
630 remotescratchbookmarks.update(scratchbookmarks)
629 _saveremotebookmarks(repo, remotescratchbookmarks, source)
631 _saveremotebookmarks(repo, remotescratchbookmarks, source)
630 else:
632 else:
631 _savelocalbookmarks(repo, scratchbookmarks)
633 _savelocalbookmarks(repo, scratchbookmarks)
632 return result
634 return result
633 finally:
635 finally:
634 if scratchbookmarks:
636 if scratchbookmarks:
635 extensions.unwrapfunction(discovery, 'findcommonincoming')
637 extensions.unwrapfunction(discovery, 'findcommonincoming')
636
638
637 def _readscratchremotebookmarks(ui, repo, other):
639 def _readscratchremotebookmarks(ui, repo, other):
638 if common.isremotebooksenabled(ui):
640 if common.isremotebooksenabled(ui):
639 remotenamesext = extensions.find('remotenames')
641 remotenamesext = extensions.find('remotenames')
640 remotepath = remotenamesext.activepath(repo.ui, other)
642 remotepath = remotenamesext.activepath(repo.ui, other)
641 result = {}
643 result = {}
642 # Let's refresh remotenames to make sure we have it up to date
644 # Let's refresh remotenames to make sure we have it up to date
643 # Seems that `repo.names['remotebookmarks']` may return stale bookmarks
645 # Seems that `repo.names['remotebookmarks']` may return stale bookmarks
644 # and it results in deleting scratch bookmarks. Our best guess how to
646 # and it results in deleting scratch bookmarks. Our best guess how to
645 # fix it is to use `clearnames()`
647 # fix it is to use `clearnames()`
646 repo._remotenames.clearnames()
648 repo._remotenames.clearnames()
647 for remotebookmark in repo.names['remotebookmarks'].listnames(repo):
649 for remotebookmark in repo.names['remotebookmarks'].listnames(repo):
648 path, bookname = remotenamesext.splitremotename(remotebookmark)
650 path, bookname = remotenamesext.splitremotename(remotebookmark)
649 if path == remotepath and _scratchbranchmatcher(bookname):
651 if path == remotepath and _scratchbranchmatcher(bookname):
650 nodes = repo.names['remotebookmarks'].nodes(repo,
652 nodes = repo.names['remotebookmarks'].nodes(repo,
651 remotebookmark)
653 remotebookmark)
652 if nodes:
654 if nodes:
653 result[bookname] = hex(nodes[0])
655 result[bookname] = hex(nodes[0])
654 return result
656 return result
655 else:
657 else:
656 return {}
658 return {}
657
659
658 def _saveremotebookmarks(repo, newbookmarks, remote):
660 def _saveremotebookmarks(repo, newbookmarks, remote):
659 remotenamesext = extensions.find('remotenames')
661 remotenamesext = extensions.find('remotenames')
660 remotepath = remotenamesext.activepath(repo.ui, remote)
662 remotepath = remotenamesext.activepath(repo.ui, remote)
661 branches = collections.defaultdict(list)
663 branches = collections.defaultdict(list)
662 bookmarks = {}
664 bookmarks = {}
663 remotenames = remotenamesext.readremotenames(repo)
665 remotenames = remotenamesext.readremotenames(repo)
664 for hexnode, nametype, remote, rname in remotenames:
666 for hexnode, nametype, remote, rname in remotenames:
665 if remote != remotepath:
667 if remote != remotepath:
666 continue
668 continue
667 if nametype == 'bookmarks':
669 if nametype == 'bookmarks':
668 if rname in newbookmarks:
670 if rname in newbookmarks:
669 # It's possible if we have a normal bookmark that matches
671 # It's possible if we have a normal bookmark that matches
670 # scratch branch pattern. In this case just use the current
672 # scratch branch pattern. In this case just use the current
671 # bookmark node
673 # bookmark node
672 del newbookmarks[rname]
674 del newbookmarks[rname]
673 bookmarks[rname] = hexnode
675 bookmarks[rname] = hexnode
674 elif nametype == 'branches':
676 elif nametype == 'branches':
675 # saveremotenames expects 20 byte binary nodes for branches
677 # saveremotenames expects 20 byte binary nodes for branches
676 branches[rname].append(bin(hexnode))
678 branches[rname].append(bin(hexnode))
677
679
678 for bookmark, hexnode in newbookmarks.iteritems():
680 for bookmark, hexnode in newbookmarks.iteritems():
679 bookmarks[bookmark] = hexnode
681 bookmarks[bookmark] = hexnode
680 remotenamesext.saveremotenames(repo, remotepath, branches, bookmarks)
682 remotenamesext.saveremotenames(repo, remotepath, branches, bookmarks)
681
683
682 def _savelocalbookmarks(repo, bookmarks):
684 def _savelocalbookmarks(repo, bookmarks):
683 if not bookmarks:
685 if not bookmarks:
684 return
686 return
685 with repo.wlock(), repo.lock(), repo.transaction('bookmark') as tr:
687 with repo.wlock(), repo.lock(), repo.transaction('bookmark') as tr:
686 changes = []
688 changes = []
687 for scratchbook, node in bookmarks.iteritems():
689 for scratchbook, node in bookmarks.iteritems():
688 changectx = repo[node]
690 changectx = repo[node]
689 changes.append((scratchbook, changectx.node()))
691 changes.append((scratchbook, changectx.node()))
690 repo._bookmarks.applychanges(repo, tr, changes)
692 repo._bookmarks.applychanges(repo, tr, changes)
691
693
692 def _findcommonincoming(orig, *args, **kwargs):
694 def _findcommonincoming(orig, *args, **kwargs):
693 common, inc, remoteheads = orig(*args, **kwargs)
695 common, inc, remoteheads = orig(*args, **kwargs)
694 return common, True, remoteheads
696 return common, True, remoteheads
695
697
696 def _push(orig, ui, repo, dest=None, *args, **opts):
698 def _push(orig, ui, repo, dest=None, *args, **opts):
697
699
698 bookmark = opts.get('bookmark')
700 bookmark = opts.get('bookmark')
699 # we only support pushing one infinitepush bookmark at once
701 # we only support pushing one infinitepush bookmark at once
700 if len(bookmark) == 1:
702 if len(bookmark) == 1:
701 bookmark = bookmark[0]
703 bookmark = bookmark[0]
702 else:
704 else:
703 bookmark = ''
705 bookmark = ''
704
706
705 oldphasemove = None
707 oldphasemove = None
706 overrides = {(experimental, configbookmark): bookmark}
708 overrides = {(experimental, configbookmark): bookmark}
707
709
708 with ui.configoverride(overrides, 'infinitepush'):
710 with ui.configoverride(overrides, 'infinitepush'):
709 scratchpush = opts.get('bundle_store')
711 scratchpush = opts.get('bundle_store')
710 if _scratchbranchmatcher(bookmark):
712 if _scratchbranchmatcher(bookmark):
711 scratchpush = True
713 scratchpush = True
712 # bundle2 can be sent back after push (for example, bundle2
714 # bundle2 can be sent back after push (for example, bundle2
713 # containing `pushkey` part to update bookmarks)
715 # containing `pushkey` part to update bookmarks)
714 ui.setconfig(experimental, 'bundle2.pushback', True)
716 ui.setconfig(experimental, 'bundle2.pushback', True)
715
717
716 if scratchpush:
718 if scratchpush:
717 # this is an infinitepush, we don't want the bookmark to be applied
719 # this is an infinitepush, we don't want the bookmark to be applied
718 # rather that should be stored in the bundlestore
720 # rather that should be stored in the bundlestore
719 opts['bookmark'] = []
721 opts['bookmark'] = []
720 ui.setconfig(experimental, configscratchpush, True)
722 ui.setconfig(experimental, configscratchpush, True)
721 oldphasemove = extensions.wrapfunction(exchange,
723 oldphasemove = extensions.wrapfunction(exchange,
722 '_localphasemove',
724 '_localphasemove',
723 _phasemove)
725 _phasemove)
724 # Copy-paste from `push` command
726 # Copy-paste from `push` command
725 path = ui.paths.getpath(dest, default=('default-push', 'default'))
727 path = ui.paths.getpath(dest, default=('default-push', 'default'))
726 if not path:
728 if not path:
727 raise error.Abort(_('default repository not configured!'),
729 raise error.Abort(_('default repository not configured!'),
728 hint=_("see 'hg help config.paths'"))
730 hint=_("see 'hg help config.paths'"))
729 destpath = path.pushloc or path.loc
731 destpath = path.pushloc or path.loc
730 # Remote scratch bookmarks will be deleted because remotenames doesn't
732 # Remote scratch bookmarks will be deleted because remotenames doesn't
731 # know about them. Let's save it before push and restore after
733 # know about them. Let's save it before push and restore after
732 remotescratchbookmarks = _readscratchremotebookmarks(ui, repo, destpath)
734 remotescratchbookmarks = _readscratchremotebookmarks(ui, repo, destpath)
733 result = orig(ui, repo, dest, *args, **opts)
735 result = orig(ui, repo, dest, *args, **opts)
734 if common.isremotebooksenabled(ui):
736 if common.isremotebooksenabled(ui):
735 if bookmark and scratchpush:
737 if bookmark and scratchpush:
736 other = hg.peer(repo, opts, destpath)
738 other = hg.peer(repo, opts, destpath)
737 fetchedbookmarks = other.listkeyspatterns('bookmarks',
739 fetchedbookmarks = other.listkeyspatterns('bookmarks',
738 patterns=[bookmark])
740 patterns=[bookmark])
739 remotescratchbookmarks.update(fetchedbookmarks)
741 remotescratchbookmarks.update(fetchedbookmarks)
740 _saveremotebookmarks(repo, remotescratchbookmarks, destpath)
742 _saveremotebookmarks(repo, remotescratchbookmarks, destpath)
741 if oldphasemove:
743 if oldphasemove:
742 exchange._localphasemove = oldphasemove
744 exchange._localphasemove = oldphasemove
743 return result
745 return result
744
746
745 def _deleteinfinitepushbookmarks(ui, repo, path, names):
747 def _deleteinfinitepushbookmarks(ui, repo, path, names):
746 """Prune remote names by removing the bookmarks we don't want anymore,
748 """Prune remote names by removing the bookmarks we don't want anymore,
747 then writing the result back to disk
749 then writing the result back to disk
748 """
750 """
749 remotenamesext = extensions.find('remotenames')
751 remotenamesext = extensions.find('remotenames')
750
752
751 # remotename format is:
753 # remotename format is:
752 # (node, nametype ("branches" or "bookmarks"), remote, name)
754 # (node, nametype ("branches" or "bookmarks"), remote, name)
753 nametype_idx = 1
755 nametype_idx = 1
754 remote_idx = 2
756 remote_idx = 2
755 name_idx = 3
757 name_idx = 3
756 remotenames = [remotename for remotename in \
758 remotenames = [remotename for remotename in \
757 remotenamesext.readremotenames(repo) \
759 remotenamesext.readremotenames(repo) \
758 if remotename[remote_idx] == path]
760 if remotename[remote_idx] == path]
759 remote_bm_names = [remotename[name_idx] for remotename in \
761 remote_bm_names = [remotename[name_idx] for remotename in \
760 remotenames if remotename[nametype_idx] == "bookmarks"]
762 remotenames if remotename[nametype_idx] == "bookmarks"]
761
763
762 for name in names:
764 for name in names:
763 if name not in remote_bm_names:
765 if name not in remote_bm_names:
764 raise error.Abort(_("infinitepush bookmark '{}' does not exist "
766 raise error.Abort(_("infinitepush bookmark '{}' does not exist "
765 "in path '{}'").format(name, path))
767 "in path '{}'").format(name, path))
766
768
767 bookmarks = {}
769 bookmarks = {}
768 branches = collections.defaultdict(list)
770 branches = collections.defaultdict(list)
769 for node, nametype, remote, name in remotenames:
771 for node, nametype, remote, name in remotenames:
770 if nametype == "bookmarks" and name not in names:
772 if nametype == "bookmarks" and name not in names:
771 bookmarks[name] = node
773 bookmarks[name] = node
772 elif nametype == "branches":
774 elif nametype == "branches":
773 # saveremotenames wants binary nodes for branches
775 # saveremotenames wants binary nodes for branches
774 branches[name].append(bin(node))
776 branches[name].append(bin(node))
775
777
776 remotenamesext.saveremotenames(repo, path, branches, bookmarks)
778 remotenamesext.saveremotenames(repo, path, branches, bookmarks)
777
779
778 def _phasemove(orig, pushop, nodes, phase=phases.public):
780 def _phasemove(orig, pushop, nodes, phase=phases.public):
779 """prevent commits from being marked public
781 """prevent commits from being marked public
780
782
781 Since these are going to a scratch branch, they aren't really being
783 Since these are going to a scratch branch, they aren't really being
782 published."""
784 published."""
783
785
784 if phase != phases.public:
786 if phase != phases.public:
785 orig(pushop, nodes, phase)
787 orig(pushop, nodes, phase)
786
788
787 @exchange.b2partsgenerator(scratchbranchparttype)
789 @exchange.b2partsgenerator(scratchbranchparttype)
788 def partgen(pushop, bundler):
790 def partgen(pushop, bundler):
789 bookmark = pushop.ui.config(experimental, configbookmark)
791 bookmark = pushop.ui.config(experimental, configbookmark)
790 scratchpush = pushop.ui.configbool(experimental, configscratchpush)
792 scratchpush = pushop.ui.configbool(experimental, configscratchpush)
791 if 'changesets' in pushop.stepsdone or not scratchpush:
793 if 'changesets' in pushop.stepsdone or not scratchpush:
792 return
794 return
793
795
794 if scratchbranchparttype not in bundle2.bundle2caps(pushop.remote):
796 if scratchbranchparttype not in bundle2.bundle2caps(pushop.remote):
795 return
797 return
796
798
797 pushop.stepsdone.add('changesets')
799 pushop.stepsdone.add('changesets')
798 if not pushop.outgoing.missing:
800 if not pushop.outgoing.missing:
799 pushop.ui.status(_('no changes found\n'))
801 pushop.ui.status(_('no changes found\n'))
800 pushop.cgresult = 0
802 pushop.cgresult = 0
801 return
803 return
802
804
803 # This parameter tells the server that the following bundle is an
805 # This parameter tells the server that the following bundle is an
804 # infinitepush. This let's it switch the part processing to our infinitepush
806 # infinitepush. This let's it switch the part processing to our infinitepush
805 # code path.
807 # code path.
806 bundler.addparam("infinitepush", "True")
808 bundler.addparam("infinitepush", "True")
807
809
808 scratchparts = bundleparts.getscratchbranchparts(pushop.repo,
810 scratchparts = bundleparts.getscratchbranchparts(pushop.repo,
809 pushop.remote,
811 pushop.remote,
810 pushop.outgoing,
812 pushop.outgoing,
811 pushop.ui,
813 pushop.ui,
812 bookmark)
814 bookmark)
813
815
814 for scratchpart in scratchparts:
816 for scratchpart in scratchparts:
815 bundler.addpart(scratchpart)
817 bundler.addpart(scratchpart)
816
818
817 def handlereply(op):
819 def handlereply(op):
818 # server either succeeds or aborts; no code to read
820 # server either succeeds or aborts; no code to read
819 pushop.cgresult = 1
821 pushop.cgresult = 1
820
822
821 return handlereply
823 return handlereply
822
824
823 bundle2.capabilities[bundleparts.scratchbranchparttype] = ()
825 bundle2.capabilities[bundleparts.scratchbranchparttype] = ()
824
826
825 def _getrevs(bundle, oldnode, force, bookmark):
827 def _getrevs(bundle, oldnode, force, bookmark):
826 'extracts and validates the revs to be imported'
828 'extracts and validates the revs to be imported'
827 revs = [bundle[r] for r in bundle.revs('sort(bundle())')]
829 revs = [bundle[r] for r in bundle.revs('sort(bundle())')]
828
830
829 # new bookmark
831 # new bookmark
830 if oldnode is None:
832 if oldnode is None:
831 return revs
833 return revs
832
834
833 # Fast forward update
835 # Fast forward update
834 if oldnode in bundle and list(bundle.set('bundle() & %s::', oldnode)):
836 if oldnode in bundle and list(bundle.set('bundle() & %s::', oldnode)):
835 return revs
837 return revs
836
838
837 return revs
839 return revs
838
840
839 @contextlib.contextmanager
841 @contextlib.contextmanager
840 def logservicecall(logger, service, **kwargs):
842 def logservicecall(logger, service, **kwargs):
841 start = time.time()
843 start = time.time()
842 logger(service, eventtype='start', **kwargs)
844 logger(service, eventtype='start', **kwargs)
843 try:
845 try:
844 yield
846 yield
845 logger(service, eventtype='success',
847 logger(service, eventtype='success',
846 elapsedms=(time.time() - start) * 1000, **kwargs)
848 elapsedms=(time.time() - start) * 1000, **kwargs)
847 except Exception as e:
849 except Exception as e:
848 logger(service, eventtype='failure',
850 logger(service, eventtype='failure',
849 elapsedms=(time.time() - start) * 1000, errormsg=str(e),
851 elapsedms=(time.time() - start) * 1000, errormsg=str(e),
850 **kwargs)
852 **kwargs)
851 raise
853 raise
852
854
853 def _getorcreateinfinitepushlogger(op):
855 def _getorcreateinfinitepushlogger(op):
854 logger = op.records['infinitepushlogger']
856 logger = op.records['infinitepushlogger']
855 if not logger:
857 if not logger:
856 ui = op.repo.ui
858 ui = op.repo.ui
857 try:
859 try:
858 username = procutil.getuser()
860 username = procutil.getuser()
859 except Exception:
861 except Exception:
860 username = 'unknown'
862 username = 'unknown'
861 # Generate random request id to be able to find all logged entries
863 # Generate random request id to be able to find all logged entries
862 # for the same request. Since requestid is pseudo-generated it may
864 # for the same request. Since requestid is pseudo-generated it may
863 # not be unique, but we assume that (hostname, username, requestid)
865 # not be unique, but we assume that (hostname, username, requestid)
864 # is unique.
866 # is unique.
865 random.seed()
867 random.seed()
866 requestid = random.randint(0, 2000000000)
868 requestid = random.randint(0, 2000000000)
867 hostname = socket.gethostname()
869 hostname = socket.gethostname()
868 logger = functools.partial(ui.log, 'infinitepush', user=username,
870 logger = functools.partial(ui.log, 'infinitepush', user=username,
869 requestid=requestid, hostname=hostname,
871 requestid=requestid, hostname=hostname,
870 reponame=ui.config('infinitepush',
872 reponame=ui.config('infinitepush',
871 'reponame'))
873 'reponame'))
872 op.records.add('infinitepushlogger', logger)
874 op.records.add('infinitepushlogger', logger)
873 else:
875 else:
874 logger = logger[0]
876 logger = logger[0]
875 return logger
877 return logger
876
878
877 def storetobundlestore(orig, repo, op, unbundler):
879 def storetobundlestore(orig, repo, op, unbundler):
878 """stores the incoming bundle coming from push command to the bundlestore
880 """stores the incoming bundle coming from push command to the bundlestore
879 instead of applying on the revlogs"""
881 instead of applying on the revlogs"""
880
882
881 repo.ui.status(_("storing changesets on the bundlestore\n"))
883 repo.ui.status(_("storing changesets on the bundlestore\n"))
882 bundler = bundle2.bundle20(repo.ui)
884 bundler = bundle2.bundle20(repo.ui)
883
885
884 # processing each part and storing it in bundler
886 # processing each part and storing it in bundler
885 with bundle2.partiterator(repo, op, unbundler) as parts:
887 with bundle2.partiterator(repo, op, unbundler) as parts:
886 for part in parts:
888 for part in parts:
887 bundlepart = None
889 bundlepart = None
888 if part.type == 'replycaps':
890 if part.type == 'replycaps':
889 # This configures the current operation to allow reply parts.
891 # This configures the current operation to allow reply parts.
890 bundle2._processpart(op, part)
892 bundle2._processpart(op, part)
891 else:
893 else:
892 bundlepart = bundle2.bundlepart(part.type, data=part.read())
894 bundlepart = bundle2.bundlepart(part.type, data=part.read())
893 for key, value in part.params.iteritems():
895 for key, value in part.params.iteritems():
894 bundlepart.addparam(key, value)
896 bundlepart.addparam(key, value)
895
897
896 # Certain parts require a response
898 # Certain parts require a response
897 if part.type in ('pushkey', 'changegroup'):
899 if part.type in ('pushkey', 'changegroup'):
898 if op.reply is not None:
900 if op.reply is not None:
899 rpart = op.reply.newpart('reply:%s' % part.type)
901 rpart = op.reply.newpart('reply:%s' % part.type)
900 rpart.addparam('in-reply-to', str(part.id),
902 rpart.addparam('in-reply-to', str(part.id),
901 mandatory=False)
903 mandatory=False)
902 rpart.addparam('return', '1', mandatory=False)
904 rpart.addparam('return', '1', mandatory=False)
903
905
904 op.records.add(part.type, {
906 op.records.add(part.type, {
905 'return': 1,
907 'return': 1,
906 })
908 })
907 if bundlepart:
909 if bundlepart:
908 bundler.addpart(bundlepart)
910 bundler.addpart(bundlepart)
909
911
910 # storing the bundle in the bundlestore
912 # storing the bundle in the bundlestore
911 buf = util.chunkbuffer(bundler.getchunks())
913 buf = util.chunkbuffer(bundler.getchunks())
912 fd, bundlefile = tempfile.mkstemp()
914 fd, bundlefile = tempfile.mkstemp()
913 try:
915 try:
914 try:
916 try:
915 fp = os.fdopen(fd, 'wb')
917 fp = os.fdopen(fd, 'wb')
916 fp.write(buf.read())
918 fp.write(buf.read())
917 finally:
919 finally:
918 fp.close()
920 fp.close()
919 storebundle(op, {}, bundlefile)
921 storebundle(op, {}, bundlefile)
920 finally:
922 finally:
921 try:
923 try:
922 os.unlink(bundlefile)
924 os.unlink(bundlefile)
923 except Exception:
925 except Exception:
924 # we would rather see the original exception
926 # we would rather see the original exception
925 pass
927 pass
926
928
927 def processparts(orig, repo, op, unbundler):
929 def processparts(orig, repo, op, unbundler):
928
930
929 # make sure we don't wrap processparts in case of `hg unbundle`
931 # make sure we don't wrap processparts in case of `hg unbundle`
930 if op.source == 'unbundle':
932 if op.source == 'unbundle':
931 return orig(repo, op, unbundler)
933 return orig(repo, op, unbundler)
932
934
933 # this server routes each push to bundle store
935 # this server routes each push to bundle store
934 if repo.ui.configbool('infinitepush', 'pushtobundlestore'):
936 if repo.ui.configbool('infinitepush', 'pushtobundlestore'):
935 return storetobundlestore(orig, repo, op, unbundler)
937 return storetobundlestore(orig, repo, op, unbundler)
936
938
937 if unbundler.params.get('infinitepush') != 'True':
939 if unbundler.params.get('infinitepush') != 'True':
938 return orig(repo, op, unbundler)
940 return orig(repo, op, unbundler)
939
941
940 handleallparts = repo.ui.configbool('infinitepush', 'storeallparts')
942 handleallparts = repo.ui.configbool('infinitepush', 'storeallparts')
941
943
942 bundler = bundle2.bundle20(repo.ui)
944 bundler = bundle2.bundle20(repo.ui)
943 cgparams = None
945 cgparams = None
944 with bundle2.partiterator(repo, op, unbundler) as parts:
946 with bundle2.partiterator(repo, op, unbundler) as parts:
945 for part in parts:
947 for part in parts:
946 bundlepart = None
948 bundlepart = None
947 if part.type == 'replycaps':
949 if part.type == 'replycaps':
948 # This configures the current operation to allow reply parts.
950 # This configures the current operation to allow reply parts.
949 bundle2._processpart(op, part)
951 bundle2._processpart(op, part)
950 elif part.type == bundleparts.scratchbranchparttype:
952 elif part.type == bundleparts.scratchbranchparttype:
951 # Scratch branch parts need to be converted to normal
953 # Scratch branch parts need to be converted to normal
952 # changegroup parts, and the extra parameters stored for later
954 # changegroup parts, and the extra parameters stored for later
953 # when we upload to the store. Eventually those parameters will
955 # when we upload to the store. Eventually those parameters will
954 # be put on the actual bundle instead of this part, then we can
956 # be put on the actual bundle instead of this part, then we can
955 # send a vanilla changegroup instead of the scratchbranch part.
957 # send a vanilla changegroup instead of the scratchbranch part.
956 cgversion = part.params.get('cgversion', '01')
958 cgversion = part.params.get('cgversion', '01')
957 bundlepart = bundle2.bundlepart('changegroup', data=part.read())
959 bundlepart = bundle2.bundlepart('changegroup', data=part.read())
958 bundlepart.addparam('version', cgversion)
960 bundlepart.addparam('version', cgversion)
959 cgparams = part.params
961 cgparams = part.params
960
962
961 # If we're not dumping all parts into the new bundle, we need to
963 # If we're not dumping all parts into the new bundle, we need to
962 # alert the future pushkey and phase-heads handler to skip
964 # alert the future pushkey and phase-heads handler to skip
963 # the part.
965 # the part.
964 if not handleallparts:
966 if not handleallparts:
965 op.records.add(scratchbranchparttype + '_skippushkey', True)
967 op.records.add(scratchbranchparttype + '_skippushkey', True)
966 op.records.add(scratchbranchparttype + '_skipphaseheads',
968 op.records.add(scratchbranchparttype + '_skipphaseheads',
967 True)
969 True)
968 else:
970 else:
969 if handleallparts:
971 if handleallparts:
970 # Ideally we would not process any parts, and instead just
972 # Ideally we would not process any parts, and instead just
971 # forward them to the bundle for storage, but since this
973 # forward them to the bundle for storage, but since this
972 # differs from previous behavior, we need to put it behind a
974 # differs from previous behavior, we need to put it behind a
973 # config flag for incremental rollout.
975 # config flag for incremental rollout.
974 bundlepart = bundle2.bundlepart(part.type, data=part.read())
976 bundlepart = bundle2.bundlepart(part.type, data=part.read())
975 for key, value in part.params.iteritems():
977 for key, value in part.params.iteritems():
976 bundlepart.addparam(key, value)
978 bundlepart.addparam(key, value)
977
979
978 # Certain parts require a response
980 # Certain parts require a response
979 if part.type == 'pushkey':
981 if part.type == 'pushkey':
980 if op.reply is not None:
982 if op.reply is not None:
981 rpart = op.reply.newpart('reply:pushkey')
983 rpart = op.reply.newpart('reply:pushkey')
982 rpart.addparam('in-reply-to', str(part.id),
984 rpart.addparam('in-reply-to', str(part.id),
983 mandatory=False)
985 mandatory=False)
984 rpart.addparam('return', '1', mandatory=False)
986 rpart.addparam('return', '1', mandatory=False)
985 else:
987 else:
986 bundle2._processpart(op, part)
988 bundle2._processpart(op, part)
987
989
988 if handleallparts:
990 if handleallparts:
989 op.records.add(part.type, {
991 op.records.add(part.type, {
990 'return': 1,
992 'return': 1,
991 })
993 })
992 if bundlepart:
994 if bundlepart:
993 bundler.addpart(bundlepart)
995 bundler.addpart(bundlepart)
994
996
995 # If commits were sent, store them
997 # If commits were sent, store them
996 if cgparams:
998 if cgparams:
997 buf = util.chunkbuffer(bundler.getchunks())
999 buf = util.chunkbuffer(bundler.getchunks())
998 fd, bundlefile = tempfile.mkstemp()
1000 fd, bundlefile = tempfile.mkstemp()
999 try:
1001 try:
1000 try:
1002 try:
1001 fp = os.fdopen(fd, 'wb')
1003 fp = os.fdopen(fd, 'wb')
1002 fp.write(buf.read())
1004 fp.write(buf.read())
1003 finally:
1005 finally:
1004 fp.close()
1006 fp.close()
1005 storebundle(op, cgparams, bundlefile)
1007 storebundle(op, cgparams, bundlefile)
1006 finally:
1008 finally:
1007 try:
1009 try:
1008 os.unlink(bundlefile)
1010 os.unlink(bundlefile)
1009 except Exception:
1011 except Exception:
1010 # we would rather see the original exception
1012 # we would rather see the original exception
1011 pass
1013 pass
1012
1014
1013 def storebundle(op, params, bundlefile):
1015 def storebundle(op, params, bundlefile):
1014 log = _getorcreateinfinitepushlogger(op)
1016 log = _getorcreateinfinitepushlogger(op)
1015 parthandlerstart = time.time()
1017 parthandlerstart = time.time()
1016 log(scratchbranchparttype, eventtype='start')
1018 log(scratchbranchparttype, eventtype='start')
1017 index = op.repo.bundlestore.index
1019 index = op.repo.bundlestore.index
1018 store = op.repo.bundlestore.store
1020 store = op.repo.bundlestore.store
1019 op.records.add(scratchbranchparttype + '_skippushkey', True)
1021 op.records.add(scratchbranchparttype + '_skippushkey', True)
1020
1022
1021 bundle = None
1023 bundle = None
1022 try: # guards bundle
1024 try: # guards bundle
1023 bundlepath = "bundle:%s+%s" % (op.repo.root, bundlefile)
1025 bundlepath = "bundle:%s+%s" % (op.repo.root, bundlefile)
1024 bundle = hg.repository(op.repo.ui, bundlepath)
1026 bundle = hg.repository(op.repo.ui, bundlepath)
1025
1027
1026 bookmark = params.get('bookmark')
1028 bookmark = params.get('bookmark')
1027 bookprevnode = params.get('bookprevnode', '')
1029 bookprevnode = params.get('bookprevnode', '')
1028 force = params.get('force')
1030 force = params.get('force')
1029
1031
1030 if bookmark:
1032 if bookmark:
1031 oldnode = index.getnode(bookmark)
1033 oldnode = index.getnode(bookmark)
1032 else:
1034 else:
1033 oldnode = None
1035 oldnode = None
1034 bundleheads = bundle.revs('heads(bundle())')
1036 bundleheads = bundle.revs('heads(bundle())')
1035 if bookmark and len(bundleheads) > 1:
1037 if bookmark and len(bundleheads) > 1:
1036 raise error.Abort(
1038 raise error.Abort(
1037 _('cannot push more than one head to a scratch branch'))
1039 _('cannot push more than one head to a scratch branch'))
1038
1040
1039 revs = _getrevs(bundle, oldnode, force, bookmark)
1041 revs = _getrevs(bundle, oldnode, force, bookmark)
1040
1042
1041 # Notify the user of what is being pushed
1043 # Notify the user of what is being pushed
1042 plural = 's' if len(revs) > 1 else ''
1044 plural = 's' if len(revs) > 1 else ''
1043 op.repo.ui.warn(_("pushing %s commit%s:\n") % (len(revs), plural))
1045 op.repo.ui.warn(_("pushing %s commit%s:\n") % (len(revs), plural))
1044 maxoutput = 10
1046 maxoutput = 10
1045 for i in range(0, min(len(revs), maxoutput)):
1047 for i in range(0, min(len(revs), maxoutput)):
1046 firstline = bundle[revs[i]].description().split('\n')[0][:50]
1048 firstline = bundle[revs[i]].description().split('\n')[0][:50]
1047 op.repo.ui.warn((" %s %s\n") % (revs[i], firstline))
1049 op.repo.ui.warn((" %s %s\n") % (revs[i], firstline))
1048
1050
1049 if len(revs) > maxoutput + 1:
1051 if len(revs) > maxoutput + 1:
1050 op.repo.ui.warn((" ...\n"))
1052 op.repo.ui.warn((" ...\n"))
1051 firstline = bundle[revs[-1]].description().split('\n')[0][:50]
1053 firstline = bundle[revs[-1]].description().split('\n')[0][:50]
1052 op.repo.ui.warn((" %s %s\n") % (revs[-1], firstline))
1054 op.repo.ui.warn((" %s %s\n") % (revs[-1], firstline))
1053
1055
1054 nodesctx = [bundle[rev] for rev in revs]
1056 nodesctx = [bundle[rev] for rev in revs]
1055 inindex = lambda rev: bool(index.getbundle(bundle[rev].hex()))
1057 inindex = lambda rev: bool(index.getbundle(bundle[rev].hex()))
1056 if bundleheads:
1058 if bundleheads:
1057 newheadscount = sum(not inindex(rev) for rev in bundleheads)
1059 newheadscount = sum(not inindex(rev) for rev in bundleheads)
1058 else:
1060 else:
1059 newheadscount = 0
1061 newheadscount = 0
1060 # If there's a bookmark specified, there should be only one head,
1062 # If there's a bookmark specified, there should be only one head,
1061 # so we choose the last node, which will be that head.
1063 # so we choose the last node, which will be that head.
1062 # If a bug or malicious client allows there to be a bookmark
1064 # If a bug or malicious client allows there to be a bookmark
1063 # with multiple heads, we will place the bookmark on the last head.
1065 # with multiple heads, we will place the bookmark on the last head.
1064 bookmarknode = nodesctx[-1].hex() if nodesctx else None
1066 bookmarknode = nodesctx[-1].hex() if nodesctx else None
1065 key = None
1067 key = None
1066 if newheadscount:
1068 if newheadscount:
1067 with open(bundlefile, 'r') as f:
1069 with open(bundlefile, 'r') as f:
1068 bundledata = f.read()
1070 bundledata = f.read()
1069 with logservicecall(log, 'bundlestore',
1071 with logservicecall(log, 'bundlestore',
1070 bundlesize=len(bundledata)):
1072 bundlesize=len(bundledata)):
1071 bundlesizelimit = 100 * 1024 * 1024 # 100 MB
1073 bundlesizelimit = 100 * 1024 * 1024 # 100 MB
1072 if len(bundledata) > bundlesizelimit:
1074 if len(bundledata) > bundlesizelimit:
1073 error_msg = ('bundle is too big: %d bytes. ' +
1075 error_msg = ('bundle is too big: %d bytes. ' +
1074 'max allowed size is 100 MB')
1076 'max allowed size is 100 MB')
1075 raise error.Abort(error_msg % (len(bundledata),))
1077 raise error.Abort(error_msg % (len(bundledata),))
1076 key = store.write(bundledata)
1078 key = store.write(bundledata)
1077
1079
1078 with logservicecall(log, 'index', newheadscount=newheadscount), index:
1080 with logservicecall(log, 'index', newheadscount=newheadscount), index:
1079 if key:
1081 if key:
1080 index.addbundle(key, nodesctx)
1082 index.addbundle(key, nodesctx)
1081 if bookmark:
1083 if bookmark:
1082 index.addbookmark(bookmark, bookmarknode)
1084 index.addbookmark(bookmark, bookmarknode)
1083 _maybeaddpushbackpart(op, bookmark, bookmarknode,
1085 _maybeaddpushbackpart(op, bookmark, bookmarknode,
1084 bookprevnode, params)
1086 bookprevnode, params)
1085 log(scratchbranchparttype, eventtype='success',
1087 log(scratchbranchparttype, eventtype='success',
1086 elapsedms=(time.time() - parthandlerstart) * 1000)
1088 elapsedms=(time.time() - parthandlerstart) * 1000)
1087
1089
1088 except Exception as e:
1090 except Exception as e:
1089 log(scratchbranchparttype, eventtype='failure',
1091 log(scratchbranchparttype, eventtype='failure',
1090 elapsedms=(time.time() - parthandlerstart) * 1000,
1092 elapsedms=(time.time() - parthandlerstart) * 1000,
1091 errormsg=str(e))
1093 errormsg=str(e))
1092 raise
1094 raise
1093 finally:
1095 finally:
1094 if bundle:
1096 if bundle:
1095 bundle.close()
1097 bundle.close()
1096
1098
1097 @bundle2.parthandler(scratchbranchparttype,
1099 @bundle2.parthandler(scratchbranchparttype,
1098 ('bookmark', 'bookprevnode', 'force',
1100 ('bookmark', 'bookprevnode', 'force',
1099 'pushbackbookmarks', 'cgversion'))
1101 'pushbackbookmarks', 'cgversion'))
1100 def bundle2scratchbranch(op, part):
1102 def bundle2scratchbranch(op, part):
1101 '''unbundle a bundle2 part containing a changegroup to store'''
1103 '''unbundle a bundle2 part containing a changegroup to store'''
1102
1104
1103 bundler = bundle2.bundle20(op.repo.ui)
1105 bundler = bundle2.bundle20(op.repo.ui)
1104 cgversion = part.params.get('cgversion', '01')
1106 cgversion = part.params.get('cgversion', '01')
1105 cgpart = bundle2.bundlepart('changegroup', data=part.read())
1107 cgpart = bundle2.bundlepart('changegroup', data=part.read())
1106 cgpart.addparam('version', cgversion)
1108 cgpart.addparam('version', cgversion)
1107 bundler.addpart(cgpart)
1109 bundler.addpart(cgpart)
1108 buf = util.chunkbuffer(bundler.getchunks())
1110 buf = util.chunkbuffer(bundler.getchunks())
1109
1111
1110 fd, bundlefile = tempfile.mkstemp()
1112 fd, bundlefile = tempfile.mkstemp()
1111 try:
1113 try:
1112 try:
1114 try:
1113 fp = os.fdopen(fd, 'wb')
1115 fp = os.fdopen(fd, 'wb')
1114 fp.write(buf.read())
1116 fp.write(buf.read())
1115 finally:
1117 finally:
1116 fp.close()
1118 fp.close()
1117 storebundle(op, part.params, bundlefile)
1119 storebundle(op, part.params, bundlefile)
1118 finally:
1120 finally:
1119 try:
1121 try:
1120 os.unlink(bundlefile)
1122 os.unlink(bundlefile)
1121 except OSError as e:
1123 except OSError as e:
1122 if e.errno != errno.ENOENT:
1124 if e.errno != errno.ENOENT:
1123 raise
1125 raise
1124
1126
1125 return 1
1127 return 1
1126
1128
1127 def _maybeaddpushbackpart(op, bookmark, newnode, oldnode, params):
1129 def _maybeaddpushbackpart(op, bookmark, newnode, oldnode, params):
1128 if params.get('pushbackbookmarks'):
1130 if params.get('pushbackbookmarks'):
1129 if op.reply and 'pushback' in op.reply.capabilities:
1131 if op.reply and 'pushback' in op.reply.capabilities:
1130 params = {
1132 params = {
1131 'namespace': 'bookmarks',
1133 'namespace': 'bookmarks',
1132 'key': bookmark,
1134 'key': bookmark,
1133 'new': newnode,
1135 'new': newnode,
1134 'old': oldnode,
1136 'old': oldnode,
1135 }
1137 }
1136 op.reply.newpart('pushkey', mandatoryparams=params.iteritems())
1138 op.reply.newpart('pushkey', mandatoryparams=params.iteritems())
1137
1139
1138 def bundle2pushkey(orig, op, part):
1140 def bundle2pushkey(orig, op, part):
1139 '''Wrapper of bundle2.handlepushkey()
1141 '''Wrapper of bundle2.handlepushkey()
1140
1142
1141 The only goal is to skip calling the original function if flag is set.
1143 The only goal is to skip calling the original function if flag is set.
1142 It's set if infinitepush push is happening.
1144 It's set if infinitepush push is happening.
1143 '''
1145 '''
1144 if op.records[scratchbranchparttype + '_skippushkey']:
1146 if op.records[scratchbranchparttype + '_skippushkey']:
1145 if op.reply is not None:
1147 if op.reply is not None:
1146 rpart = op.reply.newpart('reply:pushkey')
1148 rpart = op.reply.newpart('reply:pushkey')
1147 rpart.addparam('in-reply-to', str(part.id), mandatory=False)
1149 rpart.addparam('in-reply-to', str(part.id), mandatory=False)
1148 rpart.addparam('return', '1', mandatory=False)
1150 rpart.addparam('return', '1', mandatory=False)
1149 return 1
1151 return 1
1150
1152
1151 return orig(op, part)
1153 return orig(op, part)
1152
1154
1153 def bundle2handlephases(orig, op, part):
1155 def bundle2handlephases(orig, op, part):
1154 '''Wrapper of bundle2.handlephases()
1156 '''Wrapper of bundle2.handlephases()
1155
1157
1156 The only goal is to skip calling the original function if flag is set.
1158 The only goal is to skip calling the original function if flag is set.
1157 It's set if infinitepush push is happening.
1159 It's set if infinitepush push is happening.
1158 '''
1160 '''
1159
1161
1160 if op.records[scratchbranchparttype + '_skipphaseheads']:
1162 if op.records[scratchbranchparttype + '_skipphaseheads']:
1161 return
1163 return
1162
1164
1163 return orig(op, part)
1165 return orig(op, part)
1164
1166
1165 def _asyncsavemetadata(root, nodes):
1167 def _asyncsavemetadata(root, nodes):
1166 '''starts a separate process that fills metadata for the nodes
1168 '''starts a separate process that fills metadata for the nodes
1167
1169
1168 This function creates a separate process and doesn't wait for it's
1170 This function creates a separate process and doesn't wait for it's
1169 completion. This was done to avoid slowing down pushes
1171 completion. This was done to avoid slowing down pushes
1170 '''
1172 '''
1171
1173
1172 maxnodes = 50
1174 maxnodes = 50
1173 if len(nodes) > maxnodes:
1175 if len(nodes) > maxnodes:
1174 return
1176 return
1175 nodesargs = []
1177 nodesargs = []
1176 for node in nodes:
1178 for node in nodes:
1177 nodesargs.append('--node')
1179 nodesargs.append('--node')
1178 nodesargs.append(node)
1180 nodesargs.append(node)
1179 with open(os.devnull, 'w+b') as devnull:
1181 with open(os.devnull, 'w+b') as devnull:
1180 cmdline = [util.hgexecutable(), 'debugfillinfinitepushmetadata',
1182 cmdline = [util.hgexecutable(), 'debugfillinfinitepushmetadata',
1181 '-R', root] + nodesargs
1183 '-R', root] + nodesargs
1182 # Process will run in background. We don't care about the return code
1184 # Process will run in background. We don't care about the return code
1183 subprocess.Popen(cmdline, close_fds=True, shell=False,
1185 subprocess.Popen(cmdline, close_fds=True, shell=False,
1184 stdin=devnull, stdout=devnull, stderr=devnull)
1186 stdin=devnull, stdout=devnull, stderr=devnull)
General Comments 0
You need to be logged in to leave comments. Login now