##// END OF EJS Templates
py3: use default dict iterator instead of iterkeys...
Augie Fackler -
r36313:23864678 default
parent child Browse files
Show More
@@ -1,516 +1,516
1 1 # synthrepo.py - repo synthesis
2 2 #
3 3 # Copyright 2012 Facebook
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 '''synthesize structurally interesting change history
9 9
10 10 This extension is useful for creating a repository with properties
11 11 that are statistically similar to an existing repository. During
12 12 analysis, a simple probability table is constructed from the history
13 13 of an existing repository. During synthesis, these properties are
14 14 reconstructed.
15 15
16 16 Properties that are analyzed and synthesized include the following:
17 17
18 18 - Lines added or removed when an existing file is modified
19 19 - Number and sizes of files added
20 20 - Number of files removed
21 21 - Line lengths
22 22 - Topological distance to parent changeset(s)
23 23 - Probability of a commit being a merge
24 24 - Probability of a newly added file being added to a new directory
25 25 - Interarrival time, and time zone, of commits
26 26 - Number of files in each directory
27 27
28 28 A few obvious properties that are not currently handled realistically:
29 29
30 30 - Merges are treated as regular commits with two parents, which is not
31 31 realistic
32 32 - Modifications are not treated as operations on hunks of lines, but
33 33 as insertions and deletions of randomly chosen single lines
34 34 - Committer ID (always random)
35 35 - Executability of files
36 36 - Symlinks and binary files are ignored
37 37 '''
38 38
39 39 from __future__ import absolute_import
40 40 import bisect
41 41 import collections
42 42 import itertools
43 43 import json
44 44 import os
45 45 import random
46 46 import sys
47 47 import time
48 48
49 49 from mercurial.i18n import _
50 50 from mercurial.node import (
51 51 nullid,
52 52 nullrev,
53 53 short,
54 54 )
55 55 from mercurial import (
56 56 context,
57 57 error,
58 58 hg,
59 59 patch,
60 60 registrar,
61 61 scmutil,
62 62 util,
63 63 )
64 64
65 65 # Note for extension authors: ONLY specify testedwith = 'ships-with-hg-core' for
66 66 # extensions which SHIP WITH MERCURIAL. Non-mainline extensions should
67 67 # be specifying the version(s) of Mercurial they are tested with, or
68 68 # leave the attribute unspecified.
69 69 testedwith = 'ships-with-hg-core'
70 70
71 71 cmdtable = {}
72 72 command = registrar.command(cmdtable)
73 73
74 74 newfile = {'new fi', 'rename', 'copy f', 'copy t'}
75 75
76 76 def zerodict():
77 77 return collections.defaultdict(lambda: 0)
78 78
79 79 def roundto(x, k):
80 80 if x > k * 2:
81 81 return int(round(x / float(k)) * k)
82 82 return int(round(x))
83 83
84 84 def parsegitdiff(lines):
85 85 filename, mar, lineadd, lineremove = None, None, zerodict(), 0
86 86 binary = False
87 87 for line in lines:
88 88 start = line[:6]
89 89 if start == 'diff -':
90 90 if filename:
91 91 yield filename, mar, lineadd, lineremove, binary
92 92 mar, lineadd, lineremove, binary = 'm', zerodict(), 0, False
93 93 filename = patch.gitre.match(line).group(1)
94 94 elif start in newfile:
95 95 mar = 'a'
96 96 elif start == 'GIT bi':
97 97 binary = True
98 98 elif start == 'delete':
99 99 mar = 'r'
100 100 elif start:
101 101 s = start[0]
102 102 if s == '-' and not line.startswith('--- '):
103 103 lineremove += 1
104 104 elif s == '+' and not line.startswith('+++ '):
105 105 lineadd[roundto(len(line) - 1, 5)] += 1
106 106 if filename:
107 107 yield filename, mar, lineadd, lineremove, binary
108 108
109 109 @command('analyze',
110 110 [('o', 'output', '', _('write output to given file'), _('FILE')),
111 111 ('r', 'rev', [], _('analyze specified revisions'), _('REV'))],
112 112 _('hg analyze'), optionalrepo=True)
113 113 def analyze(ui, repo, *revs, **opts):
114 114 '''create a simple model of a repository to use for later synthesis
115 115
116 116 This command examines every changeset in the given range (or all
117 117 of history if none are specified) and creates a simple statistical
118 118 model of the history of the repository. It also measures the directory
119 119 structure of the repository as checked out.
120 120
121 121 The model is written out to a JSON file, and can be used by
122 122 :hg:`synthesize` to create or augment a repository with synthetic
123 123 commits that have a structure that is statistically similar to the
124 124 analyzed repository.
125 125 '''
126 126 root = repo.root
127 127 if not root.endswith(os.path.sep):
128 128 root += os.path.sep
129 129
130 130 revs = list(revs)
131 131 revs.extend(opts['rev'])
132 132 if not revs:
133 133 revs = [':']
134 134
135 135 output = opts['output']
136 136 if not output:
137 137 output = os.path.basename(root) + '.json'
138 138
139 139 if output == '-':
140 140 fp = sys.stdout
141 141 else:
142 142 fp = open(output, 'w')
143 143
144 144 # Always obtain file counts of each directory in the given root directory.
145 145 def onerror(e):
146 146 ui.warn(_('error walking directory structure: %s\n') % e)
147 147
148 148 dirs = {}
149 149 rootprefixlen = len(root)
150 150 for dirpath, dirnames, filenames in os.walk(root, onerror=onerror):
151 151 dirpathfromroot = dirpath[rootprefixlen:]
152 152 dirs[dirpathfromroot] = len(filenames)
153 153 if '.hg' in dirnames:
154 154 dirnames.remove('.hg')
155 155
156 156 lineschanged = zerodict()
157 157 children = zerodict()
158 158 p1distance = zerodict()
159 159 p2distance = zerodict()
160 160 linesinfilesadded = zerodict()
161 161 fileschanged = zerodict()
162 162 filesadded = zerodict()
163 163 filesremoved = zerodict()
164 164 linelengths = zerodict()
165 165 interarrival = zerodict()
166 166 parents = zerodict()
167 167 dirsadded = zerodict()
168 168 tzoffset = zerodict()
169 169
170 170 # If a mercurial repo is available, also model the commit history.
171 171 if repo:
172 172 revs = scmutil.revrange(repo, revs)
173 173 revs.sort()
174 174
175 175 progress = ui.progress
176 176 _analyzing = _('analyzing')
177 177 _changesets = _('changesets')
178 178 _total = len(revs)
179 179
180 180 for i, rev in enumerate(revs):
181 181 progress(_analyzing, i, unit=_changesets, total=_total)
182 182 ctx = repo[rev]
183 183 pl = ctx.parents()
184 184 pctx = pl[0]
185 185 prev = pctx.rev()
186 186 children[prev] += 1
187 187 p1distance[rev - prev] += 1
188 188 parents[len(pl)] += 1
189 189 tzoffset[ctx.date()[1]] += 1
190 190 if len(pl) > 1:
191 191 p2distance[rev - pl[1].rev()] += 1
192 192 if prev == rev - 1:
193 193 lastctx = pctx
194 194 else:
195 195 lastctx = repo[rev - 1]
196 196 if lastctx.rev() != nullrev:
197 197 timedelta = ctx.date()[0] - lastctx.date()[0]
198 198 interarrival[roundto(timedelta, 300)] += 1
199 199 diff = sum((d.splitlines() for d in ctx.diff(pctx, git=True)), [])
200 200 fileadds, diradds, fileremoves, filechanges = 0, 0, 0, 0
201 201 for filename, mar, lineadd, lineremove, isbin in parsegitdiff(diff):
202 202 if isbin:
203 203 continue
204 204 added = sum(lineadd.itervalues(), 0)
205 205 if mar == 'm':
206 206 if added and lineremove:
207 207 lineschanged[roundto(added, 5),
208 208 roundto(lineremove, 5)] += 1
209 209 filechanges += 1
210 210 elif mar == 'a':
211 211 fileadds += 1
212 212 if '/' in filename:
213 213 filedir = filename.rsplit('/', 1)[0]
214 214 if filedir not in pctx.dirs():
215 215 diradds += 1
216 216 linesinfilesadded[roundto(added, 5)] += 1
217 217 elif mar == 'r':
218 218 fileremoves += 1
219 219 for length, count in lineadd.iteritems():
220 220 linelengths[length] += count
221 221 fileschanged[filechanges] += 1
222 222 filesadded[fileadds] += 1
223 223 dirsadded[diradds] += 1
224 224 filesremoved[fileremoves] += 1
225 225
226 226 invchildren = zerodict()
227 227
228 228 for rev, count in children.iteritems():
229 229 invchildren[count] += 1
230 230
231 231 if output != '-':
232 232 ui.status(_('writing output to %s\n') % output)
233 233
234 234 def pronk(d):
235 235 return sorted(d.iteritems(), key=lambda x: x[1], reverse=True)
236 236
237 237 json.dump({'revs': len(revs),
238 238 'initdirs': pronk(dirs),
239 239 'lineschanged': pronk(lineschanged),
240 240 'children': pronk(invchildren),
241 241 'fileschanged': pronk(fileschanged),
242 242 'filesadded': pronk(filesadded),
243 243 'linesinfilesadded': pronk(linesinfilesadded),
244 244 'dirsadded': pronk(dirsadded),
245 245 'filesremoved': pronk(filesremoved),
246 246 'linelengths': pronk(linelengths),
247 247 'parents': pronk(parents),
248 248 'p1distance': pronk(p1distance),
249 249 'p2distance': pronk(p2distance),
250 250 'interarrival': pronk(interarrival),
251 251 'tzoffset': pronk(tzoffset),
252 252 },
253 253 fp)
254 254 fp.close()
255 255
256 256 @command('synthesize',
257 257 [('c', 'count', 0, _('create given number of commits'), _('COUNT')),
258 258 ('', 'dict', '', _('path to a dictionary of words'), _('FILE')),
259 259 ('', 'initfiles', 0, _('initial file count to create'), _('COUNT'))],
260 260 _('hg synthesize [OPTION].. DESCFILE'))
261 261 def synthesize(ui, repo, descpath, **opts):
262 262 '''synthesize commits based on a model of an existing repository
263 263
264 264 The model must have been generated by :hg:`analyze`. Commits will
265 265 be generated randomly according to the probabilities described in
266 266 the model. If --initfiles is set, the repository will be seeded with
267 267 the given number files following the modeled repository's directory
268 268 structure.
269 269
270 270 When synthesizing new content, commit descriptions, and user
271 271 names, words will be chosen randomly from a dictionary that is
272 272 presumed to contain one word per line. Use --dict to specify the
273 273 path to an alternate dictionary to use.
274 274 '''
275 275 try:
276 276 fp = hg.openpath(ui, descpath)
277 277 except Exception as err:
278 278 raise error.Abort('%s: %s' % (descpath, err[0].strerror))
279 279 desc = json.load(fp)
280 280 fp.close()
281 281
282 282 def cdf(l):
283 283 if not l:
284 284 return [], []
285 285 vals, probs = zip(*sorted(l, key=lambda x: x[1], reverse=True))
286 286 t = float(sum(probs, 0))
287 287 s, cdfs = 0, []
288 288 for v in probs:
289 289 s += v
290 290 cdfs.append(s / t)
291 291 return vals, cdfs
292 292
293 293 lineschanged = cdf(desc['lineschanged'])
294 294 fileschanged = cdf(desc['fileschanged'])
295 295 filesadded = cdf(desc['filesadded'])
296 296 dirsadded = cdf(desc['dirsadded'])
297 297 filesremoved = cdf(desc['filesremoved'])
298 298 linelengths = cdf(desc['linelengths'])
299 299 parents = cdf(desc['parents'])
300 300 p1distance = cdf(desc['p1distance'])
301 301 p2distance = cdf(desc['p2distance'])
302 302 interarrival = cdf(desc['interarrival'])
303 303 linesinfilesadded = cdf(desc['linesinfilesadded'])
304 304 tzoffset = cdf(desc['tzoffset'])
305 305
306 306 dictfile = opts.get('dict') or '/usr/share/dict/words'
307 307 try:
308 308 fp = open(dictfile, 'rU')
309 309 except IOError as err:
310 310 raise error.Abort('%s: %s' % (dictfile, err.strerror))
311 311 words = fp.read().splitlines()
312 312 fp.close()
313 313
314 314 initdirs = {}
315 315 if desc['initdirs']:
316 316 for k, v in desc['initdirs']:
317 317 initdirs[k.encode('utf-8').replace('.hg', '_hg')] = v
318 318 initdirs = renamedirs(initdirs, words)
319 319 initdirscdf = cdf(initdirs)
320 320
321 321 def pick(cdf):
322 322 return cdf[0][bisect.bisect_left(cdf[1], random.random())]
323 323
324 324 def pickpath():
325 325 return os.path.join(pick(initdirscdf), random.choice(words))
326 326
327 327 def makeline(minimum=0):
328 328 total = max(minimum, pick(linelengths))
329 329 c, l = 0, []
330 330 while c < total:
331 331 w = random.choice(words)
332 332 c += len(w) + 1
333 333 l.append(w)
334 334 return ' '.join(l)
335 335
336 336 wlock = repo.wlock()
337 337 lock = repo.lock()
338 338
339 339 nevertouch = {'.hgsub', '.hgignore', '.hgtags'}
340 340
341 341 progress = ui.progress
342 342 _synthesizing = _('synthesizing')
343 343 _files = _('initial files')
344 344 _changesets = _('changesets')
345 345
346 346 # Synthesize a single initial revision adding files to the repo according
347 347 # to the modeled directory structure.
348 348 initcount = int(opts['initfiles'])
349 349 if initcount and initdirs:
350 350 pctx = repo[None].parents()[0]
351 351 dirs = set(pctx.dirs())
352 352 files = {}
353 353
354 354 def validpath(path):
355 355 # Don't pick filenames which are already directory names.
356 356 if path in dirs:
357 357 return False
358 358 # Don't pick directories which were used as file names.
359 359 while path:
360 360 if path in files:
361 361 return False
362 362 path = os.path.dirname(path)
363 363 return True
364 364
365 365 for i in xrange(0, initcount):
366 366 ui.progress(_synthesizing, i, unit=_files, total=initcount)
367 367
368 368 path = pickpath()
369 369 while not validpath(path):
370 370 path = pickpath()
371 371 data = '%s contents\n' % path
372 372 files[path] = data
373 373 dir = os.path.dirname(path)
374 374 while dir and dir not in dirs:
375 375 dirs.add(dir)
376 376 dir = os.path.dirname(dir)
377 377
378 378 def filectxfn(repo, memctx, path):
379 379 return context.memfilectx(repo, memctx, path, files[path])
380 380
381 381 ui.progress(_synthesizing, None)
382 382 message = 'synthesized wide repo with %d files' % (len(files),)
383 383 mc = context.memctx(repo, [pctx.node(), nullid], message,
384 files.iterkeys(), filectxfn, ui.username(),
384 files, filectxfn, ui.username(),
385 385 '%d %d' % util.makedate())
386 386 initnode = mc.commit()
387 387 if ui.debugflag:
388 388 hexfn = hex
389 389 else:
390 390 hexfn = short
391 391 ui.status(_('added commit %s with %d files\n')
392 392 % (hexfn(initnode), len(files)))
393 393
394 394 # Synthesize incremental revisions to the repository, adding repo depth.
395 395 count = int(opts['count'])
396 396 heads = set(map(repo.changelog.rev, repo.heads()))
397 397 for i in xrange(count):
398 398 progress(_synthesizing, i, unit=_changesets, total=count)
399 399
400 400 node = repo.changelog.node
401 401 revs = len(repo)
402 402
403 403 def pickhead(heads, distance):
404 404 if heads:
405 405 lheads = sorted(heads)
406 406 rev = revs - min(pick(distance), revs)
407 407 if rev < lheads[-1]:
408 408 rev = lheads[bisect.bisect_left(lheads, rev)]
409 409 else:
410 410 rev = lheads[-1]
411 411 return rev, node(rev)
412 412 return nullrev, nullid
413 413
414 414 r1 = revs - min(pick(p1distance), revs)
415 415 p1 = node(r1)
416 416
417 417 # the number of heads will grow without bound if we use a pure
418 418 # model, so artificially constrain their proliferation
419 419 toomanyheads = len(heads) > random.randint(1, 20)
420 420 if p2distance[0] and (pick(parents) == 2 or toomanyheads):
421 421 r2, p2 = pickhead(heads.difference([r1]), p2distance)
422 422 else:
423 423 r2, p2 = nullrev, nullid
424 424
425 425 pl = [p1, p2]
426 426 pctx = repo[r1]
427 427 mf = pctx.manifest()
428 428 mfk = mf.keys()
429 429 changes = {}
430 430 if mfk:
431 431 for __ in xrange(pick(fileschanged)):
432 432 for __ in xrange(10):
433 433 fctx = pctx.filectx(random.choice(mfk))
434 434 path = fctx.path()
435 435 if not (path in nevertouch or fctx.isbinary() or
436 436 'l' in fctx.flags()):
437 437 break
438 438 lines = fctx.data().splitlines()
439 439 add, remove = pick(lineschanged)
440 440 for __ in xrange(remove):
441 441 if not lines:
442 442 break
443 443 del lines[random.randrange(0, len(lines))]
444 444 for __ in xrange(add):
445 445 lines.insert(random.randint(0, len(lines)), makeline())
446 446 path = fctx.path()
447 447 changes[path] = '\n'.join(lines) + '\n'
448 448 for __ in xrange(pick(filesremoved)):
449 449 path = random.choice(mfk)
450 450 for __ in xrange(10):
451 451 path = random.choice(mfk)
452 452 if path not in changes:
453 453 break
454 454 if filesadded:
455 455 dirs = list(pctx.dirs())
456 456 dirs.insert(0, '')
457 457 for __ in xrange(pick(filesadded)):
458 458 pathstr = ''
459 459 while pathstr in dirs:
460 460 path = [random.choice(dirs)]
461 461 if pick(dirsadded):
462 462 path.append(random.choice(words))
463 463 path.append(random.choice(words))
464 464 pathstr = '/'.join(filter(None, path))
465 465 data = '\n'.join(makeline()
466 466 for __ in xrange(pick(linesinfilesadded))) + '\n'
467 467 changes[pathstr] = data
468 468 def filectxfn(repo, memctx, path):
469 469 if path not in changes:
470 470 return None
471 471 return context.memfilectx(repo, memctx, path, changes[path])
472 472 if not changes:
473 473 continue
474 474 if revs:
475 475 date = repo['tip'].date()[0] + pick(interarrival)
476 476 else:
477 477 date = time.time() - (86400 * count)
478 478 # dates in mercurial must be positive, fit in 32-bit signed integers.
479 479 date = min(0x7fffffff, max(0, date))
480 480 user = random.choice(words) + '@' + random.choice(words)
481 481 mc = context.memctx(repo, pl, makeline(minimum=2),
482 482 sorted(changes),
483 483 filectxfn, user, '%d %d' % (date, pick(tzoffset)))
484 484 newnode = mc.commit()
485 485 heads.add(repo.changelog.rev(newnode))
486 486 heads.discard(r1)
487 487 heads.discard(r2)
488 488
489 489 lock.release()
490 490 wlock.release()
491 491
492 492 def renamedirs(dirs, words):
493 493 '''Randomly rename the directory names in the per-dir file count dict.'''
494 494 wordgen = itertools.cycle(words)
495 495 replacements = {'': ''}
496 496 def rename(dirpath):
497 497 '''Recursively rename the directory and all path prefixes.
498 498
499 499 The mapping from path to renamed path is stored for all path prefixes
500 500 as in dynamic programming, ensuring linear runtime and consistent
501 501 renaming regardless of iteration order through the model.
502 502 '''
503 503 if dirpath in replacements:
504 504 return replacements[dirpath]
505 505 head, _ = os.path.split(dirpath)
506 506 if head:
507 507 head = rename(head)
508 508 else:
509 509 head = ''
510 510 renamed = os.path.join(head, next(wordgen))
511 511 replacements[dirpath] = renamed
512 512 return renamed
513 513 result = []
514 514 for dirpath, count in dirs.iteritems():
515 515 result.append([rename(dirpath.lstrip(os.sep)), count])
516 516 return result
@@ -1,1356 +1,1356
1 1 # Subversion 1.4/1.5 Python API backend
2 2 #
3 3 # Copyright(C) 2007 Daniel Holth et al
4 4 from __future__ import absolute_import
5 5
6 6 import os
7 7 import re
8 8 import tempfile
9 9 import xml.dom.minidom
10 10
11 11 from mercurial.i18n import _
12 12 from mercurial import (
13 13 encoding,
14 14 error,
15 15 pycompat,
16 16 util,
17 17 vfs as vfsmod,
18 18 )
19 19
20 20 from . import common
21 21
22 22 pickle = util.pickle
23 23 stringio = util.stringio
24 24 propertycache = util.propertycache
25 25 urlerr = util.urlerr
26 26 urlreq = util.urlreq
27 27
28 28 commandline = common.commandline
29 29 commit = common.commit
30 30 converter_sink = common.converter_sink
31 31 converter_source = common.converter_source
32 32 decodeargs = common.decodeargs
33 33 encodeargs = common.encodeargs
34 34 makedatetimestamp = common.makedatetimestamp
35 35 mapfile = common.mapfile
36 36 MissingTool = common.MissingTool
37 37 NoRepo = common.NoRepo
38 38
39 39 # Subversion stuff. Works best with very recent Python SVN bindings
40 40 # e.g. SVN 1.5 or backports. Thanks to the bzr folks for enhancing
41 41 # these bindings.
42 42
43 43 try:
44 44 import svn
45 45 import svn.client
46 46 import svn.core
47 47 import svn.ra
48 48 import svn.delta
49 49 from . import transport
50 50 import warnings
51 51 warnings.filterwarnings('ignore',
52 52 module='svn.core',
53 53 category=DeprecationWarning)
54 54 svn.core.SubversionException # trigger import to catch error
55 55
56 56 except ImportError:
57 57 svn = None
58 58
59 59 class SvnPathNotFound(Exception):
60 60 pass
61 61
62 62 def revsplit(rev):
63 63 """Parse a revision string and return (uuid, path, revnum).
64 64 >>> revsplit(b'svn:a2147622-4a9f-4db4-a8d3-13562ff547b2'
65 65 ... b'/proj%20B/mytrunk/mytrunk@1')
66 66 ('a2147622-4a9f-4db4-a8d3-13562ff547b2', '/proj%20B/mytrunk/mytrunk', 1)
67 67 >>> revsplit(b'svn:8af66a51-67f5-4354-b62c-98d67cc7be1d@1')
68 68 ('', '', 1)
69 69 >>> revsplit(b'@7')
70 70 ('', '', 7)
71 71 >>> revsplit(b'7')
72 72 ('', '', 0)
73 73 >>> revsplit(b'bad')
74 74 ('', '', 0)
75 75 """
76 76 parts = rev.rsplit('@', 1)
77 77 revnum = 0
78 78 if len(parts) > 1:
79 79 revnum = int(parts[1])
80 80 parts = parts[0].split('/', 1)
81 81 uuid = ''
82 82 mod = ''
83 83 if len(parts) > 1 and parts[0].startswith('svn:'):
84 84 uuid = parts[0][4:]
85 85 mod = '/' + parts[1]
86 86 return uuid, mod, revnum
87 87
88 88 def quote(s):
89 89 # As of svn 1.7, many svn calls expect "canonical" paths. In
90 90 # theory, we should call svn.core.*canonicalize() on all paths
91 91 # before passing them to the API. Instead, we assume the base url
92 92 # is canonical and copy the behaviour of svn URL encoding function
93 93 # so we can extend it safely with new components. The "safe"
94 94 # characters were taken from the "svn_uri__char_validity" table in
95 95 # libsvn_subr/path.c.
96 96 return urlreq.quote(s, "!$&'()*+,-./:=@_~")
97 97
98 98 def geturl(path):
99 99 try:
100 100 return svn.client.url_from_path(svn.core.svn_path_canonicalize(path))
101 101 except svn.core.SubversionException:
102 102 # svn.client.url_from_path() fails with local repositories
103 103 pass
104 104 if os.path.isdir(path):
105 105 path = os.path.normpath(os.path.abspath(path))
106 106 if pycompat.iswindows:
107 107 path = '/' + util.normpath(path)
108 108 # Module URL is later compared with the repository URL returned
109 109 # by svn API, which is UTF-8.
110 110 path = encoding.tolocal(path)
111 111 path = 'file://%s' % quote(path)
112 112 return svn.core.svn_path_canonicalize(path)
113 113
114 114 def optrev(number):
115 115 optrev = svn.core.svn_opt_revision_t()
116 116 optrev.kind = svn.core.svn_opt_revision_number
117 117 optrev.value.number = number
118 118 return optrev
119 119
120 120 class changedpath(object):
121 121 def __init__(self, p):
122 122 self.copyfrom_path = p.copyfrom_path
123 123 self.copyfrom_rev = p.copyfrom_rev
124 124 self.action = p.action
125 125
126 126 def get_log_child(fp, url, paths, start, end, limit=0,
127 127 discover_changed_paths=True, strict_node_history=False):
128 128 protocol = -1
129 129 def receiver(orig_paths, revnum, author, date, message, pool):
130 130 paths = {}
131 131 if orig_paths is not None:
132 132 for k, v in orig_paths.iteritems():
133 133 paths[k] = changedpath(v)
134 134 pickle.dump((paths, revnum, author, date, message),
135 135 fp, protocol)
136 136
137 137 try:
138 138 # Use an ra of our own so that our parent can consume
139 139 # our results without confusing the server.
140 140 t = transport.SvnRaTransport(url=url)
141 141 svn.ra.get_log(t.ra, paths, start, end, limit,
142 142 discover_changed_paths,
143 143 strict_node_history,
144 144 receiver)
145 145 except IOError:
146 146 # Caller may interrupt the iteration
147 147 pickle.dump(None, fp, protocol)
148 148 except Exception as inst:
149 149 pickle.dump(str(inst), fp, protocol)
150 150 else:
151 151 pickle.dump(None, fp, protocol)
152 152 fp.close()
153 153 # With large history, cleanup process goes crazy and suddenly
154 154 # consumes *huge* amount of memory. The output file being closed,
155 155 # there is no need for clean termination.
156 156 os._exit(0)
157 157
158 158 def debugsvnlog(ui, **opts):
159 159 """Fetch SVN log in a subprocess and channel them back to parent to
160 160 avoid memory collection issues.
161 161 """
162 162 if svn is None:
163 163 raise error.Abort(_('debugsvnlog could not load Subversion python '
164 164 'bindings'))
165 165
166 166 args = decodeargs(ui.fin.read())
167 167 get_log_child(ui.fout, *args)
168 168
169 169 class logstream(object):
170 170 """Interruptible revision log iterator."""
171 171 def __init__(self, stdout):
172 172 self._stdout = stdout
173 173
174 174 def __iter__(self):
175 175 while True:
176 176 try:
177 177 entry = pickle.load(self._stdout)
178 178 except EOFError:
179 179 raise error.Abort(_('Mercurial failed to run itself, check'
180 180 ' hg executable is in PATH'))
181 181 try:
182 182 orig_paths, revnum, author, date, message = entry
183 183 except (TypeError, ValueError):
184 184 if entry is None:
185 185 break
186 186 raise error.Abort(_("log stream exception '%s'") % entry)
187 187 yield entry
188 188
189 189 def close(self):
190 190 if self._stdout:
191 191 self._stdout.close()
192 192 self._stdout = None
193 193
194 194 class directlogstream(list):
195 195 """Direct revision log iterator.
196 196 This can be used for debugging and development but it will probably leak
197 197 memory and is not suitable for real conversions."""
198 198 def __init__(self, url, paths, start, end, limit=0,
199 199 discover_changed_paths=True, strict_node_history=False):
200 200
201 201 def receiver(orig_paths, revnum, author, date, message, pool):
202 202 paths = {}
203 203 if orig_paths is not None:
204 204 for k, v in orig_paths.iteritems():
205 205 paths[k] = changedpath(v)
206 206 self.append((paths, revnum, author, date, message))
207 207
208 208 # Use an ra of our own so that our parent can consume
209 209 # our results without confusing the server.
210 210 t = transport.SvnRaTransport(url=url)
211 211 svn.ra.get_log(t.ra, paths, start, end, limit,
212 212 discover_changed_paths,
213 213 strict_node_history,
214 214 receiver)
215 215
216 216 def close(self):
217 217 pass
218 218
219 219 # Check to see if the given path is a local Subversion repo. Verify this by
220 220 # looking for several svn-specific files and directories in the given
221 221 # directory.
222 222 def filecheck(ui, path, proto):
223 223 for x in ('locks', 'hooks', 'format', 'db'):
224 224 if not os.path.exists(os.path.join(path, x)):
225 225 return False
226 226 return True
227 227
228 228 # Check to see if a given path is the root of an svn repo over http. We verify
229 229 # this by requesting a version-controlled URL we know can't exist and looking
230 230 # for the svn-specific "not found" XML.
231 231 def httpcheck(ui, path, proto):
232 232 try:
233 233 opener = urlreq.buildopener()
234 234 rsp = opener.open('%s://%s/!svn/ver/0/.svn' % (proto, path), 'rb')
235 235 data = rsp.read()
236 236 except urlerr.httperror as inst:
237 237 if inst.code != 404:
238 238 # Except for 404 we cannot know for sure this is not an svn repo
239 239 ui.warn(_('svn: cannot probe remote repository, assume it could '
240 240 'be a subversion repository. Use --source-type if you '
241 241 'know better.\n'))
242 242 return True
243 243 data = inst.fp.read()
244 244 except Exception:
245 245 # Could be urlerr.urlerror if the URL is invalid or anything else.
246 246 return False
247 247 return '<m:human-readable errcode="160013">' in data
248 248
249 249 protomap = {'http': httpcheck,
250 250 'https': httpcheck,
251 251 'file': filecheck,
252 252 }
253 253 def issvnurl(ui, url):
254 254 try:
255 255 proto, path = url.split('://', 1)
256 256 if proto == 'file':
257 257 if (pycompat.iswindows and path[:1] == '/'
258 258 and path[1:2].isalpha() and path[2:6].lower() == '%3a/'):
259 259 path = path[:2] + ':/' + path[6:]
260 260 path = urlreq.url2pathname(path)
261 261 except ValueError:
262 262 proto = 'file'
263 263 path = os.path.abspath(url)
264 264 if proto == 'file':
265 265 path = util.pconvert(path)
266 266 check = protomap.get(proto, lambda *args: False)
267 267 while '/' in path:
268 268 if check(ui, path, proto):
269 269 return True
270 270 path = path.rsplit('/', 1)[0]
271 271 return False
272 272
273 273 # SVN conversion code stolen from bzr-svn and tailor
274 274 #
275 275 # Subversion looks like a versioned filesystem, branches structures
276 276 # are defined by conventions and not enforced by the tool. First,
277 277 # we define the potential branches (modules) as "trunk" and "branches"
278 278 # children directories. Revisions are then identified by their
279 279 # module and revision number (and a repository identifier).
280 280 #
281 281 # The revision graph is really a tree (or a forest). By default, a
282 282 # revision parent is the previous revision in the same module. If the
283 283 # module directory is copied/moved from another module then the
284 284 # revision is the module root and its parent the source revision in
285 285 # the parent module. A revision has at most one parent.
286 286 #
287 287 class svn_source(converter_source):
288 288 def __init__(self, ui, repotype, url, revs=None):
289 289 super(svn_source, self).__init__(ui, repotype, url, revs=revs)
290 290
291 291 if not (url.startswith('svn://') or url.startswith('svn+ssh://') or
292 292 (os.path.exists(url) and
293 293 os.path.exists(os.path.join(url, '.svn'))) or
294 294 issvnurl(ui, url)):
295 295 raise NoRepo(_("%s does not look like a Subversion repository")
296 296 % url)
297 297 if svn is None:
298 298 raise MissingTool(_('could not load Subversion python bindings'))
299 299
300 300 try:
301 301 version = svn.core.SVN_VER_MAJOR, svn.core.SVN_VER_MINOR
302 302 if version < (1, 4):
303 303 raise MissingTool(_('Subversion python bindings %d.%d found, '
304 304 '1.4 or later required') % version)
305 305 except AttributeError:
306 306 raise MissingTool(_('Subversion python bindings are too old, 1.4 '
307 307 'or later required'))
308 308
309 309 self.lastrevs = {}
310 310
311 311 latest = None
312 312 try:
313 313 # Support file://path@rev syntax. Useful e.g. to convert
314 314 # deleted branches.
315 315 at = url.rfind('@')
316 316 if at >= 0:
317 317 latest = int(url[at + 1:])
318 318 url = url[:at]
319 319 except ValueError:
320 320 pass
321 321 self.url = geturl(url)
322 322 self.encoding = 'UTF-8' # Subversion is always nominal UTF-8
323 323 try:
324 324 self.transport = transport.SvnRaTransport(url=self.url)
325 325 self.ra = self.transport.ra
326 326 self.ctx = self.transport.client
327 327 self.baseurl = svn.ra.get_repos_root(self.ra)
328 328 # Module is either empty or a repository path starting with
329 329 # a slash and not ending with a slash.
330 330 self.module = urlreq.unquote(self.url[len(self.baseurl):])
331 331 self.prevmodule = None
332 332 self.rootmodule = self.module
333 333 self.commits = {}
334 334 self.paths = {}
335 335 self.uuid = svn.ra.get_uuid(self.ra)
336 336 except svn.core.SubversionException:
337 337 ui.traceback()
338 338 svnversion = '%d.%d.%d' % (svn.core.SVN_VER_MAJOR,
339 339 svn.core.SVN_VER_MINOR,
340 340 svn.core.SVN_VER_MICRO)
341 341 raise NoRepo(_("%s does not look like a Subversion repository "
342 342 "to libsvn version %s")
343 343 % (self.url, svnversion))
344 344
345 345 if revs:
346 346 if len(revs) > 1:
347 347 raise error.Abort(_('subversion source does not support '
348 348 'specifying multiple revisions'))
349 349 try:
350 350 latest = int(revs[0])
351 351 except ValueError:
352 352 raise error.Abort(_('svn: revision %s is not an integer') %
353 353 revs[0])
354 354
355 355 trunkcfg = self.ui.config('convert', 'svn.trunk')
356 356 if trunkcfg is None:
357 357 trunkcfg = 'trunk'
358 358 self.trunkname = trunkcfg.strip('/')
359 359 self.startrev = self.ui.config('convert', 'svn.startrev')
360 360 try:
361 361 self.startrev = int(self.startrev)
362 362 if self.startrev < 0:
363 363 self.startrev = 0
364 364 except ValueError:
365 365 raise error.Abort(_('svn: start revision %s is not an integer')
366 366 % self.startrev)
367 367
368 368 try:
369 369 self.head = self.latest(self.module, latest)
370 370 except SvnPathNotFound:
371 371 self.head = None
372 372 if not self.head:
373 373 raise error.Abort(_('no revision found in module %s')
374 374 % self.module)
375 375 self.last_changed = self.revnum(self.head)
376 376
377 377 self._changescache = (None, None)
378 378
379 379 if os.path.exists(os.path.join(url, '.svn/entries')):
380 380 self.wc = url
381 381 else:
382 382 self.wc = None
383 383 self.convertfp = None
384 384
385 385 def setrevmap(self, revmap):
386 386 lastrevs = {}
387 for revid in revmap.iterkeys():
387 for revid in revmap:
388 388 uuid, module, revnum = revsplit(revid)
389 389 lastrevnum = lastrevs.setdefault(module, revnum)
390 390 if revnum > lastrevnum:
391 391 lastrevs[module] = revnum
392 392 self.lastrevs = lastrevs
393 393
394 394 def exists(self, path, optrev):
395 395 try:
396 396 svn.client.ls(self.url.rstrip('/') + '/' + quote(path),
397 397 optrev, False, self.ctx)
398 398 return True
399 399 except svn.core.SubversionException:
400 400 return False
401 401
402 402 def getheads(self):
403 403
404 404 def isdir(path, revnum):
405 405 kind = self._checkpath(path, revnum)
406 406 return kind == svn.core.svn_node_dir
407 407
408 408 def getcfgpath(name, rev):
409 409 cfgpath = self.ui.config('convert', 'svn.' + name)
410 410 if cfgpath is not None and cfgpath.strip() == '':
411 411 return None
412 412 path = (cfgpath or name).strip('/')
413 413 if not self.exists(path, rev):
414 414 if self.module.endswith(path) and name == 'trunk':
415 415 # we are converting from inside this directory
416 416 return None
417 417 if cfgpath:
418 418 raise error.Abort(_('expected %s to be at %r, but not found'
419 419 ) % (name, path))
420 420 return None
421 421 self.ui.note(_('found %s at %r\n') % (name, path))
422 422 return path
423 423
424 424 rev = optrev(self.last_changed)
425 425 oldmodule = ''
426 426 trunk = getcfgpath('trunk', rev)
427 427 self.tags = getcfgpath('tags', rev)
428 428 branches = getcfgpath('branches', rev)
429 429
430 430 # If the project has a trunk or branches, we will extract heads
431 431 # from them. We keep the project root otherwise.
432 432 if trunk:
433 433 oldmodule = self.module or ''
434 434 self.module += '/' + trunk
435 435 self.head = self.latest(self.module, self.last_changed)
436 436 if not self.head:
437 437 raise error.Abort(_('no revision found in module %s')
438 438 % self.module)
439 439
440 440 # First head in the list is the module's head
441 441 self.heads = [self.head]
442 442 if self.tags is not None:
443 443 self.tags = '%s/%s' % (oldmodule , (self.tags or 'tags'))
444 444
445 445 # Check if branches bring a few more heads to the list
446 446 if branches:
447 447 rpath = self.url.strip('/')
448 448 branchnames = svn.client.ls(rpath + '/' + quote(branches),
449 449 rev, False, self.ctx)
450 450 for branch in sorted(branchnames):
451 451 module = '%s/%s/%s' % (oldmodule, branches, branch)
452 452 if not isdir(module, self.last_changed):
453 453 continue
454 454 brevid = self.latest(module, self.last_changed)
455 455 if not brevid:
456 456 self.ui.note(_('ignoring empty branch %s\n') % branch)
457 457 continue
458 458 self.ui.note(_('found branch %s at %d\n') %
459 459 (branch, self.revnum(brevid)))
460 460 self.heads.append(brevid)
461 461
462 462 if self.startrev and self.heads:
463 463 if len(self.heads) > 1:
464 464 raise error.Abort(_('svn: start revision is not supported '
465 465 'with more than one branch'))
466 466 revnum = self.revnum(self.heads[0])
467 467 if revnum < self.startrev:
468 468 raise error.Abort(
469 469 _('svn: no revision found after start revision %d')
470 470 % self.startrev)
471 471
472 472 return self.heads
473 473
474 474 def _getchanges(self, rev, full):
475 475 (paths, parents) = self.paths[rev]
476 476 copies = {}
477 477 if parents:
478 478 files, self.removed, copies = self.expandpaths(rev, paths, parents)
479 479 if full or not parents:
480 480 # Perform a full checkout on roots
481 481 uuid, module, revnum = revsplit(rev)
482 482 entries = svn.client.ls(self.baseurl + quote(module),
483 483 optrev(revnum), True, self.ctx)
484 484 files = [n for n, e in entries.iteritems()
485 485 if e.kind == svn.core.svn_node_file]
486 486 self.removed = set()
487 487
488 488 files.sort()
489 489 files = zip(files, [rev] * len(files))
490 490 return (files, copies)
491 491
492 492 def getchanges(self, rev, full):
493 493 # reuse cache from getchangedfiles
494 494 if self._changescache[0] == rev and not full:
495 495 (files, copies) = self._changescache[1]
496 496 else:
497 497 (files, copies) = self._getchanges(rev, full)
498 498 # caller caches the result, so free it here to release memory
499 499 del self.paths[rev]
500 500 return (files, copies, set())
501 501
502 502 def getchangedfiles(self, rev, i):
503 503 # called from filemap - cache computed values for reuse in getchanges
504 504 (files, copies) = self._getchanges(rev, False)
505 505 self._changescache = (rev, (files, copies))
506 506 return [f[0] for f in files]
507 507
508 508 def getcommit(self, rev):
509 509 if rev not in self.commits:
510 510 uuid, module, revnum = revsplit(rev)
511 511 self.module = module
512 512 self.reparent(module)
513 513 # We assume that:
514 514 # - requests for revisions after "stop" come from the
515 515 # revision graph backward traversal. Cache all of them
516 516 # down to stop, they will be used eventually.
517 517 # - requests for revisions before "stop" come to get
518 518 # isolated branches parents. Just fetch what is needed.
519 519 stop = self.lastrevs.get(module, 0)
520 520 if revnum < stop:
521 521 stop = revnum + 1
522 522 self._fetch_revisions(revnum, stop)
523 523 if rev not in self.commits:
524 524 raise error.Abort(_('svn: revision %s not found') % revnum)
525 525 revcommit = self.commits[rev]
526 526 # caller caches the result, so free it here to release memory
527 527 del self.commits[rev]
528 528 return revcommit
529 529
530 530 def checkrevformat(self, revstr, mapname='splicemap'):
531 531 """ fails if revision format does not match the correct format"""
532 532 if not re.match(r'svn:[0-9a-f]{8,8}-[0-9a-f]{4,4}-'
533 533 r'[0-9a-f]{4,4}-[0-9a-f]{4,4}-[0-9a-f]'
534 534 r'{12,12}(.*)\@[0-9]+$',revstr):
535 535 raise error.Abort(_('%s entry %s is not a valid revision'
536 536 ' identifier') % (mapname, revstr))
537 537
538 538 def numcommits(self):
539 539 return int(self.head.rsplit('@', 1)[1]) - self.startrev
540 540
541 541 def gettags(self):
542 542 tags = {}
543 543 if self.tags is None:
544 544 return tags
545 545
546 546 # svn tags are just a convention, project branches left in a
547 547 # 'tags' directory. There is no other relationship than
548 548 # ancestry, which is expensive to discover and makes them hard
549 549 # to update incrementally. Worse, past revisions may be
550 550 # referenced by tags far away in the future, requiring a deep
551 551 # history traversal on every calculation. Current code
552 552 # performs a single backward traversal, tracking moves within
553 553 # the tags directory (tag renaming) and recording a new tag
554 554 # everytime a project is copied from outside the tags
555 555 # directory. It also lists deleted tags, this behaviour may
556 556 # change in the future.
557 557 pendings = []
558 558 tagspath = self.tags
559 559 start = svn.ra.get_latest_revnum(self.ra)
560 560 stream = self._getlog([self.tags], start, self.startrev)
561 561 try:
562 562 for entry in stream:
563 563 origpaths, revnum, author, date, message = entry
564 564 if not origpaths:
565 565 origpaths = []
566 566 copies = [(e.copyfrom_path, e.copyfrom_rev, p) for p, e
567 567 in origpaths.iteritems() if e.copyfrom_path]
568 568 # Apply moves/copies from more specific to general
569 569 copies.sort(reverse=True)
570 570
571 571 srctagspath = tagspath
572 572 if copies and copies[-1][2] == tagspath:
573 573 # Track tags directory moves
574 574 srctagspath = copies.pop()[0]
575 575
576 576 for source, sourcerev, dest in copies:
577 577 if not dest.startswith(tagspath + '/'):
578 578 continue
579 579 for tag in pendings:
580 580 if tag[0].startswith(dest):
581 581 tagpath = source + tag[0][len(dest):]
582 582 tag[:2] = [tagpath, sourcerev]
583 583 break
584 584 else:
585 585 pendings.append([source, sourcerev, dest])
586 586
587 587 # Filter out tags with children coming from different
588 588 # parts of the repository like:
589 589 # /tags/tag.1 (from /trunk:10)
590 590 # /tags/tag.1/foo (from /branches/foo:12)
591 591 # Here/tags/tag.1 discarded as well as its children.
592 592 # It happens with tools like cvs2svn. Such tags cannot
593 593 # be represented in mercurial.
594 594 addeds = dict((p, e.copyfrom_path) for p, e
595 595 in origpaths.iteritems()
596 596 if e.action == 'A' and e.copyfrom_path)
597 597 badroots = set()
598 598 for destroot in addeds:
599 599 for source, sourcerev, dest in pendings:
600 600 if (not dest.startswith(destroot + '/')
601 601 or source.startswith(addeds[destroot] + '/')):
602 602 continue
603 603 badroots.add(destroot)
604 604 break
605 605
606 606 for badroot in badroots:
607 607 pendings = [p for p in pendings if p[2] != badroot
608 608 and not p[2].startswith(badroot + '/')]
609 609
610 610 # Tell tag renamings from tag creations
611 611 renamings = []
612 612 for source, sourcerev, dest in pendings:
613 613 tagname = dest.split('/')[-1]
614 614 if source.startswith(srctagspath):
615 615 renamings.append([source, sourcerev, tagname])
616 616 continue
617 617 if tagname in tags:
618 618 # Keep the latest tag value
619 619 continue
620 620 # From revision may be fake, get one with changes
621 621 try:
622 622 tagid = self.latest(source, sourcerev)
623 623 if tagid and tagname not in tags:
624 624 tags[tagname] = tagid
625 625 except SvnPathNotFound:
626 626 # It happens when we are following directories
627 627 # we assumed were copied with their parents
628 628 # but were really created in the tag
629 629 # directory.
630 630 pass
631 631 pendings = renamings
632 632 tagspath = srctagspath
633 633 finally:
634 634 stream.close()
635 635 return tags
636 636
637 637 def converted(self, rev, destrev):
638 638 if not self.wc:
639 639 return
640 640 if self.convertfp is None:
641 641 self.convertfp = open(os.path.join(self.wc, '.svn', 'hg-shamap'),
642 642 'ab')
643 643 self.convertfp.write(util.tonativeeol('%s %d\n'
644 644 % (destrev, self.revnum(rev))))
645 645 self.convertfp.flush()
646 646
647 647 def revid(self, revnum, module=None):
648 648 return 'svn:%s%s@%s' % (self.uuid, module or self.module, revnum)
649 649
650 650 def revnum(self, rev):
651 651 return int(rev.split('@')[-1])
652 652
653 653 def latest(self, path, stop=None):
654 654 """Find the latest revid affecting path, up to stop revision
655 655 number. If stop is None, default to repository latest
656 656 revision. It may return a revision in a different module,
657 657 since a branch may be moved without a change being
658 658 reported. Return None if computed module does not belong to
659 659 rootmodule subtree.
660 660 """
661 661 def findchanges(path, start, stop=None):
662 662 stream = self._getlog([path], start, stop or 1)
663 663 try:
664 664 for entry in stream:
665 665 paths, revnum, author, date, message = entry
666 666 if stop is None and paths:
667 667 # We do not know the latest changed revision,
668 668 # keep the first one with changed paths.
669 669 break
670 670 if revnum <= stop:
671 671 break
672 672
673 673 for p in paths:
674 674 if (not path.startswith(p) or
675 675 not paths[p].copyfrom_path):
676 676 continue
677 677 newpath = paths[p].copyfrom_path + path[len(p):]
678 678 self.ui.debug("branch renamed from %s to %s at %d\n" %
679 679 (path, newpath, revnum))
680 680 path = newpath
681 681 break
682 682 if not paths:
683 683 revnum = None
684 684 return revnum, path
685 685 finally:
686 686 stream.close()
687 687
688 688 if not path.startswith(self.rootmodule):
689 689 # Requests on foreign branches may be forbidden at server level
690 690 self.ui.debug('ignoring foreign branch %r\n' % path)
691 691 return None
692 692
693 693 if stop is None:
694 694 stop = svn.ra.get_latest_revnum(self.ra)
695 695 try:
696 696 prevmodule = self.reparent('')
697 697 dirent = svn.ra.stat(self.ra, path.strip('/'), stop)
698 698 self.reparent(prevmodule)
699 699 except svn.core.SubversionException:
700 700 dirent = None
701 701 if not dirent:
702 702 raise SvnPathNotFound(_('%s not found up to revision %d')
703 703 % (path, stop))
704 704
705 705 # stat() gives us the previous revision on this line of
706 706 # development, but it might be in *another module*. Fetch the
707 707 # log and detect renames down to the latest revision.
708 708 revnum, realpath = findchanges(path, stop, dirent.created_rev)
709 709 if revnum is None:
710 710 # Tools like svnsync can create empty revision, when
711 711 # synchronizing only a subtree for instance. These empty
712 712 # revisions created_rev still have their original values
713 713 # despite all changes having disappeared and can be
714 714 # returned by ra.stat(), at least when stating the root
715 715 # module. In that case, do not trust created_rev and scan
716 716 # the whole history.
717 717 revnum, realpath = findchanges(path, stop)
718 718 if revnum is None:
719 719 self.ui.debug('ignoring empty branch %r\n' % realpath)
720 720 return None
721 721
722 722 if not realpath.startswith(self.rootmodule):
723 723 self.ui.debug('ignoring foreign branch %r\n' % realpath)
724 724 return None
725 725 return self.revid(revnum, realpath)
726 726
727 727 def reparent(self, module):
728 728 """Reparent the svn transport and return the previous parent."""
729 729 if self.prevmodule == module:
730 730 return module
731 731 svnurl = self.baseurl + quote(module)
732 732 prevmodule = self.prevmodule
733 733 if prevmodule is None:
734 734 prevmodule = ''
735 735 self.ui.debug("reparent to %s\n" % svnurl)
736 736 svn.ra.reparent(self.ra, svnurl)
737 737 self.prevmodule = module
738 738 return prevmodule
739 739
740 740 def expandpaths(self, rev, paths, parents):
741 741 changed, removed = set(), set()
742 742 copies = {}
743 743
744 744 new_module, revnum = revsplit(rev)[1:]
745 745 if new_module != self.module:
746 746 self.module = new_module
747 747 self.reparent(self.module)
748 748
749 749 for i, (path, ent) in enumerate(paths):
750 750 self.ui.progress(_('scanning paths'), i, item=path,
751 751 total=len(paths), unit=_('paths'))
752 752 entrypath = self.getrelpath(path)
753 753
754 754 kind = self._checkpath(entrypath, revnum)
755 755 if kind == svn.core.svn_node_file:
756 756 changed.add(self.recode(entrypath))
757 757 if not ent.copyfrom_path or not parents:
758 758 continue
759 759 # Copy sources not in parent revisions cannot be
760 760 # represented, ignore their origin for now
761 761 pmodule, prevnum = revsplit(parents[0])[1:]
762 762 if ent.copyfrom_rev < prevnum:
763 763 continue
764 764 copyfrom_path = self.getrelpath(ent.copyfrom_path, pmodule)
765 765 if not copyfrom_path:
766 766 continue
767 767 self.ui.debug("copied to %s from %s@%s\n" %
768 768 (entrypath, copyfrom_path, ent.copyfrom_rev))
769 769 copies[self.recode(entrypath)] = self.recode(copyfrom_path)
770 770 elif kind == 0: # gone, but had better be a deleted *file*
771 771 self.ui.debug("gone from %s\n" % ent.copyfrom_rev)
772 772 pmodule, prevnum = revsplit(parents[0])[1:]
773 773 parentpath = pmodule + "/" + entrypath
774 774 fromkind = self._checkpath(entrypath, prevnum, pmodule)
775 775
776 776 if fromkind == svn.core.svn_node_file:
777 777 removed.add(self.recode(entrypath))
778 778 elif fromkind == svn.core.svn_node_dir:
779 779 oroot = parentpath.strip('/')
780 780 nroot = path.strip('/')
781 781 children = self._iterfiles(oroot, prevnum)
782 782 for childpath in children:
783 783 childpath = childpath.replace(oroot, nroot)
784 784 childpath = self.getrelpath("/" + childpath, pmodule)
785 785 if childpath:
786 786 removed.add(self.recode(childpath))
787 787 else:
788 788 self.ui.debug('unknown path in revision %d: %s\n' % \
789 789 (revnum, path))
790 790 elif kind == svn.core.svn_node_dir:
791 791 if ent.action == 'M':
792 792 # If the directory just had a prop change,
793 793 # then we shouldn't need to look for its children.
794 794 continue
795 795 if ent.action == 'R' and parents:
796 796 # If a directory is replacing a file, mark the previous
797 797 # file as deleted
798 798 pmodule, prevnum = revsplit(parents[0])[1:]
799 799 pkind = self._checkpath(entrypath, prevnum, pmodule)
800 800 if pkind == svn.core.svn_node_file:
801 801 removed.add(self.recode(entrypath))
802 802 elif pkind == svn.core.svn_node_dir:
803 803 # We do not know what files were kept or removed,
804 804 # mark them all as changed.
805 805 for childpath in self._iterfiles(pmodule, prevnum):
806 806 childpath = self.getrelpath("/" + childpath)
807 807 if childpath:
808 808 changed.add(self.recode(childpath))
809 809
810 810 for childpath in self._iterfiles(path, revnum):
811 811 childpath = self.getrelpath("/" + childpath)
812 812 if childpath:
813 813 changed.add(self.recode(childpath))
814 814
815 815 # Handle directory copies
816 816 if not ent.copyfrom_path or not parents:
817 817 continue
818 818 # Copy sources not in parent revisions cannot be
819 819 # represented, ignore their origin for now
820 820 pmodule, prevnum = revsplit(parents[0])[1:]
821 821 if ent.copyfrom_rev < prevnum:
822 822 continue
823 823 copyfrompath = self.getrelpath(ent.copyfrom_path, pmodule)
824 824 if not copyfrompath:
825 825 continue
826 826 self.ui.debug("mark %s came from %s:%d\n"
827 827 % (path, copyfrompath, ent.copyfrom_rev))
828 828 children = self._iterfiles(ent.copyfrom_path, ent.copyfrom_rev)
829 829 for childpath in children:
830 830 childpath = self.getrelpath("/" + childpath, pmodule)
831 831 if not childpath:
832 832 continue
833 833 copytopath = path + childpath[len(copyfrompath):]
834 834 copytopath = self.getrelpath(copytopath)
835 835 copies[self.recode(copytopath)] = self.recode(childpath)
836 836
837 837 self.ui.progress(_('scanning paths'), None)
838 838 changed.update(removed)
839 839 return (list(changed), removed, copies)
840 840
841 841 def _fetch_revisions(self, from_revnum, to_revnum):
842 842 if from_revnum < to_revnum:
843 843 from_revnum, to_revnum = to_revnum, from_revnum
844 844
845 845 self.child_cset = None
846 846
847 847 def parselogentry(orig_paths, revnum, author, date, message):
848 848 """Return the parsed commit object or None, and True if
849 849 the revision is a branch root.
850 850 """
851 851 self.ui.debug("parsing revision %d (%d changes)\n" %
852 852 (revnum, len(orig_paths)))
853 853
854 854 branched = False
855 855 rev = self.revid(revnum)
856 856 # branch log might return entries for a parent we already have
857 857
858 858 if rev in self.commits or revnum < to_revnum:
859 859 return None, branched
860 860
861 861 parents = []
862 862 # check whether this revision is the start of a branch or part
863 863 # of a branch renaming
864 864 orig_paths = sorted(orig_paths.iteritems())
865 865 root_paths = [(p, e) for p, e in orig_paths
866 866 if self.module.startswith(p)]
867 867 if root_paths:
868 868 path, ent = root_paths[-1]
869 869 if ent.copyfrom_path:
870 870 branched = True
871 871 newpath = ent.copyfrom_path + self.module[len(path):]
872 872 # ent.copyfrom_rev may not be the actual last revision
873 873 previd = self.latest(newpath, ent.copyfrom_rev)
874 874 if previd is not None:
875 875 prevmodule, prevnum = revsplit(previd)[1:]
876 876 if prevnum >= self.startrev:
877 877 parents = [previd]
878 878 self.ui.note(
879 879 _('found parent of branch %s at %d: %s\n') %
880 880 (self.module, prevnum, prevmodule))
881 881 else:
882 882 self.ui.debug("no copyfrom path, don't know what to do.\n")
883 883
884 884 paths = []
885 885 # filter out unrelated paths
886 886 for path, ent in orig_paths:
887 887 if self.getrelpath(path) is None:
888 888 continue
889 889 paths.append((path, ent))
890 890
891 891 # Example SVN datetime. Includes microseconds.
892 892 # ISO-8601 conformant
893 893 # '2007-01-04T17:35:00.902377Z'
894 894 date = util.parsedate(date[:19] + " UTC", ["%Y-%m-%dT%H:%M:%S"])
895 895 if self.ui.configbool('convert', 'localtimezone'):
896 896 date = makedatetimestamp(date[0])
897 897
898 898 if message:
899 899 log = self.recode(message)
900 900 else:
901 901 log = ''
902 902
903 903 if author:
904 904 author = self.recode(author)
905 905 else:
906 906 author = ''
907 907
908 908 try:
909 909 branch = self.module.split("/")[-1]
910 910 if branch == self.trunkname:
911 911 branch = None
912 912 except IndexError:
913 913 branch = None
914 914
915 915 cset = commit(author=author,
916 916 date=util.datestr(date, '%Y-%m-%d %H:%M:%S %1%2'),
917 917 desc=log,
918 918 parents=parents,
919 919 branch=branch,
920 920 rev=rev)
921 921
922 922 self.commits[rev] = cset
923 923 # The parents list is *shared* among self.paths and the
924 924 # commit object. Both will be updated below.
925 925 self.paths[rev] = (paths, cset.parents)
926 926 if self.child_cset and not self.child_cset.parents:
927 927 self.child_cset.parents[:] = [rev]
928 928 self.child_cset = cset
929 929 return cset, branched
930 930
931 931 self.ui.note(_('fetching revision log for "%s" from %d to %d\n') %
932 932 (self.module, from_revnum, to_revnum))
933 933
934 934 try:
935 935 firstcset = None
936 936 lastonbranch = False
937 937 stream = self._getlog([self.module], from_revnum, to_revnum)
938 938 try:
939 939 for entry in stream:
940 940 paths, revnum, author, date, message = entry
941 941 if revnum < self.startrev:
942 942 lastonbranch = True
943 943 break
944 944 if not paths:
945 945 self.ui.debug('revision %d has no entries\n' % revnum)
946 946 # If we ever leave the loop on an empty
947 947 # revision, do not try to get a parent branch
948 948 lastonbranch = lastonbranch or revnum == 0
949 949 continue
950 950 cset, lastonbranch = parselogentry(paths, revnum, author,
951 951 date, message)
952 952 if cset:
953 953 firstcset = cset
954 954 if lastonbranch:
955 955 break
956 956 finally:
957 957 stream.close()
958 958
959 959 if not lastonbranch and firstcset and not firstcset.parents:
960 960 # The first revision of the sequence (the last fetched one)
961 961 # has invalid parents if not a branch root. Find the parent
962 962 # revision now, if any.
963 963 try:
964 964 firstrevnum = self.revnum(firstcset.rev)
965 965 if firstrevnum > 1:
966 966 latest = self.latest(self.module, firstrevnum - 1)
967 967 if latest:
968 968 firstcset.parents.append(latest)
969 969 except SvnPathNotFound:
970 970 pass
971 971 except svn.core.SubversionException as xxx_todo_changeme:
972 972 (inst, num) = xxx_todo_changeme.args
973 973 if num == svn.core.SVN_ERR_FS_NO_SUCH_REVISION:
974 974 raise error.Abort(_('svn: branch has no revision %s')
975 975 % to_revnum)
976 976 raise
977 977
978 978 def getfile(self, file, rev):
979 979 # TODO: ra.get_file transmits the whole file instead of diffs.
980 980 if file in self.removed:
981 981 return None, None
982 982 mode = ''
983 983 try:
984 984 new_module, revnum = revsplit(rev)[1:]
985 985 if self.module != new_module:
986 986 self.module = new_module
987 987 self.reparent(self.module)
988 988 io = stringio()
989 989 info = svn.ra.get_file(self.ra, file, revnum, io)
990 990 data = io.getvalue()
991 991 # ra.get_file() seems to keep a reference on the input buffer
992 992 # preventing collection. Release it explicitly.
993 993 io.close()
994 994 if isinstance(info, list):
995 995 info = info[-1]
996 996 mode = ("svn:executable" in info) and 'x' or ''
997 997 mode = ("svn:special" in info) and 'l' or mode
998 998 except svn.core.SubversionException as e:
999 999 notfound = (svn.core.SVN_ERR_FS_NOT_FOUND,
1000 1000 svn.core.SVN_ERR_RA_DAV_PATH_NOT_FOUND)
1001 1001 if e.apr_err in notfound: # File not found
1002 1002 return None, None
1003 1003 raise
1004 1004 if mode == 'l':
1005 1005 link_prefix = "link "
1006 1006 if data.startswith(link_prefix):
1007 1007 data = data[len(link_prefix):]
1008 1008 return data, mode
1009 1009
1010 1010 def _iterfiles(self, path, revnum):
1011 1011 """Enumerate all files in path at revnum, recursively."""
1012 1012 path = path.strip('/')
1013 1013 pool = svn.core.Pool()
1014 1014 rpath = '/'.join([self.baseurl, quote(path)]).strip('/')
1015 1015 entries = svn.client.ls(rpath, optrev(revnum), True, self.ctx, pool)
1016 1016 if path:
1017 1017 path += '/'
1018 1018 return ((path + p) for p, e in entries.iteritems()
1019 1019 if e.kind == svn.core.svn_node_file)
1020 1020
1021 1021 def getrelpath(self, path, module=None):
1022 1022 if module is None:
1023 1023 module = self.module
1024 1024 # Given the repository url of this wc, say
1025 1025 # "http://server/plone/CMFPlone/branches/Plone-2_0-branch"
1026 1026 # extract the "entry" portion (a relative path) from what
1027 1027 # svn log --xml says, i.e.
1028 1028 # "/CMFPlone/branches/Plone-2_0-branch/tests/PloneTestCase.py"
1029 1029 # that is to say "tests/PloneTestCase.py"
1030 1030 if path.startswith(module):
1031 1031 relative = path.rstrip('/')[len(module):]
1032 1032 if relative.startswith('/'):
1033 1033 return relative[1:]
1034 1034 elif relative == '':
1035 1035 return relative
1036 1036
1037 1037 # The path is outside our tracked tree...
1038 1038 self.ui.debug('%r is not under %r, ignoring\n' % (path, module))
1039 1039 return None
1040 1040
1041 1041 def _checkpath(self, path, revnum, module=None):
1042 1042 if module is not None:
1043 1043 prevmodule = self.reparent('')
1044 1044 path = module + '/' + path
1045 1045 try:
1046 1046 # ra.check_path does not like leading slashes very much, it leads
1047 1047 # to PROPFIND subversion errors
1048 1048 return svn.ra.check_path(self.ra, path.strip('/'), revnum)
1049 1049 finally:
1050 1050 if module is not None:
1051 1051 self.reparent(prevmodule)
1052 1052
1053 1053 def _getlog(self, paths, start, end, limit=0, discover_changed_paths=True,
1054 1054 strict_node_history=False):
1055 1055 # Normalize path names, svn >= 1.5 only wants paths relative to
1056 1056 # supplied URL
1057 1057 relpaths = []
1058 1058 for p in paths:
1059 1059 if not p.startswith('/'):
1060 1060 p = self.module + '/' + p
1061 1061 relpaths.append(p.strip('/'))
1062 1062 args = [self.baseurl, relpaths, start, end, limit,
1063 1063 discover_changed_paths, strict_node_history]
1064 1064 # developer config: convert.svn.debugsvnlog
1065 1065 if not self.ui.configbool('convert', 'svn.debugsvnlog'):
1066 1066 return directlogstream(*args)
1067 1067 arg = encodeargs(args)
1068 1068 hgexe = util.hgexecutable()
1069 1069 cmd = '%s debugsvnlog' % util.shellquote(hgexe)
1070 1070 stdin, stdout = util.popen2(util.quotecommand(cmd))
1071 1071 stdin.write(arg)
1072 1072 try:
1073 1073 stdin.close()
1074 1074 except IOError:
1075 1075 raise error.Abort(_('Mercurial failed to run itself, check'
1076 1076 ' hg executable is in PATH'))
1077 1077 return logstream(stdout)
1078 1078
1079 1079 pre_revprop_change = '''#!/bin/sh
1080 1080
1081 1081 REPOS="$1"
1082 1082 REV="$2"
1083 1083 USER="$3"
1084 1084 PROPNAME="$4"
1085 1085 ACTION="$5"
1086 1086
1087 1087 if [ "$ACTION" = "M" -a "$PROPNAME" = "svn:log" ]; then exit 0; fi
1088 1088 if [ "$ACTION" = "A" -a "$PROPNAME" = "hg:convert-branch" ]; then exit 0; fi
1089 1089 if [ "$ACTION" = "A" -a "$PROPNAME" = "hg:convert-rev" ]; then exit 0; fi
1090 1090
1091 1091 echo "Changing prohibited revision property" >&2
1092 1092 exit 1
1093 1093 '''
1094 1094
1095 1095 class svn_sink(converter_sink, commandline):
1096 1096 commit_re = re.compile(r'Committed revision (\d+).', re.M)
1097 1097 uuid_re = re.compile(r'Repository UUID:\s*(\S+)', re.M)
1098 1098
1099 1099 def prerun(self):
1100 1100 if self.wc:
1101 1101 os.chdir(self.wc)
1102 1102
1103 1103 def postrun(self):
1104 1104 if self.wc:
1105 1105 os.chdir(self.cwd)
1106 1106
1107 1107 def join(self, name):
1108 1108 return os.path.join(self.wc, '.svn', name)
1109 1109
1110 1110 def revmapfile(self):
1111 1111 return self.join('hg-shamap')
1112 1112
1113 1113 def authorfile(self):
1114 1114 return self.join('hg-authormap')
1115 1115
1116 1116 def __init__(self, ui, repotype, path):
1117 1117
1118 1118 converter_sink.__init__(self, ui, repotype, path)
1119 1119 commandline.__init__(self, ui, 'svn')
1120 1120 self.delete = []
1121 1121 self.setexec = []
1122 1122 self.delexec = []
1123 1123 self.copies = []
1124 1124 self.wc = None
1125 1125 self.cwd = pycompat.getcwd()
1126 1126
1127 1127 created = False
1128 1128 if os.path.isfile(os.path.join(path, '.svn', 'entries')):
1129 1129 self.wc = os.path.realpath(path)
1130 1130 self.run0('update')
1131 1131 else:
1132 1132 if not re.search(r'^(file|http|https|svn|svn\+ssh)\://', path):
1133 1133 path = os.path.realpath(path)
1134 1134 if os.path.isdir(os.path.dirname(path)):
1135 1135 if not os.path.exists(os.path.join(path, 'db', 'fs-type')):
1136 1136 ui.status(_('initializing svn repository %r\n') %
1137 1137 os.path.basename(path))
1138 1138 commandline(ui, 'svnadmin').run0('create', path)
1139 1139 created = path
1140 1140 path = util.normpath(path)
1141 1141 if not path.startswith('/'):
1142 1142 path = '/' + path
1143 1143 path = 'file://' + path
1144 1144
1145 1145 wcpath = os.path.join(pycompat.getcwd(), os.path.basename(path) +
1146 1146 '-wc')
1147 1147 ui.status(_('initializing svn working copy %r\n')
1148 1148 % os.path.basename(wcpath))
1149 1149 self.run0('checkout', path, wcpath)
1150 1150
1151 1151 self.wc = wcpath
1152 1152 self.opener = vfsmod.vfs(self.wc)
1153 1153 self.wopener = vfsmod.vfs(self.wc)
1154 1154 self.childmap = mapfile(ui, self.join('hg-childmap'))
1155 1155 if util.checkexec(self.wc):
1156 1156 self.is_exec = util.isexec
1157 1157 else:
1158 1158 self.is_exec = None
1159 1159
1160 1160 if created:
1161 1161 hook = os.path.join(created, 'hooks', 'pre-revprop-change')
1162 1162 fp = open(hook, 'wb')
1163 1163 fp.write(pre_revprop_change)
1164 1164 fp.close()
1165 1165 util.setflags(hook, False, True)
1166 1166
1167 1167 output = self.run0('info')
1168 1168 self.uuid = self.uuid_re.search(output).group(1).strip()
1169 1169
1170 1170 def wjoin(self, *names):
1171 1171 return os.path.join(self.wc, *names)
1172 1172
1173 1173 @propertycache
1174 1174 def manifest(self):
1175 1175 # As of svn 1.7, the "add" command fails when receiving
1176 1176 # already tracked entries, so we have to track and filter them
1177 1177 # ourselves.
1178 1178 m = set()
1179 1179 output = self.run0('ls', recursive=True, xml=True)
1180 1180 doc = xml.dom.minidom.parseString(output)
1181 1181 for e in doc.getElementsByTagName('entry'):
1182 1182 for n in e.childNodes:
1183 1183 if n.nodeType != n.ELEMENT_NODE or n.tagName != 'name':
1184 1184 continue
1185 1185 name = ''.join(c.data for c in n.childNodes
1186 1186 if c.nodeType == c.TEXT_NODE)
1187 1187 # Entries are compared with names coming from
1188 1188 # mercurial, so bytes with undefined encoding. Our
1189 1189 # best bet is to assume they are in local
1190 1190 # encoding. They will be passed to command line calls
1191 1191 # later anyway, so they better be.
1192 1192 m.add(encoding.unitolocal(name))
1193 1193 break
1194 1194 return m
1195 1195
1196 1196 def putfile(self, filename, flags, data):
1197 1197 if 'l' in flags:
1198 1198 self.wopener.symlink(data, filename)
1199 1199 else:
1200 1200 try:
1201 1201 if os.path.islink(self.wjoin(filename)):
1202 1202 os.unlink(filename)
1203 1203 except OSError:
1204 1204 pass
1205 1205 self.wopener.write(filename, data)
1206 1206
1207 1207 if self.is_exec:
1208 1208 if self.is_exec(self.wjoin(filename)):
1209 1209 if 'x' not in flags:
1210 1210 self.delexec.append(filename)
1211 1211 else:
1212 1212 if 'x' in flags:
1213 1213 self.setexec.append(filename)
1214 1214 util.setflags(self.wjoin(filename), False, 'x' in flags)
1215 1215
1216 1216 def _copyfile(self, source, dest):
1217 1217 # SVN's copy command pukes if the destination file exists, but
1218 1218 # our copyfile method expects to record a copy that has
1219 1219 # already occurred. Cross the semantic gap.
1220 1220 wdest = self.wjoin(dest)
1221 1221 exists = os.path.lexists(wdest)
1222 1222 if exists:
1223 1223 fd, tempname = tempfile.mkstemp(
1224 1224 prefix='hg-copy-', dir=os.path.dirname(wdest))
1225 1225 os.close(fd)
1226 1226 os.unlink(tempname)
1227 1227 os.rename(wdest, tempname)
1228 1228 try:
1229 1229 self.run0('copy', source, dest)
1230 1230 finally:
1231 1231 self.manifest.add(dest)
1232 1232 if exists:
1233 1233 try:
1234 1234 os.unlink(wdest)
1235 1235 except OSError:
1236 1236 pass
1237 1237 os.rename(tempname, wdest)
1238 1238
1239 1239 def dirs_of(self, files):
1240 1240 dirs = set()
1241 1241 for f in files:
1242 1242 if os.path.isdir(self.wjoin(f)):
1243 1243 dirs.add(f)
1244 1244 i = len(f)
1245 1245 for i in iter(lambda: f.rfind('/', 0, i), -1):
1246 1246 dirs.add(f[:i])
1247 1247 return dirs
1248 1248
1249 1249 def add_dirs(self, files):
1250 1250 add_dirs = [d for d in sorted(self.dirs_of(files))
1251 1251 if d not in self.manifest]
1252 1252 if add_dirs:
1253 1253 self.manifest.update(add_dirs)
1254 1254 self.xargs(add_dirs, 'add', non_recursive=True, quiet=True)
1255 1255 return add_dirs
1256 1256
1257 1257 def add_files(self, files):
1258 1258 files = [f for f in files if f not in self.manifest]
1259 1259 if files:
1260 1260 self.manifest.update(files)
1261 1261 self.xargs(files, 'add', quiet=True)
1262 1262 return files
1263 1263
1264 1264 def addchild(self, parent, child):
1265 1265 self.childmap[parent] = child
1266 1266
1267 1267 def revid(self, rev):
1268 1268 return u"svn:%s@%s" % (self.uuid, rev)
1269 1269
1270 1270 def putcommit(self, files, copies, parents, commit, source, revmap, full,
1271 1271 cleanp2):
1272 1272 for parent in parents:
1273 1273 try:
1274 1274 return self.revid(self.childmap[parent])
1275 1275 except KeyError:
1276 1276 pass
1277 1277
1278 1278 # Apply changes to working copy
1279 1279 for f, v in files:
1280 1280 data, mode = source.getfile(f, v)
1281 1281 if data is None:
1282 1282 self.delete.append(f)
1283 1283 else:
1284 1284 self.putfile(f, mode, data)
1285 1285 if f in copies:
1286 1286 self.copies.append([copies[f], f])
1287 1287 if full:
1288 1288 self.delete.extend(sorted(self.manifest.difference(files)))
1289 1289 files = [f[0] for f in files]
1290 1290
1291 1291 entries = set(self.delete)
1292 1292 files = frozenset(files)
1293 1293 entries.update(self.add_dirs(files.difference(entries)))
1294 1294 if self.copies:
1295 1295 for s, d in self.copies:
1296 1296 self._copyfile(s, d)
1297 1297 self.copies = []
1298 1298 if self.delete:
1299 1299 self.xargs(self.delete, 'delete')
1300 1300 for f in self.delete:
1301 1301 self.manifest.remove(f)
1302 1302 self.delete = []
1303 1303 entries.update(self.add_files(files.difference(entries)))
1304 1304 if self.delexec:
1305 1305 self.xargs(self.delexec, 'propdel', 'svn:executable')
1306 1306 self.delexec = []
1307 1307 if self.setexec:
1308 1308 self.xargs(self.setexec, 'propset', 'svn:executable', '*')
1309 1309 self.setexec = []
1310 1310
1311 1311 fd, messagefile = tempfile.mkstemp(prefix='hg-convert-')
1312 1312 fp = os.fdopen(fd, pycompat.sysstr('wb'))
1313 1313 fp.write(util.tonativeeol(commit.desc))
1314 1314 fp.close()
1315 1315 try:
1316 1316 output = self.run0('commit',
1317 1317 username=util.shortuser(commit.author),
1318 1318 file=messagefile,
1319 1319 encoding='utf-8')
1320 1320 try:
1321 1321 rev = self.commit_re.search(output).group(1)
1322 1322 except AttributeError:
1323 1323 if parents and not files:
1324 1324 return parents[0]
1325 1325 self.ui.warn(_('unexpected svn output:\n'))
1326 1326 self.ui.warn(output)
1327 1327 raise error.Abort(_('unable to cope with svn output'))
1328 1328 if commit.rev:
1329 1329 self.run('propset', 'hg:convert-rev', commit.rev,
1330 1330 revprop=True, revision=rev)
1331 1331 if commit.branch and commit.branch != 'default':
1332 1332 self.run('propset', 'hg:convert-branch', commit.branch,
1333 1333 revprop=True, revision=rev)
1334 1334 for parent in parents:
1335 1335 self.addchild(parent, rev)
1336 1336 return self.revid(rev)
1337 1337 finally:
1338 1338 os.unlink(messagefile)
1339 1339
1340 1340 def puttags(self, tags):
1341 1341 self.ui.warn(_('writing Subversion tags is not yet implemented\n'))
1342 1342 return None, None
1343 1343
1344 1344 def hascommitfrommap(self, rev):
1345 1345 # We trust that revisions referenced in a map still is present
1346 1346 # TODO: implement something better if necessary and feasible
1347 1347 return True
1348 1348
1349 1349 def hascommitforsplicemap(self, rev):
1350 1350 # This is not correct as one can convert to an existing subversion
1351 1351 # repository and childmap would not list all revisions. Too bad.
1352 1352 if rev in self.childmap:
1353 1353 return True
1354 1354 raise error.Abort(_('splice map revision %s not found in subversion '
1355 1355 'child map (revision lookups are not implemented)')
1356 1356 % rev)
@@ -1,1870 +1,1870
1 1 # rebase.py - rebasing feature for mercurial
2 2 #
3 3 # Copyright 2008 Stefano Tortarolo <stefano.tortarolo at gmail dot com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 '''command to move sets of revisions to a different ancestor
9 9
10 10 This extension lets you rebase changesets in an existing Mercurial
11 11 repository.
12 12
13 13 For more information:
14 14 https://mercurial-scm.org/wiki/RebaseExtension
15 15 '''
16 16
17 17 from __future__ import absolute_import
18 18
19 19 import errno
20 20 import os
21 21
22 22 from mercurial.i18n import _
23 23 from mercurial.node import (
24 24 nullid,
25 25 nullrev,
26 26 short,
27 27 )
28 28 from mercurial import (
29 29 bookmarks,
30 30 cmdutil,
31 31 commands,
32 32 copies,
33 33 destutil,
34 34 dirstateguard,
35 35 error,
36 36 extensions,
37 37 hg,
38 38 lock,
39 39 merge as mergemod,
40 40 mergeutil,
41 41 obsolete,
42 42 obsutil,
43 43 patch,
44 44 phases,
45 45 pycompat,
46 46 registrar,
47 47 repair,
48 48 revset,
49 49 revsetlang,
50 50 scmutil,
51 51 smartset,
52 52 util,
53 53 )
54 54
55 55 release = lock.release
56 56
57 57 # The following constants are used throughout the rebase module. The ordering of
58 58 # their values must be maintained.
59 59
60 60 # Indicates that a revision needs to be rebased
61 61 revtodo = -1
62 62 revtodostr = '-1'
63 63
64 64 # legacy revstates no longer needed in current code
65 65 # -2: nullmerge, -3: revignored, -4: revprecursor, -5: revpruned
66 66 legacystates = {'-2', '-3', '-4', '-5'}
67 67
68 68 cmdtable = {}
69 69 command = registrar.command(cmdtable)
70 70 # Note for extension authors: ONLY specify testedwith = 'ships-with-hg-core' for
71 71 # extensions which SHIP WITH MERCURIAL. Non-mainline extensions should
72 72 # be specifying the version(s) of Mercurial they are tested with, or
73 73 # leave the attribute unspecified.
74 74 testedwith = 'ships-with-hg-core'
75 75
76 76 def _nothingtorebase():
77 77 return 1
78 78
79 79 def _savegraft(ctx, extra):
80 80 s = ctx.extra().get('source', None)
81 81 if s is not None:
82 82 extra['source'] = s
83 83 s = ctx.extra().get('intermediate-source', None)
84 84 if s is not None:
85 85 extra['intermediate-source'] = s
86 86
87 87 def _savebranch(ctx, extra):
88 88 extra['branch'] = ctx.branch()
89 89
90 90 def _makeextrafn(copiers):
91 91 """make an extrafn out of the given copy-functions.
92 92
93 93 A copy function takes a context and an extra dict, and mutates the
94 94 extra dict as needed based on the given context.
95 95 """
96 96 def extrafn(ctx, extra):
97 97 for c in copiers:
98 98 c(ctx, extra)
99 99 return extrafn
100 100
101 101 def _destrebase(repo, sourceset, destspace=None):
102 102 """small wrapper around destmerge to pass the right extra args
103 103
104 104 Please wrap destutil.destmerge instead."""
105 105 return destutil.destmerge(repo, action='rebase', sourceset=sourceset,
106 106 onheadcheck=False, destspace=destspace)
107 107
108 108 revsetpredicate = registrar.revsetpredicate()
109 109
110 110 @revsetpredicate('_destrebase')
111 111 def _revsetdestrebase(repo, subset, x):
112 112 # ``_rebasedefaultdest()``
113 113
114 114 # default destination for rebase.
115 115 # # XXX: Currently private because I expect the signature to change.
116 116 # # XXX: - bailing out in case of ambiguity vs returning all data.
117 117 # i18n: "_rebasedefaultdest" is a keyword
118 118 sourceset = None
119 119 if x is not None:
120 120 sourceset = revset.getset(repo, smartset.fullreposet(repo), x)
121 121 return subset & smartset.baseset([_destrebase(repo, sourceset)])
122 122
123 123 def _ctxdesc(ctx):
124 124 """short description for a context"""
125 125 desc = '%d:%s "%s"' % (ctx.rev(), ctx,
126 126 ctx.description().split('\n', 1)[0])
127 127 repo = ctx.repo()
128 128 names = []
129 129 for nsname, ns in repo.names.iteritems():
130 130 if nsname == 'branches':
131 131 continue
132 132 names.extend(ns.names(repo, ctx.node()))
133 133 if names:
134 134 desc += ' (%s)' % ' '.join(names)
135 135 return desc
136 136
137 137 class rebaseruntime(object):
138 138 """This class is a container for rebase runtime state"""
139 139 def __init__(self, repo, ui, inmemory=False, opts=None):
140 140 if opts is None:
141 141 opts = {}
142 142
143 143 # prepared: whether we have rebasestate prepared or not. Currently it
144 144 # decides whether "self.repo" is unfiltered or not.
145 145 # The rebasestate has explicit hash to hash instructions not depending
146 146 # on visibility. If rebasestate exists (in-memory or on-disk), use
147 147 # unfiltered repo to avoid visibility issues.
148 148 # Before knowing rebasestate (i.e. when starting a new rebase (not
149 149 # --continue or --abort)), the original repo should be used so
150 150 # visibility-dependent revsets are correct.
151 151 self.prepared = False
152 152 self._repo = repo
153 153
154 154 self.ui = ui
155 155 self.opts = opts
156 156 self.originalwd = None
157 157 self.external = nullrev
158 158 # Mapping between the old revision id and either what is the new rebased
159 159 # revision or what needs to be done with the old revision. The state
160 160 # dict will be what contains most of the rebase progress state.
161 161 self.state = {}
162 162 self.activebookmark = None
163 163 self.destmap = {}
164 164 self.skipped = set()
165 165
166 166 self.collapsef = opts.get('collapse', False)
167 167 self.collapsemsg = cmdutil.logmessage(ui, opts)
168 168 self.date = opts.get('date', None)
169 169
170 170 e = opts.get('extrafn') # internal, used by e.g. hgsubversion
171 171 self.extrafns = [_savegraft]
172 172 if e:
173 173 self.extrafns = [e]
174 174
175 175 self.keepf = opts.get('keep', False)
176 176 self.keepbranchesf = opts.get('keepbranches', False)
177 177 # keepopen is not meant for use on the command line, but by
178 178 # other extensions
179 179 self.keepopen = opts.get('keepopen', False)
180 180 self.obsoletenotrebased = {}
181 181 self.obsoletewithoutsuccessorindestination = set()
182 182 self.inmemory = inmemory
183 183
184 184 @property
185 185 def repo(self):
186 186 if self.prepared:
187 187 return self._repo.unfiltered()
188 188 else:
189 189 return self._repo
190 190
191 191 def storestatus(self, tr=None):
192 192 """Store the current status to allow recovery"""
193 193 if tr:
194 194 tr.addfilegenerator('rebasestate', ('rebasestate',),
195 195 self._writestatus, location='plain')
196 196 else:
197 197 with self.repo.vfs("rebasestate", "w") as f:
198 198 self._writestatus(f)
199 199
200 200 def _writestatus(self, f):
201 201 repo = self.repo
202 202 assert repo.filtername is None
203 203 f.write(repo[self.originalwd].hex() + '\n')
204 204 # was "dest". we now write dest per src root below.
205 205 f.write('\n')
206 206 f.write(repo[self.external].hex() + '\n')
207 207 f.write('%d\n' % int(self.collapsef))
208 208 f.write('%d\n' % int(self.keepf))
209 209 f.write('%d\n' % int(self.keepbranchesf))
210 210 f.write('%s\n' % (self.activebookmark or ''))
211 211 destmap = self.destmap
212 212 for d, v in self.state.iteritems():
213 213 oldrev = repo[d].hex()
214 214 if v >= 0:
215 215 newrev = repo[v].hex()
216 216 else:
217 217 newrev = "%d" % v
218 218 destnode = repo[destmap[d]].hex()
219 219 f.write("%s:%s:%s\n" % (oldrev, newrev, destnode))
220 220 repo.ui.debug('rebase status stored\n')
221 221
222 222 def restorestatus(self):
223 223 """Restore a previously stored status"""
224 224 self.prepared = True
225 225 repo = self.repo
226 226 assert repo.filtername is None
227 227 keepbranches = None
228 228 legacydest = None
229 229 collapse = False
230 230 external = nullrev
231 231 activebookmark = None
232 232 state = {}
233 233 destmap = {}
234 234
235 235 try:
236 236 f = repo.vfs("rebasestate")
237 237 for i, l in enumerate(f.read().splitlines()):
238 238 if i == 0:
239 239 originalwd = repo[l].rev()
240 240 elif i == 1:
241 241 # this line should be empty in newer version. but legacy
242 242 # clients may still use it
243 243 if l:
244 244 legacydest = repo[l].rev()
245 245 elif i == 2:
246 246 external = repo[l].rev()
247 247 elif i == 3:
248 248 collapse = bool(int(l))
249 249 elif i == 4:
250 250 keep = bool(int(l))
251 251 elif i == 5:
252 252 keepbranches = bool(int(l))
253 253 elif i == 6 and not (len(l) == 81 and ':' in l):
254 254 # line 6 is a recent addition, so for backwards
255 255 # compatibility check that the line doesn't look like the
256 256 # oldrev:newrev lines
257 257 activebookmark = l
258 258 else:
259 259 args = l.split(':')
260 260 oldrev = args[0]
261 261 newrev = args[1]
262 262 if newrev in legacystates:
263 263 continue
264 264 if len(args) > 2:
265 265 destnode = args[2]
266 266 else:
267 267 destnode = legacydest
268 268 destmap[repo[oldrev].rev()] = repo[destnode].rev()
269 269 if newrev in (nullid, revtodostr):
270 270 state[repo[oldrev].rev()] = revtodo
271 271 # Legacy compat special case
272 272 else:
273 273 state[repo[oldrev].rev()] = repo[newrev].rev()
274 274
275 275 except IOError as err:
276 276 if err.errno != errno.ENOENT:
277 277 raise
278 278 cmdutil.wrongtooltocontinue(repo, _('rebase'))
279 279
280 280 if keepbranches is None:
281 281 raise error.Abort(_('.hg/rebasestate is incomplete'))
282 282
283 283 skipped = set()
284 284 # recompute the set of skipped revs
285 285 if not collapse:
286 286 seen = set(destmap.values())
287 287 for old, new in sorted(state.items()):
288 288 if new != revtodo and new in seen:
289 289 skipped.add(old)
290 290 seen.add(new)
291 291 repo.ui.debug('computed skipped revs: %s\n' %
292 292 (' '.join('%d' % r for r in sorted(skipped)) or ''))
293 293 repo.ui.debug('rebase status resumed\n')
294 294
295 295 self.originalwd = originalwd
296 296 self.destmap = destmap
297 297 self.state = state
298 298 self.skipped = skipped
299 299 self.collapsef = collapse
300 300 self.keepf = keep
301 301 self.keepbranchesf = keepbranches
302 302 self.external = external
303 303 self.activebookmark = activebookmark
304 304
305 305 def _handleskippingobsolete(self, obsoleterevs, destmap):
306 306 """Compute structures necessary for skipping obsolete revisions
307 307
308 308 obsoleterevs: iterable of all obsolete revisions in rebaseset
309 309 destmap: {srcrev: destrev} destination revisions
310 310 """
311 311 self.obsoletenotrebased = {}
312 312 if not self.ui.configbool('experimental', 'rebaseskipobsolete'):
313 313 return
314 314 obsoleteset = set(obsoleterevs)
315 315 (self.obsoletenotrebased,
316 316 self.obsoletewithoutsuccessorindestination,
317 317 obsoleteextinctsuccessors) = _computeobsoletenotrebased(
318 318 self.repo, obsoleteset, destmap)
319 319 skippedset = set(self.obsoletenotrebased)
320 320 skippedset.update(self.obsoletewithoutsuccessorindestination)
321 321 skippedset.update(obsoleteextinctsuccessors)
322 322 _checkobsrebase(self.repo, self.ui, obsoleteset, skippedset)
323 323
324 324 def _prepareabortorcontinue(self, isabort):
325 325 try:
326 326 self.restorestatus()
327 327 self.collapsemsg = restorecollapsemsg(self.repo, isabort)
328 328 except error.RepoLookupError:
329 329 if isabort:
330 330 clearstatus(self.repo)
331 331 clearcollapsemsg(self.repo)
332 332 self.repo.ui.warn(_('rebase aborted (no revision is removed,'
333 333 ' only broken state is cleared)\n'))
334 334 return 0
335 335 else:
336 336 msg = _('cannot continue inconsistent rebase')
337 337 hint = _('use "hg rebase --abort" to clear broken state')
338 338 raise error.Abort(msg, hint=hint)
339 339 if isabort:
340 340 return abort(self.repo, self.originalwd, self.destmap,
341 341 self.state, activebookmark=self.activebookmark)
342 342
343 343 def _preparenewrebase(self, destmap):
344 344 if not destmap:
345 345 return _nothingtorebase()
346 346
347 347 rebaseset = destmap.keys()
348 348 allowunstable = obsolete.isenabled(self.repo, obsolete.allowunstableopt)
349 349 if (not (self.keepf or allowunstable)
350 350 and self.repo.revs('first(children(%ld) - %ld)',
351 351 rebaseset, rebaseset)):
352 352 raise error.Abort(
353 353 _("can't remove original changesets with"
354 354 " unrebased descendants"),
355 355 hint=_('use --keep to keep original changesets'))
356 356
357 357 result = buildstate(self.repo, destmap, self.collapsef)
358 358
359 359 if not result:
360 360 # Empty state built, nothing to rebase
361 361 self.ui.status(_('nothing to rebase\n'))
362 362 return _nothingtorebase()
363 363
364 364 for root in self.repo.set('roots(%ld)', rebaseset):
365 365 if not self.keepf and not root.mutable():
366 366 raise error.Abort(_("can't rebase public changeset %s")
367 367 % root,
368 368 hint=_("see 'hg help phases' for details"))
369 369
370 370 (self.originalwd, self.destmap, self.state) = result
371 371 if self.collapsef:
372 372 dests = set(self.destmap.values())
373 373 if len(dests) != 1:
374 374 raise error.Abort(
375 375 _('--collapse does not work with multiple destinations'))
376 376 destrev = next(iter(dests))
377 377 destancestors = self.repo.changelog.ancestors([destrev],
378 378 inclusive=True)
379 379 self.external = externalparent(self.repo, self.state, destancestors)
380 380
381 381 for destrev in sorted(set(destmap.values())):
382 382 dest = self.repo[destrev]
383 383 if dest.closesbranch() and not self.keepbranchesf:
384 384 self.ui.status(_('reopening closed branch head %s\n') % dest)
385 385
386 386 self.prepared = True
387 387
388 388 def _assignworkingcopy(self):
389 389 if self.inmemory:
390 390 from mercurial.context import overlayworkingctx
391 391 self.wctx = overlayworkingctx(self.repo)
392 392 self.repo.ui.debug("rebasing in-memory\n")
393 393 else:
394 394 self.wctx = self.repo[None]
395 395 self.repo.ui.debug("rebasing on disk\n")
396 396 self.repo.ui.log("rebase", "", rebase_imm_used=self.wctx.isinmemory())
397 397
398 398 def _performrebase(self, tr):
399 399 self._assignworkingcopy()
400 400 repo, ui = self.repo, self.ui
401 401 if self.keepbranchesf:
402 402 # insert _savebranch at the start of extrafns so if
403 403 # there's a user-provided extrafn it can clobber branch if
404 404 # desired
405 405 self.extrafns.insert(0, _savebranch)
406 406 if self.collapsef:
407 407 branches = set()
408 408 for rev in self.state:
409 409 branches.add(repo[rev].branch())
410 410 if len(branches) > 1:
411 411 raise error.Abort(_('cannot collapse multiple named '
412 412 'branches'))
413 413
414 414 # Calculate self.obsoletenotrebased
415 415 obsrevs = _filterobsoleterevs(self.repo, self.state)
416 416 self._handleskippingobsolete(obsrevs, self.destmap)
417 417
418 418 # Keep track of the active bookmarks in order to reset them later
419 419 self.activebookmark = self.activebookmark or repo._activebookmark
420 420 if self.activebookmark:
421 421 bookmarks.deactivate(repo)
422 422
423 423 # Store the state before we begin so users can run 'hg rebase --abort'
424 424 # if we fail before the transaction closes.
425 425 self.storestatus()
426 426
427 427 cands = [k for k, v in self.state.iteritems() if v == revtodo]
428 428 total = len(cands)
429 429 pos = 0
430 430 for subset in sortsource(self.destmap):
431 431 pos = self._performrebasesubset(tr, subset, pos, total)
432 432 ui.progress(_('rebasing'), None)
433 433 ui.note(_('rebase merging completed\n'))
434 434
435 435 def _performrebasesubset(self, tr, subset, pos, total):
436 436 repo, ui, opts = self.repo, self.ui, self.opts
437 437 sortedrevs = repo.revs('sort(%ld, -topo)', subset)
438 438 allowdivergence = self.ui.configbool(
439 439 'experimental', 'evolution.allowdivergence')
440 440 if not allowdivergence:
441 441 sortedrevs -= repo.revs(
442 442 'descendants(%ld) and not %ld',
443 443 self.obsoletewithoutsuccessorindestination,
444 444 self.obsoletewithoutsuccessorindestination,
445 445 )
446 446 for rev in sortedrevs:
447 447 dest = self.destmap[rev]
448 448 ctx = repo[rev]
449 449 desc = _ctxdesc(ctx)
450 450 if self.state[rev] == rev:
451 451 ui.status(_('already rebased %s\n') % desc)
452 452 elif (not allowdivergence
453 453 and rev in self.obsoletewithoutsuccessorindestination):
454 454 msg = _('note: not rebasing %s and its descendants as '
455 455 'this would cause divergence\n') % desc
456 456 repo.ui.status(msg)
457 457 self.skipped.add(rev)
458 458 elif rev in self.obsoletenotrebased:
459 459 succ = self.obsoletenotrebased[rev]
460 460 if succ is None:
461 461 msg = _('note: not rebasing %s, it has no '
462 462 'successor\n') % desc
463 463 else:
464 464 succdesc = _ctxdesc(repo[succ])
465 465 msg = (_('note: not rebasing %s, already in '
466 466 'destination as %s\n') % (desc, succdesc))
467 467 repo.ui.status(msg)
468 468 # Make clearrebased aware state[rev] is not a true successor
469 469 self.skipped.add(rev)
470 470 # Record rev as moved to its desired destination in self.state.
471 471 # This helps bookmark and working parent movement.
472 472 dest = max(adjustdest(repo, rev, self.destmap, self.state,
473 473 self.skipped))
474 474 self.state[rev] = dest
475 475 elif self.state[rev] == revtodo:
476 476 pos += 1
477 477 ui.status(_('rebasing %s\n') % desc)
478 478 ui.progress(_("rebasing"), pos, ("%d:%s" % (rev, ctx)),
479 479 _('changesets'), total)
480 480 p1, p2, base = defineparents(repo, rev, self.destmap,
481 481 self.state, self.skipped,
482 482 self.obsoletenotrebased)
483 483 self.storestatus(tr=tr)
484 484 storecollapsemsg(repo, self.collapsemsg)
485 485 if len(repo[None].parents()) == 2:
486 486 repo.ui.debug('resuming interrupted rebase\n')
487 487 else:
488 488 try:
489 489 ui.setconfig('ui', 'forcemerge', opts.get('tool', ''),
490 490 'rebase')
491 491 stats = rebasenode(repo, rev, p1, base, self.state,
492 492 self.collapsef, dest, wctx=self.wctx)
493 493 if stats and stats[3] > 0:
494 494 if self.wctx.isinmemory():
495 495 raise error.InMemoryMergeConflictsError()
496 496 else:
497 497 raise error.InterventionRequired(
498 498 _('unresolved conflicts (see hg '
499 499 'resolve, then hg rebase --continue)'))
500 500 finally:
501 501 ui.setconfig('ui', 'forcemerge', '', 'rebase')
502 502 if not self.collapsef:
503 503 merging = p2 != nullrev
504 504 editform = cmdutil.mergeeditform(merging, 'rebase')
505 505 editor = cmdutil.getcommiteditor(editform=editform,
506 506 **pycompat.strkwargs(opts))
507 507 if self.wctx.isinmemory():
508 508 newnode = concludememorynode(repo, rev, p1, p2,
509 509 wctx=self.wctx,
510 510 extrafn=_makeextrafn(self.extrafns),
511 511 editor=editor,
512 512 keepbranches=self.keepbranchesf,
513 513 date=self.date)
514 514 mergemod.mergestate.clean(repo)
515 515 else:
516 516 newnode = concludenode(repo, rev, p1, p2,
517 517 extrafn=_makeextrafn(self.extrafns),
518 518 editor=editor,
519 519 keepbranches=self.keepbranchesf,
520 520 date=self.date)
521 521
522 522 if newnode is None:
523 523 # If it ended up being a no-op commit, then the normal
524 524 # merge state clean-up path doesn't happen, so do it
525 525 # here. Fix issue5494
526 526 mergemod.mergestate.clean(repo)
527 527 else:
528 528 # Skip commit if we are collapsing
529 529 if self.wctx.isinmemory():
530 530 self.wctx.setbase(repo[p1])
531 531 else:
532 532 repo.setparents(repo[p1].node())
533 533 newnode = None
534 534 # Update the state
535 535 if newnode is not None:
536 536 self.state[rev] = repo[newnode].rev()
537 537 ui.debug('rebased as %s\n' % short(newnode))
538 538 else:
539 539 if not self.collapsef:
540 540 ui.warn(_('note: rebase of %d:%s created no changes '
541 541 'to commit\n') % (rev, ctx))
542 542 self.skipped.add(rev)
543 543 self.state[rev] = p1
544 544 ui.debug('next revision set to %d\n' % p1)
545 545 else:
546 546 ui.status(_('already rebased %s as %s\n') %
547 547 (desc, repo[self.state[rev]]))
548 548 return pos
549 549
550 550 def _finishrebase(self):
551 551 repo, ui, opts = self.repo, self.ui, self.opts
552 552 fm = ui.formatter('rebase', opts)
553 553 fm.startitem()
554 554 if self.collapsef and not self.keepopen:
555 555 p1, p2, _base = defineparents(repo, min(self.state), self.destmap,
556 556 self.state, self.skipped,
557 557 self.obsoletenotrebased)
558 558 editopt = opts.get('edit')
559 559 editform = 'rebase.collapse'
560 560 if self.collapsemsg:
561 561 commitmsg = self.collapsemsg
562 562 else:
563 563 commitmsg = 'Collapsed revision'
564 564 for rebased in sorted(self.state):
565 565 if rebased not in self.skipped:
566 566 commitmsg += '\n* %s' % repo[rebased].description()
567 567 editopt = True
568 568 editor = cmdutil.getcommiteditor(edit=editopt, editform=editform)
569 569 revtoreuse = max(self.state)
570 570
571 571 dsguard = None
572 572 if self.inmemory:
573 573 newnode = concludememorynode(repo, revtoreuse, p1,
574 574 self.external,
575 575 commitmsg=commitmsg,
576 576 extrafn=_makeextrafn(self.extrafns),
577 577 editor=editor,
578 578 keepbranches=self.keepbranchesf,
579 579 date=self.date, wctx=self.wctx)
580 580 else:
581 581 if ui.configbool('rebase', 'singletransaction'):
582 582 dsguard = dirstateguard.dirstateguard(repo, 'rebase')
583 583 with util.acceptintervention(dsguard):
584 584 newnode = concludenode(repo, revtoreuse, p1, self.external,
585 585 commitmsg=commitmsg,
586 586 extrafn=_makeextrafn(self.extrafns),
587 587 editor=editor,
588 588 keepbranches=self.keepbranchesf,
589 589 date=self.date)
590 590 if newnode is not None:
591 591 newrev = repo[newnode].rev()
592 for oldrev in self.state.iterkeys():
592 for oldrev in self.state:
593 593 self.state[oldrev] = newrev
594 594
595 595 if 'qtip' in repo.tags():
596 596 updatemq(repo, self.state, self.skipped, **opts)
597 597
598 598 # restore original working directory
599 599 # (we do this before stripping)
600 600 newwd = self.state.get(self.originalwd, self.originalwd)
601 601 if newwd < 0:
602 602 # original directory is a parent of rebase set root or ignored
603 603 newwd = self.originalwd
604 604 if (newwd not in [c.rev() for c in repo[None].parents()] and
605 605 not self.inmemory):
606 606 ui.note(_("update back to initial working directory parent\n"))
607 607 hg.updaterepo(repo, newwd, False)
608 608
609 609 collapsedas = None
610 610 if not self.keepf:
611 611 if self.collapsef:
612 612 collapsedas = newnode
613 613 clearrebased(ui, repo, self.destmap, self.state, self.skipped,
614 614 collapsedas, self.keepf, fm=fm)
615 615
616 616 clearstatus(repo)
617 617 clearcollapsemsg(repo)
618 618
619 619 ui.note(_("rebase completed\n"))
620 620 util.unlinkpath(repo.sjoin('undo'), ignoremissing=True)
621 621 if self.skipped:
622 622 skippedlen = len(self.skipped)
623 623 ui.note(_("%d revisions have been skipped\n") % skippedlen)
624 624 fm.end()
625 625
626 626 if (self.activebookmark and self.activebookmark in repo._bookmarks and
627 627 repo['.'].node() == repo._bookmarks[self.activebookmark]):
628 628 bookmarks.activate(repo, self.activebookmark)
629 629
630 630 @command('rebase',
631 631 [('s', 'source', '',
632 632 _('rebase the specified changeset and descendants'), _('REV')),
633 633 ('b', 'base', '',
634 634 _('rebase everything from branching point of specified changeset'),
635 635 _('REV')),
636 636 ('r', 'rev', [],
637 637 _('rebase these revisions'),
638 638 _('REV')),
639 639 ('d', 'dest', '',
640 640 _('rebase onto the specified changeset'), _('REV')),
641 641 ('', 'collapse', False, _('collapse the rebased changesets')),
642 642 ('m', 'message', '',
643 643 _('use text as collapse commit message'), _('TEXT')),
644 644 ('e', 'edit', False, _('invoke editor on commit messages')),
645 645 ('l', 'logfile', '',
646 646 _('read collapse commit message from file'), _('FILE')),
647 647 ('k', 'keep', False, _('keep original changesets')),
648 648 ('', 'keepbranches', False, _('keep original branch names')),
649 649 ('D', 'detach', False, _('(DEPRECATED)')),
650 650 ('i', 'interactive', False, _('(DEPRECATED)')),
651 651 ('t', 'tool', '', _('specify merge tool')),
652 652 ('c', 'continue', False, _('continue an interrupted rebase')),
653 653 ('a', 'abort', False, _('abort an interrupted rebase'))] +
654 654 cmdutil.formatteropts,
655 655 _('[-s REV | -b REV] [-d REV] [OPTION]'))
656 656 def rebase(ui, repo, **opts):
657 657 """move changeset (and descendants) to a different branch
658 658
659 659 Rebase uses repeated merging to graft changesets from one part of
660 660 history (the source) onto another (the destination). This can be
661 661 useful for linearizing *local* changes relative to a master
662 662 development tree.
663 663
664 664 Published commits cannot be rebased (see :hg:`help phases`).
665 665 To copy commits, see :hg:`help graft`.
666 666
667 667 If you don't specify a destination changeset (``-d/--dest``), rebase
668 668 will use the same logic as :hg:`merge` to pick a destination. if
669 669 the current branch contains exactly one other head, the other head
670 670 is merged with by default. Otherwise, an explicit revision with
671 671 which to merge with must be provided. (destination changeset is not
672 672 modified by rebasing, but new changesets are added as its
673 673 descendants.)
674 674
675 675 Here are the ways to select changesets:
676 676
677 677 1. Explicitly select them using ``--rev``.
678 678
679 679 2. Use ``--source`` to select a root changeset and include all of its
680 680 descendants.
681 681
682 682 3. Use ``--base`` to select a changeset; rebase will find ancestors
683 683 and their descendants which are not also ancestors of the destination.
684 684
685 685 4. If you do not specify any of ``--rev``, ``source``, or ``--base``,
686 686 rebase will use ``--base .`` as above.
687 687
688 688 If ``--source`` or ``--rev`` is used, special names ``SRC`` and ``ALLSRC``
689 689 can be used in ``--dest``. Destination would be calculated per source
690 690 revision with ``SRC`` substituted by that single source revision and
691 691 ``ALLSRC`` substituted by all source revisions.
692 692
693 693 Rebase will destroy original changesets unless you use ``--keep``.
694 694 It will also move your bookmarks (even if you do).
695 695
696 696 Some changesets may be dropped if they do not contribute changes
697 697 (e.g. merges from the destination branch).
698 698
699 699 Unlike ``merge``, rebase will do nothing if you are at the branch tip of
700 700 a named branch with two heads. You will need to explicitly specify source
701 701 and/or destination.
702 702
703 703 If you need to use a tool to automate merge/conflict decisions, you
704 704 can specify one with ``--tool``, see :hg:`help merge-tools`.
705 705 As a caveat: the tool will not be used to mediate when a file was
706 706 deleted, there is no hook presently available for this.
707 707
708 708 If a rebase is interrupted to manually resolve a conflict, it can be
709 709 continued with --continue/-c or aborted with --abort/-a.
710 710
711 711 .. container:: verbose
712 712
713 713 Examples:
714 714
715 715 - move "local changes" (current commit back to branching point)
716 716 to the current branch tip after a pull::
717 717
718 718 hg rebase
719 719
720 720 - move a single changeset to the stable branch::
721 721
722 722 hg rebase -r 5f493448 -d stable
723 723
724 724 - splice a commit and all its descendants onto another part of history::
725 725
726 726 hg rebase --source c0c3 --dest 4cf9
727 727
728 728 - rebase everything on a branch marked by a bookmark onto the
729 729 default branch::
730 730
731 731 hg rebase --base myfeature --dest default
732 732
733 733 - collapse a sequence of changes into a single commit::
734 734
735 735 hg rebase --collapse -r 1520:1525 -d .
736 736
737 737 - move a named branch while preserving its name::
738 738
739 739 hg rebase -r "branch(featureX)" -d 1.3 --keepbranches
740 740
741 741 - stabilize orphaned changesets so history looks linear::
742 742
743 743 hg rebase -r 'orphan()-obsolete()'\
744 744 -d 'first(max((successors(max(roots(ALLSRC) & ::SRC)^)-obsolete())::) +\
745 745 max(::((roots(ALLSRC) & ::SRC)^)-obsolete()))'
746 746
747 747 Configuration Options:
748 748
749 749 You can make rebase require a destination if you set the following config
750 750 option::
751 751
752 752 [commands]
753 753 rebase.requiredest = True
754 754
755 755 By default, rebase will close the transaction after each commit. For
756 756 performance purposes, you can configure rebase to use a single transaction
757 757 across the entire rebase. WARNING: This setting introduces a significant
758 758 risk of losing the work you've done in a rebase if the rebase aborts
759 759 unexpectedly::
760 760
761 761 [rebase]
762 762 singletransaction = True
763 763
764 764 By default, rebase writes to the working copy, but you can configure it to
765 765 run in-memory for for better performance, and to allow it to run if the
766 766 working copy is dirty::
767 767
768 768 [rebase]
769 769 experimental.inmemory = True
770 770
771 771 Return Values:
772 772
773 773 Returns 0 on success, 1 if nothing to rebase or there are
774 774 unresolved conflicts.
775 775
776 776 """
777 777 inmemory = ui.configbool('rebase', 'experimental.inmemory')
778 778 if (opts.get('continue') or opts.get('abort') or
779 779 repo.currenttransaction() is not None):
780 780 # in-memory rebase is not compatible with resuming rebases.
781 781 # (Or if it is run within a transaction, since the restart logic can
782 782 # fail the entire transaction.)
783 783 inmemory = False
784 784
785 785 if inmemory:
786 786 try:
787 787 # in-memory merge doesn't support conflicts, so if we hit any, abort
788 788 # and re-run as an on-disk merge.
789 789 return _origrebase(ui, repo, inmemory=inmemory, **opts)
790 790 except error.InMemoryMergeConflictsError:
791 791 ui.warn(_('hit merge conflicts; re-running rebase without in-memory'
792 792 ' merge\n'))
793 793 _origrebase(ui, repo, **{'abort': True})
794 794 return _origrebase(ui, repo, inmemory=False, **opts)
795 795 else:
796 796 return _origrebase(ui, repo, **opts)
797 797
798 798 def _origrebase(ui, repo, inmemory=False, **opts):
799 799 opts = pycompat.byteskwargs(opts)
800 800 rbsrt = rebaseruntime(repo, ui, inmemory, opts)
801 801
802 802 with repo.wlock(), repo.lock():
803 803 # Validate input and define rebasing points
804 804 destf = opts.get('dest', None)
805 805 srcf = opts.get('source', None)
806 806 basef = opts.get('base', None)
807 807 revf = opts.get('rev', [])
808 808 # search default destination in this space
809 809 # used in the 'hg pull --rebase' case, see issue 5214.
810 810 destspace = opts.get('_destspace')
811 811 contf = opts.get('continue')
812 812 abortf = opts.get('abort')
813 813 if opts.get('interactive'):
814 814 try:
815 815 if extensions.find('histedit'):
816 816 enablehistedit = ''
817 817 except KeyError:
818 818 enablehistedit = " --config extensions.histedit="
819 819 help = "hg%s help -e histedit" % enablehistedit
820 820 msg = _("interactive history editing is supported by the "
821 821 "'histedit' extension (see \"%s\")") % help
822 822 raise error.Abort(msg)
823 823
824 824 if rbsrt.collapsemsg and not rbsrt.collapsef:
825 825 raise error.Abort(
826 826 _('message can only be specified with collapse'))
827 827
828 828 if contf or abortf:
829 829 if contf and abortf:
830 830 raise error.Abort(_('cannot use both abort and continue'))
831 831 if rbsrt.collapsef:
832 832 raise error.Abort(
833 833 _('cannot use collapse with continue or abort'))
834 834 if srcf or basef or destf:
835 835 raise error.Abort(
836 836 _('abort and continue do not allow specifying revisions'))
837 837 if abortf and opts.get('tool', False):
838 838 ui.warn(_('tool option will be ignored\n'))
839 839 if contf:
840 840 ms = mergemod.mergestate.read(repo)
841 841 mergeutil.checkunresolved(ms)
842 842
843 843 retcode = rbsrt._prepareabortorcontinue(abortf)
844 844 if retcode is not None:
845 845 return retcode
846 846 else:
847 847 destmap = _definedestmap(ui, repo, rbsrt, destf, srcf, basef, revf,
848 848 destspace=destspace)
849 849 retcode = rbsrt._preparenewrebase(destmap)
850 850 if retcode is not None:
851 851 return retcode
852 852
853 853 tr = None
854 854 dsguard = None
855 855
856 856 singletr = ui.configbool('rebase', 'singletransaction')
857 857 if singletr:
858 858 tr = repo.transaction('rebase')
859 859
860 860 # If `rebase.singletransaction` is enabled, wrap the entire operation in
861 861 # one transaction here. Otherwise, transactions are obtained when
862 862 # committing each node, which is slower but allows partial success.
863 863 with util.acceptintervention(tr):
864 864 # Same logic for the dirstate guard, except we don't create one when
865 865 # rebasing in-memory (it's not needed).
866 866 if singletr and not inmemory:
867 867 dsguard = dirstateguard.dirstateguard(repo, 'rebase')
868 868 with util.acceptintervention(dsguard):
869 869 rbsrt._performrebase(tr)
870 870
871 871 rbsrt._finishrebase()
872 872
873 873 def _definedestmap(ui, repo, rbsrt, destf=None, srcf=None, basef=None,
874 874 revf=None, destspace=None):
875 875 """use revisions argument to define destmap {srcrev: destrev}"""
876 876 if revf is None:
877 877 revf = []
878 878
879 879 # destspace is here to work around issues with `hg pull --rebase` see
880 880 # issue5214 for details
881 881 if srcf and basef:
882 882 raise error.Abort(_('cannot specify both a source and a base'))
883 883 if revf and basef:
884 884 raise error.Abort(_('cannot specify both a revision and a base'))
885 885 if revf and srcf:
886 886 raise error.Abort(_('cannot specify both a revision and a source'))
887 887
888 888 if not rbsrt.inmemory:
889 889 cmdutil.checkunfinished(repo)
890 890 cmdutil.bailifchanged(repo)
891 891
892 892 if ui.configbool('commands', 'rebase.requiredest') and not destf:
893 893 raise error.Abort(_('you must specify a destination'),
894 894 hint=_('use: hg rebase -d REV'))
895 895
896 896 dest = None
897 897
898 898 if revf:
899 899 rebaseset = scmutil.revrange(repo, revf)
900 900 if not rebaseset:
901 901 ui.status(_('empty "rev" revision set - nothing to rebase\n'))
902 902 return None
903 903 elif srcf:
904 904 src = scmutil.revrange(repo, [srcf])
905 905 if not src:
906 906 ui.status(_('empty "source" revision set - nothing to rebase\n'))
907 907 return None
908 908 rebaseset = repo.revs('(%ld)::', src)
909 909 assert rebaseset
910 910 else:
911 911 base = scmutil.revrange(repo, [basef or '.'])
912 912 if not base:
913 913 ui.status(_('empty "base" revision set - '
914 914 "can't compute rebase set\n"))
915 915 return None
916 916 if destf:
917 917 # --base does not support multiple destinations
918 918 dest = scmutil.revsingle(repo, destf)
919 919 else:
920 920 dest = repo[_destrebase(repo, base, destspace=destspace)]
921 921 destf = str(dest)
922 922
923 923 roots = [] # selected children of branching points
924 924 bpbase = {} # {branchingpoint: [origbase]}
925 925 for b in base: # group bases by branching points
926 926 bp = repo.revs('ancestor(%d, %d)', b, dest).first()
927 927 bpbase[bp] = bpbase.get(bp, []) + [b]
928 928 if None in bpbase:
929 929 # emulate the old behavior, showing "nothing to rebase" (a better
930 930 # behavior may be abort with "cannot find branching point" error)
931 931 bpbase.clear()
932 932 for bp, bs in bpbase.iteritems(): # calculate roots
933 933 roots += list(repo.revs('children(%d) & ancestors(%ld)', bp, bs))
934 934
935 935 rebaseset = repo.revs('%ld::', roots)
936 936
937 937 if not rebaseset:
938 938 # transform to list because smartsets are not comparable to
939 939 # lists. This should be improved to honor laziness of
940 940 # smartset.
941 941 if list(base) == [dest.rev()]:
942 942 if basef:
943 943 ui.status(_('nothing to rebase - %s is both "base"'
944 944 ' and destination\n') % dest)
945 945 else:
946 946 ui.status(_('nothing to rebase - working directory '
947 947 'parent is also destination\n'))
948 948 elif not repo.revs('%ld - ::%d', base, dest):
949 949 if basef:
950 950 ui.status(_('nothing to rebase - "base" %s is '
951 951 'already an ancestor of destination '
952 952 '%s\n') %
953 953 ('+'.join(str(repo[r]) for r in base),
954 954 dest))
955 955 else:
956 956 ui.status(_('nothing to rebase - working '
957 957 'directory parent is already an '
958 958 'ancestor of destination %s\n') % dest)
959 959 else: # can it happen?
960 960 ui.status(_('nothing to rebase from %s to %s\n') %
961 961 ('+'.join(str(repo[r]) for r in base), dest))
962 962 return None
963 963 # If rebasing the working copy parent, force in-memory merge to be off.
964 964 #
965 965 # This is because the extra work of checking out the newly rebased commit
966 966 # outweights the benefits of rebasing in-memory, and executing an extra
967 967 # update command adds a bit of overhead, so better to just do it on disk. In
968 968 # all other cases leave it on.
969 969 #
970 970 # Note that there are cases where this isn't true -- e.g., rebasing large
971 971 # stacks that include the WCP. However, I'm not yet sure where the cutoff
972 972 # is.
973 973 rebasingwcp = repo['.'].rev() in rebaseset
974 974 ui.log("rebase", "", rebase_rebasing_wcp=rebasingwcp)
975 975 if rbsrt.inmemory and rebasingwcp:
976 976 rbsrt.inmemory = False
977 977 # Check these since we did not before.
978 978 cmdutil.checkunfinished(repo)
979 979 cmdutil.bailifchanged(repo)
980 980
981 981 if not destf:
982 982 dest = repo[_destrebase(repo, rebaseset, destspace=destspace)]
983 983 destf = str(dest)
984 984
985 985 allsrc = revsetlang.formatspec('%ld', rebaseset)
986 986 alias = {'ALLSRC': allsrc}
987 987
988 988 if dest is None:
989 989 try:
990 990 # fast path: try to resolve dest without SRC alias
991 991 dest = scmutil.revsingle(repo, destf, localalias=alias)
992 992 except error.RepoLookupError:
993 993 # multi-dest path: resolve dest for each SRC separately
994 994 destmap = {}
995 995 for r in rebaseset:
996 996 alias['SRC'] = revsetlang.formatspec('%d', r)
997 997 # use repo.anyrevs instead of scmutil.revsingle because we
998 998 # don't want to abort if destset is empty.
999 999 destset = repo.anyrevs([destf], user=True, localalias=alias)
1000 1000 size = len(destset)
1001 1001 if size == 1:
1002 1002 destmap[r] = destset.first()
1003 1003 elif size == 0:
1004 1004 ui.note(_('skipping %s - empty destination\n') % repo[r])
1005 1005 else:
1006 1006 raise error.Abort(_('rebase destination for %s is not '
1007 1007 'unique') % repo[r])
1008 1008
1009 1009 if dest is not None:
1010 1010 # single-dest case: assign dest to each rev in rebaseset
1011 1011 destrev = dest.rev()
1012 1012 destmap = {r: destrev for r in rebaseset} # {srcrev: destrev}
1013 1013
1014 1014 if not destmap:
1015 1015 ui.status(_('nothing to rebase - empty destination\n'))
1016 1016 return None
1017 1017
1018 1018 return destmap
1019 1019
1020 1020 def externalparent(repo, state, destancestors):
1021 1021 """Return the revision that should be used as the second parent
1022 1022 when the revisions in state is collapsed on top of destancestors.
1023 1023 Abort if there is more than one parent.
1024 1024 """
1025 1025 parents = set()
1026 1026 source = min(state)
1027 1027 for rev in state:
1028 1028 if rev == source:
1029 1029 continue
1030 1030 for p in repo[rev].parents():
1031 1031 if (p.rev() not in state
1032 1032 and p.rev() not in destancestors):
1033 1033 parents.add(p.rev())
1034 1034 if not parents:
1035 1035 return nullrev
1036 1036 if len(parents) == 1:
1037 1037 return parents.pop()
1038 1038 raise error.Abort(_('unable to collapse on top of %s, there is more '
1039 1039 'than one external parent: %s') %
1040 1040 (max(destancestors),
1041 1041 ', '.join(str(p) for p in sorted(parents))))
1042 1042
1043 1043 def concludememorynode(repo, rev, p1, p2, wctx=None,
1044 1044 commitmsg=None, editor=None, extrafn=None,
1045 1045 keepbranches=False, date=None):
1046 1046 '''Commit the memory changes with parents p1 and p2. Reuse commit info from
1047 1047 rev but also store useful information in extra.
1048 1048 Return node of committed revision.'''
1049 1049 ctx = repo[rev]
1050 1050 if commitmsg is None:
1051 1051 commitmsg = ctx.description()
1052 1052 keepbranch = keepbranches and repo[p1].branch() != ctx.branch()
1053 1053 extra = {'rebase_source': ctx.hex()}
1054 1054 if extrafn:
1055 1055 extrafn(ctx, extra)
1056 1056
1057 1057 destphase = max(ctx.phase(), phases.draft)
1058 1058 overrides = {('phases', 'new-commit'): destphase}
1059 1059 with repo.ui.configoverride(overrides, 'rebase'):
1060 1060 if keepbranch:
1061 1061 repo.ui.setconfig('ui', 'allowemptycommit', True)
1062 1062 # Replicates the empty check in ``repo.commit``.
1063 1063 if wctx.isempty() and not repo.ui.configbool('ui', 'allowemptycommit'):
1064 1064 return None
1065 1065
1066 1066 if date is None:
1067 1067 date = ctx.date()
1068 1068
1069 1069 # By convention, ``extra['branch']`` (set by extrafn) clobbers
1070 1070 # ``branch`` (used when passing ``--keepbranches``).
1071 1071 branch = repo[p1].branch()
1072 1072 if 'branch' in extra:
1073 1073 branch = extra['branch']
1074 1074
1075 1075 memctx = wctx.tomemctx(commitmsg, parents=(p1, p2), date=date,
1076 1076 extra=extra, user=ctx.user(), branch=branch, editor=editor)
1077 1077 commitres = repo.commitctx(memctx)
1078 1078 wctx.clean() # Might be reused
1079 1079 return commitres
1080 1080
1081 1081 def concludenode(repo, rev, p1, p2, commitmsg=None, editor=None, extrafn=None,
1082 1082 keepbranches=False, date=None):
1083 1083 '''Commit the wd changes with parents p1 and p2. Reuse commit info from rev
1084 1084 but also store useful information in extra.
1085 1085 Return node of committed revision.'''
1086 1086 dsguard = util.nullcontextmanager()
1087 1087 if not repo.ui.configbool('rebase', 'singletransaction'):
1088 1088 dsguard = dirstateguard.dirstateguard(repo, 'rebase')
1089 1089 with dsguard:
1090 1090 repo.setparents(repo[p1].node(), repo[p2].node())
1091 1091 ctx = repo[rev]
1092 1092 if commitmsg is None:
1093 1093 commitmsg = ctx.description()
1094 1094 keepbranch = keepbranches and repo[p1].branch() != ctx.branch()
1095 1095 extra = {'rebase_source': ctx.hex()}
1096 1096 if extrafn:
1097 1097 extrafn(ctx, extra)
1098 1098
1099 1099 destphase = max(ctx.phase(), phases.draft)
1100 1100 overrides = {('phases', 'new-commit'): destphase}
1101 1101 with repo.ui.configoverride(overrides, 'rebase'):
1102 1102 if keepbranch:
1103 1103 repo.ui.setconfig('ui', 'allowemptycommit', True)
1104 1104 # Commit might fail if unresolved files exist
1105 1105 if date is None:
1106 1106 date = ctx.date()
1107 1107 newnode = repo.commit(text=commitmsg, user=ctx.user(),
1108 1108 date=date, extra=extra, editor=editor)
1109 1109
1110 1110 repo.dirstate.setbranch(repo[newnode].branch())
1111 1111 return newnode
1112 1112
1113 1113 def rebasenode(repo, rev, p1, base, state, collapse, dest, wctx):
1114 1114 'Rebase a single revision rev on top of p1 using base as merge ancestor'
1115 1115 # Merge phase
1116 1116 # Update to destination and merge it with local
1117 1117 if wctx.isinmemory():
1118 1118 wctx.setbase(repo[p1])
1119 1119 else:
1120 1120 if repo['.'].rev() != p1:
1121 1121 repo.ui.debug(" update to %d:%s\n" % (p1, repo[p1]))
1122 1122 mergemod.update(repo, p1, False, True)
1123 1123 else:
1124 1124 repo.ui.debug(" already in destination\n")
1125 1125 # This is, alas, necessary to invalidate workingctx's manifest cache,
1126 1126 # as well as other data we litter on it in other places.
1127 1127 wctx = repo[None]
1128 1128 repo.dirstate.write(repo.currenttransaction())
1129 1129 repo.ui.debug(" merge against %d:%s\n" % (rev, repo[rev]))
1130 1130 if base is not None:
1131 1131 repo.ui.debug(" detach base %d:%s\n" % (base, repo[base]))
1132 1132 # When collapsing in-place, the parent is the common ancestor, we
1133 1133 # have to allow merging with it.
1134 1134 stats = mergemod.update(repo, rev, True, True, base, collapse,
1135 1135 labels=['dest', 'source'], wc=wctx)
1136 1136 if collapse:
1137 1137 copies.duplicatecopies(repo, wctx, rev, dest)
1138 1138 else:
1139 1139 # If we're not using --collapse, we need to
1140 1140 # duplicate copies between the revision we're
1141 1141 # rebasing and its first parent, but *not*
1142 1142 # duplicate any copies that have already been
1143 1143 # performed in the destination.
1144 1144 p1rev = repo[rev].p1().rev()
1145 1145 copies.duplicatecopies(repo, wctx, rev, p1rev, skiprev=dest)
1146 1146 return stats
1147 1147
1148 1148 def adjustdest(repo, rev, destmap, state, skipped):
1149 1149 """adjust rebase destination given the current rebase state
1150 1150
1151 1151 rev is what is being rebased. Return a list of two revs, which are the
1152 1152 adjusted destinations for rev's p1 and p2, respectively. If a parent is
1153 1153 nullrev, return dest without adjustment for it.
1154 1154
1155 1155 For example, when doing rebasing B+E to F, C to G, rebase will first move B
1156 1156 to B1, and E's destination will be adjusted from F to B1.
1157 1157
1158 1158 B1 <- written during rebasing B
1159 1159 |
1160 1160 F <- original destination of B, E
1161 1161 |
1162 1162 | E <- rev, which is being rebased
1163 1163 | |
1164 1164 | D <- prev, one parent of rev being checked
1165 1165 | |
1166 1166 | x <- skipped, ex. no successor or successor in (::dest)
1167 1167 | |
1168 1168 | C <- rebased as C', different destination
1169 1169 | |
1170 1170 | B <- rebased as B1 C'
1171 1171 |/ |
1172 1172 A G <- destination of C, different
1173 1173
1174 1174 Another example about merge changeset, rebase -r C+G+H -d K, rebase will
1175 1175 first move C to C1, G to G1, and when it's checking H, the adjusted
1176 1176 destinations will be [C1, G1].
1177 1177
1178 1178 H C1 G1
1179 1179 /| | /
1180 1180 F G |/
1181 1181 K | | -> K
1182 1182 | C D |
1183 1183 | |/ |
1184 1184 | B | ...
1185 1185 |/ |/
1186 1186 A A
1187 1187
1188 1188 Besides, adjust dest according to existing rebase information. For example,
1189 1189
1190 1190 B C D B needs to be rebased on top of C, C needs to be rebased on top
1191 1191 \|/ of D. We will rebase C first.
1192 1192 A
1193 1193
1194 1194 C' After rebasing C, when considering B's destination, use C'
1195 1195 | instead of the original C.
1196 1196 B D
1197 1197 \ /
1198 1198 A
1199 1199 """
1200 1200 # pick already rebased revs with same dest from state as interesting source
1201 1201 dest = destmap[rev]
1202 1202 source = [s for s, d in state.items()
1203 1203 if d > 0 and destmap[s] == dest and s not in skipped]
1204 1204
1205 1205 result = []
1206 1206 for prev in repo.changelog.parentrevs(rev):
1207 1207 adjusted = dest
1208 1208 if prev != nullrev:
1209 1209 candidate = repo.revs('max(%ld and (::%d))', source, prev).first()
1210 1210 if candidate is not None:
1211 1211 adjusted = state[candidate]
1212 1212 if adjusted == dest and dest in state:
1213 1213 adjusted = state[dest]
1214 1214 if adjusted == revtodo:
1215 1215 # sortsource should produce an order that makes this impossible
1216 1216 raise error.ProgrammingError(
1217 1217 'rev %d should be rebased already at this time' % dest)
1218 1218 result.append(adjusted)
1219 1219 return result
1220 1220
1221 1221 def _checkobsrebase(repo, ui, rebaseobsrevs, rebaseobsskipped):
1222 1222 """
1223 1223 Abort if rebase will create divergence or rebase is noop because of markers
1224 1224
1225 1225 `rebaseobsrevs`: set of obsolete revision in source
1226 1226 `rebaseobsskipped`: set of revisions from source skipped because they have
1227 1227 successors in destination or no non-obsolete successor.
1228 1228 """
1229 1229 # Obsolete node with successors not in dest leads to divergence
1230 1230 divergenceok = ui.configbool('experimental',
1231 1231 'evolution.allowdivergence')
1232 1232 divergencebasecandidates = rebaseobsrevs - rebaseobsskipped
1233 1233
1234 1234 if divergencebasecandidates and not divergenceok:
1235 1235 divhashes = (str(repo[r])
1236 1236 for r in divergencebasecandidates)
1237 1237 msg = _("this rebase will cause "
1238 1238 "divergences from: %s")
1239 1239 h = _("to force the rebase please set "
1240 1240 "experimental.evolution.allowdivergence=True")
1241 1241 raise error.Abort(msg % (",".join(divhashes),), hint=h)
1242 1242
1243 1243 def successorrevs(unfi, rev):
1244 1244 """yield revision numbers for successors of rev"""
1245 1245 assert unfi.filtername is None
1246 1246 nodemap = unfi.changelog.nodemap
1247 1247 for s in obsutil.allsuccessors(unfi.obsstore, [unfi[rev].node()]):
1248 1248 if s in nodemap:
1249 1249 yield nodemap[s]
1250 1250
1251 1251 def defineparents(repo, rev, destmap, state, skipped, obsskipped):
1252 1252 """Return new parents and optionally a merge base for rev being rebased
1253 1253
1254 1254 The destination specified by "dest" cannot always be used directly because
1255 1255 previously rebase result could affect destination. For example,
1256 1256
1257 1257 D E rebase -r C+D+E -d B
1258 1258 |/ C will be rebased to C'
1259 1259 B C D's new destination will be C' instead of B
1260 1260 |/ E's new destination will be C' instead of B
1261 1261 A
1262 1262
1263 1263 The new parents of a merge is slightly more complicated. See the comment
1264 1264 block below.
1265 1265 """
1266 1266 # use unfiltered changelog since successorrevs may return filtered nodes
1267 1267 assert repo.filtername is None
1268 1268 cl = repo.changelog
1269 1269 def isancestor(a, b):
1270 1270 # take revision numbers instead of nodes
1271 1271 if a == b:
1272 1272 return True
1273 1273 elif a > b:
1274 1274 return False
1275 1275 return cl.isancestor(cl.node(a), cl.node(b))
1276 1276
1277 1277 dest = destmap[rev]
1278 1278 oldps = repo.changelog.parentrevs(rev) # old parents
1279 1279 newps = [nullrev, nullrev] # new parents
1280 1280 dests = adjustdest(repo, rev, destmap, state, skipped)
1281 1281 bases = list(oldps) # merge base candidates, initially just old parents
1282 1282
1283 1283 if all(r == nullrev for r in oldps[1:]):
1284 1284 # For non-merge changeset, just move p to adjusted dest as requested.
1285 1285 newps[0] = dests[0]
1286 1286 else:
1287 1287 # For merge changeset, if we move p to dests[i] unconditionally, both
1288 1288 # parents may change and the end result looks like "the merge loses a
1289 1289 # parent", which is a surprise. This is a limit because "--dest" only
1290 1290 # accepts one dest per src.
1291 1291 #
1292 1292 # Therefore, only move p with reasonable conditions (in this order):
1293 1293 # 1. use dest, if dest is a descendent of (p or one of p's successors)
1294 1294 # 2. use p's rebased result, if p is rebased (state[p] > 0)
1295 1295 #
1296 1296 # Comparing with adjustdest, the logic here does some additional work:
1297 1297 # 1. decide which parents will not be moved towards dest
1298 1298 # 2. if the above decision is "no", should a parent still be moved
1299 1299 # because it was rebased?
1300 1300 #
1301 1301 # For example:
1302 1302 #
1303 1303 # C # "rebase -r C -d D" is an error since none of the parents
1304 1304 # /| # can be moved. "rebase -r B+C -d D" will move C's parent
1305 1305 # A B D # B (using rule "2."), since B will be rebased.
1306 1306 #
1307 1307 # The loop tries to be not rely on the fact that a Mercurial node has
1308 1308 # at most 2 parents.
1309 1309 for i, p in enumerate(oldps):
1310 1310 np = p # new parent
1311 1311 if any(isancestor(x, dests[i]) for x in successorrevs(repo, p)):
1312 1312 np = dests[i]
1313 1313 elif p in state and state[p] > 0:
1314 1314 np = state[p]
1315 1315
1316 1316 # "bases" only record "special" merge bases that cannot be
1317 1317 # calculated from changelog DAG (i.e. isancestor(p, np) is False).
1318 1318 # For example:
1319 1319 #
1320 1320 # B' # rebase -s B -d D, when B was rebased to B'. dest for C
1321 1321 # | C # is B', but merge base for C is B, instead of
1322 1322 # D | # changelog.ancestor(C, B') == A. If changelog DAG and
1323 1323 # | B # "state" edges are merged (so there will be an edge from
1324 1324 # |/ # B to B'), the merge base is still ancestor(C, B') in
1325 1325 # A # the merged graph.
1326 1326 #
1327 1327 # Also see https://bz.mercurial-scm.org/show_bug.cgi?id=1950#c8
1328 1328 # which uses "virtual null merge" to explain this situation.
1329 1329 if isancestor(p, np):
1330 1330 bases[i] = nullrev
1331 1331
1332 1332 # If one parent becomes an ancestor of the other, drop the ancestor
1333 1333 for j, x in enumerate(newps[:i]):
1334 1334 if x == nullrev:
1335 1335 continue
1336 1336 if isancestor(np, x): # CASE-1
1337 1337 np = nullrev
1338 1338 elif isancestor(x, np): # CASE-2
1339 1339 newps[j] = np
1340 1340 np = nullrev
1341 1341 # New parents forming an ancestor relationship does not
1342 1342 # mean the old parents have a similar relationship. Do not
1343 1343 # set bases[x] to nullrev.
1344 1344 bases[j], bases[i] = bases[i], bases[j]
1345 1345
1346 1346 newps[i] = np
1347 1347
1348 1348 # "rebasenode" updates to new p1, and the old p1 will be used as merge
1349 1349 # base. If only p2 changes, merging using unchanged p1 as merge base is
1350 1350 # suboptimal. Therefore swap parents to make the merge sane.
1351 1351 if newps[1] != nullrev and oldps[0] == newps[0]:
1352 1352 assert len(newps) == 2 and len(oldps) == 2
1353 1353 newps.reverse()
1354 1354 bases.reverse()
1355 1355
1356 1356 # No parent change might be an error because we fail to make rev a
1357 1357 # descendent of requested dest. This can happen, for example:
1358 1358 #
1359 1359 # C # rebase -r C -d D
1360 1360 # /| # None of A and B will be changed to D and rebase fails.
1361 1361 # A B D
1362 1362 if set(newps) == set(oldps) and dest not in newps:
1363 1363 raise error.Abort(_('cannot rebase %d:%s without '
1364 1364 'moving at least one of its parents')
1365 1365 % (rev, repo[rev]))
1366 1366
1367 1367 # Source should not be ancestor of dest. The check here guarantees it's
1368 1368 # impossible. With multi-dest, the initial check does not cover complex
1369 1369 # cases since we don't have abstractions to dry-run rebase cheaply.
1370 1370 if any(p != nullrev and isancestor(rev, p) for p in newps):
1371 1371 raise error.Abort(_('source is ancestor of destination'))
1372 1372
1373 1373 # "rebasenode" updates to new p1, use the corresponding merge base.
1374 1374 if bases[0] != nullrev:
1375 1375 base = bases[0]
1376 1376 else:
1377 1377 base = None
1378 1378
1379 1379 # Check if the merge will contain unwanted changes. That may happen if
1380 1380 # there are multiple special (non-changelog ancestor) merge bases, which
1381 1381 # cannot be handled well by the 3-way merge algorithm. For example:
1382 1382 #
1383 1383 # F
1384 1384 # /|
1385 1385 # D E # "rebase -r D+E+F -d Z", when rebasing F, if "D" was chosen
1386 1386 # | | # as merge base, the difference between D and F will include
1387 1387 # B C # C, so the rebased F will contain C surprisingly. If "E" was
1388 1388 # |/ # chosen, the rebased F will contain B.
1389 1389 # A Z
1390 1390 #
1391 1391 # But our merge base candidates (D and E in above case) could still be
1392 1392 # better than the default (ancestor(F, Z) == null). Therefore still
1393 1393 # pick one (so choose p1 above).
1394 1394 if sum(1 for b in bases if b != nullrev) > 1:
1395 1395 unwanted = [None, None] # unwanted[i]: unwanted revs if choose bases[i]
1396 1396 for i, base in enumerate(bases):
1397 1397 if base == nullrev:
1398 1398 continue
1399 1399 # Revisions in the side (not chosen as merge base) branch that
1400 1400 # might contain "surprising" contents
1401 1401 siderevs = list(repo.revs('((%ld-%d) %% (%d+%d))',
1402 1402 bases, base, base, dest))
1403 1403
1404 1404 # If those revisions are covered by rebaseset, the result is good.
1405 1405 # A merge in rebaseset would be considered to cover its ancestors.
1406 1406 if siderevs:
1407 1407 rebaseset = [r for r, d in state.items()
1408 1408 if d > 0 and r not in obsskipped]
1409 1409 merges = [r for r in rebaseset
1410 1410 if cl.parentrevs(r)[1] != nullrev]
1411 1411 unwanted[i] = list(repo.revs('%ld - (::%ld) - %ld',
1412 1412 siderevs, merges, rebaseset))
1413 1413
1414 1414 # Choose a merge base that has a minimal number of unwanted revs.
1415 1415 l, i = min((len(revs), i)
1416 1416 for i, revs in enumerate(unwanted) if revs is not None)
1417 1417 base = bases[i]
1418 1418
1419 1419 # newps[0] should match merge base if possible. Currently, if newps[i]
1420 1420 # is nullrev, the only case is newps[i] and newps[j] (j < i), one is
1421 1421 # the other's ancestor. In that case, it's fine to not swap newps here.
1422 1422 # (see CASE-1 and CASE-2 above)
1423 1423 if i != 0 and newps[i] != nullrev:
1424 1424 newps[0], newps[i] = newps[i], newps[0]
1425 1425
1426 1426 # The merge will include unwanted revisions. Abort now. Revisit this if
1427 1427 # we have a more advanced merge algorithm that handles multiple bases.
1428 1428 if l > 0:
1429 1429 unwanteddesc = _(' or ').join(
1430 1430 (', '.join('%d:%s' % (r, repo[r]) for r in revs)
1431 1431 for revs in unwanted if revs is not None))
1432 1432 raise error.Abort(
1433 1433 _('rebasing %d:%s will include unwanted changes from %s')
1434 1434 % (rev, repo[rev], unwanteddesc))
1435 1435
1436 1436 repo.ui.debug(" future parents are %d and %d\n" % tuple(newps))
1437 1437
1438 1438 return newps[0], newps[1], base
1439 1439
1440 1440 def isagitpatch(repo, patchname):
1441 1441 'Return true if the given patch is in git format'
1442 1442 mqpatch = os.path.join(repo.mq.path, patchname)
1443 1443 for line in patch.linereader(file(mqpatch, 'rb')):
1444 1444 if line.startswith('diff --git'):
1445 1445 return True
1446 1446 return False
1447 1447
1448 1448 def updatemq(repo, state, skipped, **opts):
1449 1449 'Update rebased mq patches - finalize and then import them'
1450 1450 mqrebase = {}
1451 1451 mq = repo.mq
1452 1452 original_series = mq.fullseries[:]
1453 1453 skippedpatches = set()
1454 1454
1455 1455 for p in mq.applied:
1456 1456 rev = repo[p.node].rev()
1457 1457 if rev in state:
1458 1458 repo.ui.debug('revision %d is an mq patch (%s), finalize it.\n' %
1459 1459 (rev, p.name))
1460 1460 mqrebase[rev] = (p.name, isagitpatch(repo, p.name))
1461 1461 else:
1462 1462 # Applied but not rebased, not sure this should happen
1463 1463 skippedpatches.add(p.name)
1464 1464
1465 1465 if mqrebase:
1466 1466 mq.finish(repo, mqrebase.keys())
1467 1467
1468 1468 # We must start import from the newest revision
1469 1469 for rev in sorted(mqrebase, reverse=True):
1470 1470 if rev not in skipped:
1471 1471 name, isgit = mqrebase[rev]
1472 1472 repo.ui.note(_('updating mq patch %s to %s:%s\n') %
1473 1473 (name, state[rev], repo[state[rev]]))
1474 1474 mq.qimport(repo, (), patchname=name, git=isgit,
1475 1475 rev=[str(state[rev])])
1476 1476 else:
1477 1477 # Rebased and skipped
1478 1478 skippedpatches.add(mqrebase[rev][0])
1479 1479
1480 1480 # Patches were either applied and rebased and imported in
1481 1481 # order, applied and removed or unapplied. Discard the removed
1482 1482 # ones while preserving the original series order and guards.
1483 1483 newseries = [s for s in original_series
1484 1484 if mq.guard_re.split(s, 1)[0] not in skippedpatches]
1485 1485 mq.fullseries[:] = newseries
1486 1486 mq.seriesdirty = True
1487 1487 mq.savedirty()
1488 1488
1489 1489 def storecollapsemsg(repo, collapsemsg):
1490 1490 'Store the collapse message to allow recovery'
1491 1491 collapsemsg = collapsemsg or ''
1492 1492 f = repo.vfs("last-message.txt", "w")
1493 1493 f.write("%s\n" % collapsemsg)
1494 1494 f.close()
1495 1495
1496 1496 def clearcollapsemsg(repo):
1497 1497 'Remove collapse message file'
1498 1498 repo.vfs.unlinkpath("last-message.txt", ignoremissing=True)
1499 1499
1500 1500 def restorecollapsemsg(repo, isabort):
1501 1501 'Restore previously stored collapse message'
1502 1502 try:
1503 1503 f = repo.vfs("last-message.txt")
1504 1504 collapsemsg = f.readline().strip()
1505 1505 f.close()
1506 1506 except IOError as err:
1507 1507 if err.errno != errno.ENOENT:
1508 1508 raise
1509 1509 if isabort:
1510 1510 # Oh well, just abort like normal
1511 1511 collapsemsg = ''
1512 1512 else:
1513 1513 raise error.Abort(_('missing .hg/last-message.txt for rebase'))
1514 1514 return collapsemsg
1515 1515
1516 1516 def clearstatus(repo):
1517 1517 'Remove the status files'
1518 1518 # Make sure the active transaction won't write the state file
1519 1519 tr = repo.currenttransaction()
1520 1520 if tr:
1521 1521 tr.removefilegenerator('rebasestate')
1522 1522 repo.vfs.unlinkpath("rebasestate", ignoremissing=True)
1523 1523
1524 1524 def needupdate(repo, state):
1525 1525 '''check whether we should `update --clean` away from a merge, or if
1526 1526 somehow the working dir got forcibly updated, e.g. by older hg'''
1527 1527 parents = [p.rev() for p in repo[None].parents()]
1528 1528
1529 1529 # Are we in a merge state at all?
1530 1530 if len(parents) < 2:
1531 1531 return False
1532 1532
1533 1533 # We should be standing on the first as-of-yet unrebased commit.
1534 1534 firstunrebased = min([old for old, new in state.iteritems()
1535 1535 if new == nullrev])
1536 1536 if firstunrebased in parents:
1537 1537 return True
1538 1538
1539 1539 return False
1540 1540
1541 1541 def abort(repo, originalwd, destmap, state, activebookmark=None):
1542 1542 '''Restore the repository to its original state. Additional args:
1543 1543
1544 1544 activebookmark: the name of the bookmark that should be active after the
1545 1545 restore'''
1546 1546
1547 1547 try:
1548 1548 # If the first commits in the rebased set get skipped during the rebase,
1549 1549 # their values within the state mapping will be the dest rev id. The
1550 1550 # dstates list must must not contain the dest rev (issue4896)
1551 1551 dstates = [s for r, s in state.items() if s >= 0 and s != destmap[r]]
1552 1552 immutable = [d for d in dstates if not repo[d].mutable()]
1553 1553 cleanup = True
1554 1554 if immutable:
1555 1555 repo.ui.warn(_("warning: can't clean up public changesets %s\n")
1556 1556 % ', '.join(str(repo[r]) for r in immutable),
1557 1557 hint=_("see 'hg help phases' for details"))
1558 1558 cleanup = False
1559 1559
1560 1560 descendants = set()
1561 1561 if dstates:
1562 1562 descendants = set(repo.changelog.descendants(dstates))
1563 1563 if descendants - set(dstates):
1564 1564 repo.ui.warn(_("warning: new changesets detected on destination "
1565 1565 "branch, can't strip\n"))
1566 1566 cleanup = False
1567 1567
1568 1568 if cleanup:
1569 1569 shouldupdate = False
1570 1570 rebased = [s for r, s in state.items()
1571 1571 if s >= 0 and s != destmap[r]]
1572 1572 if rebased:
1573 1573 strippoints = [
1574 1574 c.node() for c in repo.set('roots(%ld)', rebased)]
1575 1575
1576 1576 updateifonnodes = set(rebased)
1577 1577 updateifonnodes.update(destmap.values())
1578 1578 updateifonnodes.add(originalwd)
1579 1579 shouldupdate = repo['.'].rev() in updateifonnodes
1580 1580
1581 1581 # Update away from the rebase if necessary
1582 1582 if shouldupdate or needupdate(repo, state):
1583 1583 mergemod.update(repo, originalwd, False, True)
1584 1584
1585 1585 # Strip from the first rebased revision
1586 1586 if rebased:
1587 1587 # no backup of rebased cset versions needed
1588 1588 repair.strip(repo.ui, repo, strippoints)
1589 1589
1590 1590 if activebookmark and activebookmark in repo._bookmarks:
1591 1591 bookmarks.activate(repo, activebookmark)
1592 1592
1593 1593 finally:
1594 1594 clearstatus(repo)
1595 1595 clearcollapsemsg(repo)
1596 1596 repo.ui.warn(_('rebase aborted\n'))
1597 1597 return 0
1598 1598
1599 1599 def sortsource(destmap):
1600 1600 """yield source revisions in an order that we only rebase things once
1601 1601
1602 1602 If source and destination overlaps, we should filter out revisions
1603 1603 depending on other revisions which hasn't been rebased yet.
1604 1604
1605 1605 Yield a sorted list of revisions each time.
1606 1606
1607 1607 For example, when rebasing A to B, B to C. This function yields [B], then
1608 1608 [A], indicating B needs to be rebased first.
1609 1609
1610 1610 Raise if there is a cycle so the rebase is impossible.
1611 1611 """
1612 1612 srcset = set(destmap)
1613 1613 while srcset:
1614 1614 srclist = sorted(srcset)
1615 1615 result = []
1616 1616 for r in srclist:
1617 1617 if destmap[r] not in srcset:
1618 1618 result.append(r)
1619 1619 if not result:
1620 1620 raise error.Abort(_('source and destination form a cycle'))
1621 1621 srcset -= set(result)
1622 1622 yield result
1623 1623
1624 1624 def buildstate(repo, destmap, collapse):
1625 1625 '''Define which revisions are going to be rebased and where
1626 1626
1627 1627 repo: repo
1628 1628 destmap: {srcrev: destrev}
1629 1629 '''
1630 1630 rebaseset = destmap.keys()
1631 1631 originalwd = repo['.'].rev()
1632 1632
1633 1633 # This check isn't strictly necessary, since mq detects commits over an
1634 1634 # applied patch. But it prevents messing up the working directory when
1635 1635 # a partially completed rebase is blocked by mq.
1636 1636 if 'qtip' in repo.tags():
1637 1637 mqapplied = set(repo[s.node].rev() for s in repo.mq.applied)
1638 1638 if set(destmap.values()) & mqapplied:
1639 1639 raise error.Abort(_('cannot rebase onto an applied mq patch'))
1640 1640
1641 1641 # Get "cycle" error early by exhausting the generator.
1642 1642 sortedsrc = list(sortsource(destmap)) # a list of sorted revs
1643 1643 if not sortedsrc:
1644 1644 raise error.Abort(_('no matching revisions'))
1645 1645
1646 1646 # Only check the first batch of revisions to rebase not depending on other
1647 1647 # rebaseset. This means "source is ancestor of destination" for the second
1648 1648 # (and following) batches of revisions are not checked here. We rely on
1649 1649 # "defineparents" to do that check.
1650 1650 roots = list(repo.set('roots(%ld)', sortedsrc[0]))
1651 1651 if not roots:
1652 1652 raise error.Abort(_('no matching revisions'))
1653 1653 def revof(r):
1654 1654 return r.rev()
1655 1655 roots = sorted(roots, key=revof)
1656 1656 state = dict.fromkeys(rebaseset, revtodo)
1657 1657 emptyrebase = (len(sortedsrc) == 1)
1658 1658 for root in roots:
1659 1659 dest = repo[destmap[root.rev()]]
1660 1660 commonbase = root.ancestor(dest)
1661 1661 if commonbase == root:
1662 1662 raise error.Abort(_('source is ancestor of destination'))
1663 1663 if commonbase == dest:
1664 1664 wctx = repo[None]
1665 1665 if dest == wctx.p1():
1666 1666 # when rebasing to '.', it will use the current wd branch name
1667 1667 samebranch = root.branch() == wctx.branch()
1668 1668 else:
1669 1669 samebranch = root.branch() == dest.branch()
1670 1670 if not collapse and samebranch and dest in root.parents():
1671 1671 # mark the revision as done by setting its new revision
1672 1672 # equal to its old (current) revisions
1673 1673 state[root.rev()] = root.rev()
1674 1674 repo.ui.debug('source is a child of destination\n')
1675 1675 continue
1676 1676
1677 1677 emptyrebase = False
1678 1678 repo.ui.debug('rebase onto %s starting from %s\n' % (dest, root))
1679 1679 if emptyrebase:
1680 1680 return None
1681 1681 for rev in sorted(state):
1682 1682 parents = [p for p in repo.changelog.parentrevs(rev) if p != nullrev]
1683 1683 # if all parents of this revision are done, then so is this revision
1684 1684 if parents and all((state.get(p) == p for p in parents)):
1685 1685 state[rev] = rev
1686 1686 return originalwd, destmap, state
1687 1687
1688 1688 def clearrebased(ui, repo, destmap, state, skipped, collapsedas=None,
1689 1689 keepf=False, fm=None):
1690 1690 """dispose of rebased revision at the end of the rebase
1691 1691
1692 1692 If `collapsedas` is not None, the rebase was a collapse whose result if the
1693 1693 `collapsedas` node.
1694 1694
1695 1695 If `keepf` is not True, the rebase has --keep set and no nodes should be
1696 1696 removed (but bookmarks still need to be moved).
1697 1697 """
1698 1698 tonode = repo.changelog.node
1699 1699 replacements = {}
1700 1700 moves = {}
1701 1701 for rev, newrev in sorted(state.items()):
1702 1702 if newrev >= 0 and newrev != rev:
1703 1703 oldnode = tonode(rev)
1704 1704 newnode = collapsedas or tonode(newrev)
1705 1705 moves[oldnode] = newnode
1706 1706 if not keepf:
1707 1707 if rev in skipped:
1708 1708 succs = ()
1709 1709 else:
1710 1710 succs = (newnode,)
1711 1711 replacements[oldnode] = succs
1712 1712 scmutil.cleanupnodes(repo, replacements, 'rebase', moves)
1713 1713 if fm:
1714 1714 hf = fm.hexfunc
1715 1715 fl = fm.formatlist
1716 1716 fd = fm.formatdict
1717 1717 nodechanges = fd({hf(oldn): fl([hf(n) for n in newn], name='node')
1718 1718 for oldn, newn in replacements.iteritems()},
1719 1719 key="oldnode", value="newnodes")
1720 1720 fm.data(nodechanges=nodechanges)
1721 1721
1722 1722 def pullrebase(orig, ui, repo, *args, **opts):
1723 1723 'Call rebase after pull if the latter has been invoked with --rebase'
1724 1724 ret = None
1725 1725 if opts.get(r'rebase'):
1726 1726 if ui.configbool('commands', 'rebase.requiredest'):
1727 1727 msg = _('rebase destination required by configuration')
1728 1728 hint = _('use hg pull followed by hg rebase -d DEST')
1729 1729 raise error.Abort(msg, hint=hint)
1730 1730
1731 1731 with repo.wlock(), repo.lock():
1732 1732 if opts.get(r'update'):
1733 1733 del opts[r'update']
1734 1734 ui.debug('--update and --rebase are not compatible, ignoring '
1735 1735 'the update flag\n')
1736 1736
1737 1737 cmdutil.checkunfinished(repo)
1738 1738 cmdutil.bailifchanged(repo, hint=_('cannot pull with rebase: '
1739 1739 'please commit or shelve your changes first'))
1740 1740
1741 1741 revsprepull = len(repo)
1742 1742 origpostincoming = commands.postincoming
1743 1743 def _dummy(*args, **kwargs):
1744 1744 pass
1745 1745 commands.postincoming = _dummy
1746 1746 try:
1747 1747 ret = orig(ui, repo, *args, **opts)
1748 1748 finally:
1749 1749 commands.postincoming = origpostincoming
1750 1750 revspostpull = len(repo)
1751 1751 if revspostpull > revsprepull:
1752 1752 # --rev option from pull conflict with rebase own --rev
1753 1753 # dropping it
1754 1754 if r'rev' in opts:
1755 1755 del opts[r'rev']
1756 1756 # positional argument from pull conflicts with rebase's own
1757 1757 # --source.
1758 1758 if r'source' in opts:
1759 1759 del opts[r'source']
1760 1760 # revsprepull is the len of the repo, not revnum of tip.
1761 1761 destspace = list(repo.changelog.revs(start=revsprepull))
1762 1762 opts[r'_destspace'] = destspace
1763 1763 try:
1764 1764 rebase(ui, repo, **opts)
1765 1765 except error.NoMergeDestAbort:
1766 1766 # we can maybe update instead
1767 1767 rev, _a, _b = destutil.destupdate(repo)
1768 1768 if rev == repo['.'].rev():
1769 1769 ui.status(_('nothing to rebase\n'))
1770 1770 else:
1771 1771 ui.status(_('nothing to rebase - updating instead\n'))
1772 1772 # not passing argument to get the bare update behavior
1773 1773 # with warning and trumpets
1774 1774 commands.update(ui, repo)
1775 1775 else:
1776 1776 if opts.get(r'tool'):
1777 1777 raise error.Abort(_('--tool can only be used with --rebase'))
1778 1778 ret = orig(ui, repo, *args, **opts)
1779 1779
1780 1780 return ret
1781 1781
1782 1782 def _filterobsoleterevs(repo, revs):
1783 1783 """returns a set of the obsolete revisions in revs"""
1784 1784 return set(r for r in revs if repo[r].obsolete())
1785 1785
1786 1786 def _computeobsoletenotrebased(repo, rebaseobsrevs, destmap):
1787 1787 """Return (obsoletenotrebased, obsoletewithoutsuccessorindestination).
1788 1788
1789 1789 `obsoletenotrebased` is a mapping mapping obsolete => successor for all
1790 1790 obsolete nodes to be rebased given in `rebaseobsrevs`.
1791 1791
1792 1792 `obsoletewithoutsuccessorindestination` is a set with obsolete revisions
1793 1793 without a successor in destination.
1794 1794
1795 1795 `obsoleteextinctsuccessors` is a set of obsolete revisions with only
1796 1796 obsolete successors.
1797 1797 """
1798 1798 obsoletenotrebased = {}
1799 1799 obsoletewithoutsuccessorindestination = set([])
1800 1800 obsoleteextinctsuccessors = set([])
1801 1801
1802 1802 assert repo.filtername is None
1803 1803 cl = repo.changelog
1804 1804 nodemap = cl.nodemap
1805 1805 extinctnodes = set(cl.node(r) for r in repo.revs('extinct()'))
1806 1806 for srcrev in rebaseobsrevs:
1807 1807 srcnode = cl.node(srcrev)
1808 1808 destnode = cl.node(destmap[srcrev])
1809 1809 # XXX: more advanced APIs are required to handle split correctly
1810 1810 successors = set(obsutil.allsuccessors(repo.obsstore, [srcnode]))
1811 1811 # obsutil.allsuccessors includes node itself
1812 1812 successors.remove(srcnode)
1813 1813 if successors.issubset(extinctnodes):
1814 1814 # all successors are extinct
1815 1815 obsoleteextinctsuccessors.add(srcrev)
1816 1816 if not successors:
1817 1817 # no successor
1818 1818 obsoletenotrebased[srcrev] = None
1819 1819 else:
1820 1820 for succnode in successors:
1821 1821 if succnode not in nodemap:
1822 1822 continue
1823 1823 if cl.isancestor(succnode, destnode):
1824 1824 obsoletenotrebased[srcrev] = nodemap[succnode]
1825 1825 break
1826 1826 else:
1827 1827 # If 'srcrev' has a successor in rebase set but none in
1828 1828 # destination (which would be catched above), we shall skip it
1829 1829 # and its descendants to avoid divergence.
1830 1830 if any(nodemap[s] in destmap for s in successors):
1831 1831 obsoletewithoutsuccessorindestination.add(srcrev)
1832 1832
1833 1833 return (
1834 1834 obsoletenotrebased,
1835 1835 obsoletewithoutsuccessorindestination,
1836 1836 obsoleteextinctsuccessors,
1837 1837 )
1838 1838
1839 1839 def summaryhook(ui, repo):
1840 1840 if not repo.vfs.exists('rebasestate'):
1841 1841 return
1842 1842 try:
1843 1843 rbsrt = rebaseruntime(repo, ui, {})
1844 1844 rbsrt.restorestatus()
1845 1845 state = rbsrt.state
1846 1846 except error.RepoLookupError:
1847 1847 # i18n: column positioning for "hg summary"
1848 1848 msg = _('rebase: (use "hg rebase --abort" to clear broken state)\n')
1849 1849 ui.write(msg)
1850 1850 return
1851 1851 numrebased = len([i for i in state.itervalues() if i >= 0])
1852 1852 # i18n: column positioning for "hg summary"
1853 1853 ui.write(_('rebase: %s, %s (rebase --continue)\n') %
1854 1854 (ui.label(_('%d rebased'), 'rebase.rebased') % numrebased,
1855 1855 ui.label(_('%d remaining'), 'rebase.remaining') %
1856 1856 (len(state) - numrebased)))
1857 1857
1858 1858 def uisetup(ui):
1859 1859 #Replace pull with a decorator to provide --rebase option
1860 1860 entry = extensions.wrapcommand(commands.table, 'pull', pullrebase)
1861 1861 entry[1].append(('', 'rebase', None,
1862 1862 _("rebase working directory to branch head")))
1863 1863 entry[1].append(('t', 'tool', '',
1864 1864 _("specify merge tool for rebase")))
1865 1865 cmdutil.summaryhooks.add('rebase', summaryhook)
1866 1866 cmdutil.unfinishedstates.append(
1867 1867 ['rebasestate', False, False, _('rebase in progress'),
1868 1868 _("use 'hg rebase --continue' or 'hg rebase --abort'")])
1869 1869 cmdutil.afterresolvedstates.append(
1870 1870 ['rebasestate', _('hg rebase --continue')])
General Comments 0
You need to be logged in to leave comments. Login now