##// END OF EJS Templates
blackbox: do not translate the log messages...
Durham Goode -
r18691:4f485bd6 default
parent child Browse files
Show More
@@ -1,841 +1,841
1 1 # dispatch.py - command dispatching for mercurial
2 2 #
3 3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 from i18n import _
9 9 import os, sys, atexit, signal, pdb, socket, errno, shlex, time, traceback, re
10 10 import util, commands, hg, fancyopts, extensions, hook, error
11 11 import cmdutil, encoding
12 12 import ui as uimod
13 13
14 14 class request(object):
15 15 def __init__(self, args, ui=None, repo=None, fin=None, fout=None,
16 16 ferr=None):
17 17 self.args = args
18 18 self.ui = ui
19 19 self.repo = repo
20 20
21 21 # input/output/error streams
22 22 self.fin = fin
23 23 self.fout = fout
24 24 self.ferr = ferr
25 25
26 26 def run():
27 27 "run the command in sys.argv"
28 28 sys.exit((dispatch(request(sys.argv[1:])) or 0) & 255)
29 29
30 30 def dispatch(req):
31 31 "run the command specified in req.args"
32 32 if req.ferr:
33 33 ferr = req.ferr
34 34 elif req.ui:
35 35 ferr = req.ui.ferr
36 36 else:
37 37 ferr = sys.stderr
38 38
39 39 try:
40 40 if not req.ui:
41 41 req.ui = uimod.ui()
42 42 if '--traceback' in req.args:
43 43 req.ui.setconfig('ui', 'traceback', 'on')
44 44
45 45 # set ui streams from the request
46 46 if req.fin:
47 47 req.ui.fin = req.fin
48 48 if req.fout:
49 49 req.ui.fout = req.fout
50 50 if req.ferr:
51 51 req.ui.ferr = req.ferr
52 52 except util.Abort, inst:
53 53 ferr.write(_("abort: %s\n") % inst)
54 54 if inst.hint:
55 55 ferr.write(_("(%s)\n") % inst.hint)
56 56 return -1
57 57 except error.ParseError, inst:
58 58 if len(inst.args) > 1:
59 59 ferr.write(_("hg: parse error at %s: %s\n") %
60 60 (inst.args[1], inst.args[0]))
61 61 else:
62 62 ferr.write(_("hg: parse error: %s\n") % inst.args[0])
63 63 return -1
64 64
65 65 return _runcatch(req)
66 66
67 67 def _runcatch(req):
68 68 def catchterm(*args):
69 69 raise error.SignalInterrupt
70 70
71 71 ui = req.ui
72 72 try:
73 73 for name in 'SIGBREAK', 'SIGHUP', 'SIGTERM':
74 74 num = getattr(signal, name, None)
75 75 if num:
76 76 signal.signal(num, catchterm)
77 77 except ValueError:
78 78 pass # happens if called in a thread
79 79
80 80 try:
81 81 try:
82 82 # enter the debugger before command execution
83 83 if '--debugger' in req.args:
84 84 ui.warn(_("entering debugger - "
85 85 "type c to continue starting hg or h for help\n"))
86 86 pdb.set_trace()
87 87 try:
88 88 return _dispatch(req)
89 89 finally:
90 90 ui.flush()
91 91 except: # re-raises
92 92 # enter the debugger when we hit an exception
93 93 if '--debugger' in req.args:
94 94 traceback.print_exc()
95 95 pdb.post_mortem(sys.exc_info()[2])
96 96 ui.traceback()
97 97 raise
98 98
99 99 # Global exception handling, alphabetically
100 100 # Mercurial-specific first, followed by built-in and library exceptions
101 101 except error.AmbiguousCommand, inst:
102 102 ui.warn(_("hg: command '%s' is ambiguous:\n %s\n") %
103 103 (inst.args[0], " ".join(inst.args[1])))
104 104 except error.ParseError, inst:
105 105 if len(inst.args) > 1:
106 106 ui.warn(_("hg: parse error at %s: %s\n") %
107 107 (inst.args[1], inst.args[0]))
108 108 else:
109 109 ui.warn(_("hg: parse error: %s\n") % inst.args[0])
110 110 return -1
111 111 except error.LockHeld, inst:
112 112 if inst.errno == errno.ETIMEDOUT:
113 113 reason = _('timed out waiting for lock held by %s') % inst.locker
114 114 else:
115 115 reason = _('lock held by %s') % inst.locker
116 116 ui.warn(_("abort: %s: %s\n") % (inst.desc or inst.filename, reason))
117 117 except error.LockUnavailable, inst:
118 118 ui.warn(_("abort: could not lock %s: %s\n") %
119 119 (inst.desc or inst.filename, inst.strerror))
120 120 except error.CommandError, inst:
121 121 if inst.args[0]:
122 122 ui.warn(_("hg %s: %s\n") % (inst.args[0], inst.args[1]))
123 123 commands.help_(ui, inst.args[0], full=False, command=True)
124 124 else:
125 125 ui.warn(_("hg: %s\n") % inst.args[1])
126 126 commands.help_(ui, 'shortlist')
127 127 except error.OutOfBandError, inst:
128 128 ui.warn(_("abort: remote error:\n"))
129 129 ui.warn(''.join(inst.args))
130 130 except error.RepoError, inst:
131 131 ui.warn(_("abort: %s!\n") % inst)
132 132 if inst.hint:
133 133 ui.warn(_("(%s)\n") % inst.hint)
134 134 except error.ResponseError, inst:
135 135 ui.warn(_("abort: %s") % inst.args[0])
136 136 if not isinstance(inst.args[1], basestring):
137 137 ui.warn(" %r\n" % (inst.args[1],))
138 138 elif not inst.args[1]:
139 139 ui.warn(_(" empty string\n"))
140 140 else:
141 141 ui.warn("\n%r\n" % util.ellipsis(inst.args[1]))
142 142 except error.RevlogError, inst:
143 143 ui.warn(_("abort: %s!\n") % inst)
144 144 except error.SignalInterrupt:
145 145 ui.warn(_("killed!\n"))
146 146 except error.UnknownCommand, inst:
147 147 ui.warn(_("hg: unknown command '%s'\n") % inst.args[0])
148 148 try:
149 149 # check if the command is in a disabled extension
150 150 # (but don't check for extensions themselves)
151 151 commands.help_(ui, inst.args[0], unknowncmd=True)
152 152 except error.UnknownCommand:
153 153 commands.help_(ui, 'shortlist')
154 154 except util.Abort, inst:
155 155 ui.warn(_("abort: %s\n") % inst)
156 156 if inst.hint:
157 157 ui.warn(_("(%s)\n") % inst.hint)
158 158 except ImportError, inst:
159 159 ui.warn(_("abort: %s!\n") % inst)
160 160 m = str(inst).split()[-1]
161 161 if m in "mpatch bdiff".split():
162 162 ui.warn(_("(did you forget to compile extensions?)\n"))
163 163 elif m in "zlib".split():
164 164 ui.warn(_("(is your Python install correct?)\n"))
165 165 except IOError, inst:
166 166 if util.safehasattr(inst, "code"):
167 167 ui.warn(_("abort: %s\n") % inst)
168 168 elif util.safehasattr(inst, "reason"):
169 169 try: # usually it is in the form (errno, strerror)
170 170 reason = inst.reason.args[1]
171 171 except (AttributeError, IndexError):
172 172 # it might be anything, for example a string
173 173 reason = inst.reason
174 174 ui.warn(_("abort: error: %s\n") % reason)
175 175 elif util.safehasattr(inst, "args") and inst.args[0] == errno.EPIPE:
176 176 if ui.debugflag:
177 177 ui.warn(_("broken pipe\n"))
178 178 elif getattr(inst, "strerror", None):
179 179 if getattr(inst, "filename", None):
180 180 ui.warn(_("abort: %s: %s\n") % (inst.strerror, inst.filename))
181 181 else:
182 182 ui.warn(_("abort: %s\n") % inst.strerror)
183 183 else:
184 184 raise
185 185 except OSError, inst:
186 186 if getattr(inst, "filename", None) is not None:
187 187 ui.warn(_("abort: %s: '%s'\n") % (inst.strerror, inst.filename))
188 188 else:
189 189 ui.warn(_("abort: %s\n") % inst.strerror)
190 190 except KeyboardInterrupt:
191 191 try:
192 192 ui.warn(_("interrupted!\n"))
193 193 except IOError, inst:
194 194 if inst.errno == errno.EPIPE:
195 195 if ui.debugflag:
196 196 ui.warn(_("\nbroken pipe\n"))
197 197 else:
198 198 raise
199 199 except MemoryError:
200 200 ui.warn(_("abort: out of memory\n"))
201 201 except SystemExit, inst:
202 202 # Commands shouldn't sys.exit directly, but give a return code.
203 203 # Just in case catch this and and pass exit code to caller.
204 204 return inst.code
205 205 except socket.error, inst:
206 206 ui.warn(_("abort: %s\n") % inst.args[-1])
207 207 except: # re-raises
208 208 myver = util.version()
209 209 # For compatibility checking, we discard the portion of the hg
210 210 # version after the + on the assumption that if a "normal
211 211 # user" is running a build with a + in it the packager
212 212 # probably built from fairly close to a tag and anyone with a
213 213 # 'make local' copy of hg (where the version number can be out
214 214 # of date) will be clueful enough to notice the implausible
215 215 # version number and try updating.
216 216 compare = myver.split('+')[0]
217 217 ct = tuplever(compare)
218 218 worst = None, ct, ''
219 219 for name, mod in extensions.extensions():
220 220 testedwith = getattr(mod, 'testedwith', '')
221 221 report = getattr(mod, 'buglink', _('the extension author.'))
222 222 if not testedwith.strip():
223 223 # We found an untested extension. It's likely the culprit.
224 224 worst = name, 'unknown', report
225 225 break
226 226 if compare not in testedwith.split() and testedwith != 'internal':
227 227 tested = [tuplever(v) for v in testedwith.split()]
228 228 lower = [t for t in tested if t < ct]
229 229 nearest = max(lower or tested)
230 230 if worst[0] is None or nearest < worst[1]:
231 231 worst = name, nearest, report
232 232 if worst[0] is not None:
233 233 name, testedwith, report = worst
234 234 if not isinstance(testedwith, str):
235 235 testedwith = '.'.join([str(c) for c in testedwith])
236 236 warning = (_('** Unknown exception encountered with '
237 237 'possibly-broken third-party extension %s\n'
238 238 '** which supports versions %s of Mercurial.\n'
239 239 '** Please disable %s and try your action again.\n'
240 240 '** If that fixes the bug please report it to %s\n')
241 241 % (name, testedwith, name, report))
242 242 else:
243 243 warning = (_("** unknown exception encountered, "
244 244 "please report by visiting\n") +
245 245 _("** http://mercurial.selenic.com/wiki/BugTracker\n"))
246 246 warning += ((_("** Python %s\n") % sys.version.replace('\n', '')) +
247 247 (_("** Mercurial Distributed SCM (version %s)\n") % myver) +
248 248 (_("** Extensions loaded: %s\n") %
249 249 ", ".join([x[0] for x in extensions.extensions()])))
250 250 ui.log("commandexception", "%s\n%s\n", warning, traceback.format_exc())
251 251 ui.warn(warning)
252 252 raise
253 253
254 254 return -1
255 255
256 256 def tuplever(v):
257 257 try:
258 258 return tuple([int(i) for i in v.split('.')])
259 259 except ValueError:
260 260 return tuple()
261 261
262 262 def aliasargs(fn, givenargs):
263 263 args = getattr(fn, 'args', [])
264 264 if args:
265 265 cmd = ' '.join(map(util.shellquote, args))
266 266
267 267 nums = []
268 268 def replacer(m):
269 269 num = int(m.group(1)) - 1
270 270 nums.append(num)
271 271 if num < len(givenargs):
272 272 return givenargs[num]
273 273 raise util.Abort(_('too few arguments for command alias'))
274 274 cmd = re.sub(r'\$(\d+|\$)', replacer, cmd)
275 275 givenargs = [x for i, x in enumerate(givenargs)
276 276 if i not in nums]
277 277 args = shlex.split(cmd)
278 278 return args + givenargs
279 279
280 280 class cmdalias(object):
281 281 def __init__(self, name, definition, cmdtable):
282 282 self.name = self.cmd = name
283 283 self.cmdname = ''
284 284 self.definition = definition
285 285 self.args = []
286 286 self.opts = []
287 287 self.help = ''
288 288 self.norepo = True
289 289 self.optionalrepo = False
290 290 self.badalias = False
291 291
292 292 try:
293 293 aliases, entry = cmdutil.findcmd(self.name, cmdtable)
294 294 for alias, e in cmdtable.iteritems():
295 295 if e is entry:
296 296 self.cmd = alias
297 297 break
298 298 self.shadows = True
299 299 except error.UnknownCommand:
300 300 self.shadows = False
301 301
302 302 if not self.definition:
303 303 def fn(ui, *args):
304 304 ui.warn(_("no definition for alias '%s'\n") % self.name)
305 305 return 1
306 306 self.fn = fn
307 307 self.badalias = True
308 308 return
309 309
310 310 if self.definition.startswith('!'):
311 311 self.shell = True
312 312 def fn(ui, *args):
313 313 env = {'HG_ARGS': ' '.join((self.name,) + args)}
314 314 def _checkvar(m):
315 315 if m.groups()[0] == '$':
316 316 return m.group()
317 317 elif int(m.groups()[0]) <= len(args):
318 318 return m.group()
319 319 else:
320 320 ui.debug("No argument found for substitution "
321 321 "of %i variable in alias '%s' definition."
322 322 % (int(m.groups()[0]), self.name))
323 323 return ''
324 324 cmd = re.sub(r'\$(\d+|\$)', _checkvar, self.definition[1:])
325 325 replace = dict((str(i + 1), arg) for i, arg in enumerate(args))
326 326 replace['0'] = self.name
327 327 replace['@'] = ' '.join(args)
328 328 cmd = util.interpolate(r'\$', replace, cmd, escape_prefix=True)
329 329 return util.system(cmd, environ=env, out=ui.fout)
330 330 self.fn = fn
331 331 return
332 332
333 333 args = shlex.split(self.definition)
334 334 self.cmdname = cmd = args.pop(0)
335 335 args = map(util.expandpath, args)
336 336
337 337 for invalidarg in ("--cwd", "-R", "--repository", "--repo"):
338 338 if _earlygetopt([invalidarg], args):
339 339 def fn(ui, *args):
340 340 ui.warn(_("error in definition for alias '%s': %s may only "
341 341 "be given on the command line\n")
342 342 % (self.name, invalidarg))
343 343 return 1
344 344
345 345 self.fn = fn
346 346 self.badalias = True
347 347 return
348 348
349 349 try:
350 350 tableentry = cmdutil.findcmd(cmd, cmdtable, False)[1]
351 351 if len(tableentry) > 2:
352 352 self.fn, self.opts, self.help = tableentry
353 353 else:
354 354 self.fn, self.opts = tableentry
355 355
356 356 self.args = aliasargs(self.fn, args)
357 357 if cmd not in commands.norepo.split(' '):
358 358 self.norepo = False
359 359 if cmd in commands.optionalrepo.split(' '):
360 360 self.optionalrepo = True
361 361 if self.help.startswith("hg " + cmd):
362 362 # drop prefix in old-style help lines so hg shows the alias
363 363 self.help = self.help[4 + len(cmd):]
364 364 self.__doc__ = self.fn.__doc__
365 365
366 366 except error.UnknownCommand:
367 367 def fn(ui, *args):
368 368 ui.warn(_("alias '%s' resolves to unknown command '%s'\n") \
369 369 % (self.name, cmd))
370 370 try:
371 371 # check if the command is in a disabled extension
372 372 commands.help_(ui, cmd, unknowncmd=True)
373 373 except error.UnknownCommand:
374 374 pass
375 375 return 1
376 376 self.fn = fn
377 377 self.badalias = True
378 378 except error.AmbiguousCommand:
379 379 def fn(ui, *args):
380 380 ui.warn(_("alias '%s' resolves to ambiguous command '%s'\n") \
381 381 % (self.name, cmd))
382 382 return 1
383 383 self.fn = fn
384 384 self.badalias = True
385 385
386 386 def __call__(self, ui, *args, **opts):
387 387 if self.shadows:
388 388 ui.debug("alias '%s' shadows command '%s'\n" %
389 389 (self.name, self.cmdname))
390 390
391 391 if util.safehasattr(self, 'shell'):
392 392 return self.fn(ui, *args, **opts)
393 393 else:
394 394 try:
395 395 util.checksignature(self.fn)(ui, *args, **opts)
396 396 except error.SignatureError:
397 397 args = ' '.join([self.cmdname] + self.args)
398 398 ui.debug("alias '%s' expands to '%s'\n" % (self.name, args))
399 399 raise
400 400
401 401 def addaliases(ui, cmdtable):
402 402 # aliases are processed after extensions have been loaded, so they
403 403 # may use extension commands. Aliases can also use other alias definitions,
404 404 # but only if they have been defined prior to the current definition.
405 405 for alias, definition in ui.configitems('alias'):
406 406 aliasdef = cmdalias(alias, definition, cmdtable)
407 407
408 408 try:
409 409 olddef = cmdtable[aliasdef.cmd][0]
410 410 if olddef.definition == aliasdef.definition:
411 411 continue
412 412 except (KeyError, AttributeError):
413 413 # definition might not exist or it might not be a cmdalias
414 414 pass
415 415
416 416 cmdtable[aliasdef.name] = (aliasdef, aliasdef.opts, aliasdef.help)
417 417 if aliasdef.norepo:
418 418 commands.norepo += ' %s' % alias
419 419 if aliasdef.optionalrepo:
420 420 commands.optionalrepo += ' %s' % alias
421 421
422 422 def _parse(ui, args):
423 423 options = {}
424 424 cmdoptions = {}
425 425
426 426 try:
427 427 args = fancyopts.fancyopts(args, commands.globalopts, options)
428 428 except fancyopts.getopt.GetoptError, inst:
429 429 raise error.CommandError(None, inst)
430 430
431 431 if args:
432 432 cmd, args = args[0], args[1:]
433 433 aliases, entry = cmdutil.findcmd(cmd, commands.table,
434 434 ui.configbool("ui", "strict"))
435 435 cmd = aliases[0]
436 436 args = aliasargs(entry[0], args)
437 437 defaults = ui.config("defaults", cmd)
438 438 if defaults:
439 439 args = map(util.expandpath, shlex.split(defaults)) + args
440 440 c = list(entry[1])
441 441 else:
442 442 cmd = None
443 443 c = []
444 444
445 445 # combine global options into local
446 446 for o in commands.globalopts:
447 447 c.append((o[0], o[1], options[o[1]], o[3]))
448 448
449 449 try:
450 450 args = fancyopts.fancyopts(args, c, cmdoptions, True)
451 451 except fancyopts.getopt.GetoptError, inst:
452 452 raise error.CommandError(cmd, inst)
453 453
454 454 # separate global options back out
455 455 for o in commands.globalopts:
456 456 n = o[1]
457 457 options[n] = cmdoptions[n]
458 458 del cmdoptions[n]
459 459
460 460 return (cmd, cmd and entry[0] or None, args, options, cmdoptions)
461 461
462 462 def _parseconfig(ui, config):
463 463 """parse the --config options from the command line"""
464 464 configs = []
465 465
466 466 for cfg in config:
467 467 try:
468 468 name, value = cfg.split('=', 1)
469 469 section, name = name.split('.', 1)
470 470 if not section or not name:
471 471 raise IndexError
472 472 ui.setconfig(section, name, value)
473 473 configs.append((section, name, value))
474 474 except (IndexError, ValueError):
475 475 raise util.Abort(_('malformed --config option: %r '
476 476 '(use --config section.name=value)') % cfg)
477 477
478 478 return configs
479 479
480 480 def _earlygetopt(aliases, args):
481 481 """Return list of values for an option (or aliases).
482 482
483 483 The values are listed in the order they appear in args.
484 484 The options and values are removed from args.
485 485 """
486 486 try:
487 487 argcount = args.index("--")
488 488 except ValueError:
489 489 argcount = len(args)
490 490 shortopts = [opt for opt in aliases if len(opt) == 2]
491 491 values = []
492 492 pos = 0
493 493 while pos < argcount:
494 494 if args[pos] in aliases:
495 495 if pos + 1 >= argcount:
496 496 # ignore and let getopt report an error if there is no value
497 497 break
498 498 del args[pos]
499 499 values.append(args.pop(pos))
500 500 argcount -= 2
501 501 elif args[pos][:2] in shortopts:
502 502 # short option can have no following space, e.g. hg log -Rfoo
503 503 values.append(args.pop(pos)[2:])
504 504 argcount -= 1
505 505 else:
506 506 pos += 1
507 507 return values
508 508
509 509 def runcommand(lui, repo, cmd, fullargs, ui, options, d, cmdpats, cmdoptions):
510 510 # run pre-hook, and abort if it fails
511 511 ret = hook.hook(lui, repo, "pre-%s" % cmd, False, args=" ".join(fullargs),
512 512 pats=cmdpats, opts=cmdoptions)
513 513 if ret:
514 514 return ret
515 515 ret = _runcommand(ui, options, cmd, d)
516 516 # run post-hook, passing command result
517 517 hook.hook(lui, repo, "post-%s" % cmd, False, args=" ".join(fullargs),
518 518 result=ret, pats=cmdpats, opts=cmdoptions)
519 519 return ret
520 520
521 521 def _getlocal(ui, rpath):
522 522 """Return (path, local ui object) for the given target path.
523 523
524 524 Takes paths in [cwd]/.hg/hgrc into account."
525 525 """
526 526 try:
527 527 wd = os.getcwd()
528 528 except OSError, e:
529 529 raise util.Abort(_("error getting current working directory: %s") %
530 530 e.strerror)
531 531 path = cmdutil.findrepo(wd) or ""
532 532 if not path:
533 533 lui = ui
534 534 else:
535 535 lui = ui.copy()
536 536 lui.readconfig(os.path.join(path, ".hg", "hgrc"), path)
537 537
538 538 if rpath and rpath[-1]:
539 539 path = lui.expandpath(rpath[-1])
540 540 lui = ui.copy()
541 541 lui.readconfig(os.path.join(path, ".hg", "hgrc"), path)
542 542
543 543 return path, lui
544 544
545 545 def _checkshellalias(lui, ui, args):
546 546 options = {}
547 547
548 548 try:
549 549 args = fancyopts.fancyopts(args, commands.globalopts, options)
550 550 except fancyopts.getopt.GetoptError:
551 551 return
552 552
553 553 if not args:
554 554 return
555 555
556 556 norepo = commands.norepo
557 557 optionalrepo = commands.optionalrepo
558 558 def restorecommands():
559 559 commands.norepo = norepo
560 560 commands.optionalrepo = optionalrepo
561 561
562 562 cmdtable = commands.table.copy()
563 563 addaliases(lui, cmdtable)
564 564
565 565 cmd = args[0]
566 566 try:
567 567 aliases, entry = cmdutil.findcmd(cmd, cmdtable,
568 568 lui.configbool("ui", "strict"))
569 569 except (error.AmbiguousCommand, error.UnknownCommand):
570 570 restorecommands()
571 571 return
572 572
573 573 cmd = aliases[0]
574 574 fn = entry[0]
575 575
576 576 if cmd and util.safehasattr(fn, 'shell'):
577 577 d = lambda: fn(ui, *args[1:])
578 578 return lambda: runcommand(lui, None, cmd, args[:1], ui, options, d,
579 579 [], {})
580 580
581 581 restorecommands()
582 582
583 583 _loaded = set()
584 584 def _dispatch(req):
585 585 args = req.args
586 586 ui = req.ui
587 587
588 588 # read --config before doing anything else
589 589 # (e.g. to change trust settings for reading .hg/hgrc)
590 590 cfgs = _parseconfig(ui, _earlygetopt(['--config'], args))
591 591
592 592 # check for cwd
593 593 cwd = _earlygetopt(['--cwd'], args)
594 594 if cwd:
595 595 os.chdir(cwd[-1])
596 596
597 597 rpath = _earlygetopt(["-R", "--repository", "--repo"], args)
598 598 path, lui = _getlocal(ui, rpath)
599 599
600 600 # Now that we're operating in the right directory/repository with
601 601 # the right config settings, check for shell aliases
602 602 shellaliasfn = _checkshellalias(lui, ui, args)
603 603 if shellaliasfn:
604 604 return shellaliasfn()
605 605
606 606 # Configure extensions in phases: uisetup, extsetup, cmdtable, and
607 607 # reposetup. Programs like TortoiseHg will call _dispatch several
608 608 # times so we keep track of configured extensions in _loaded.
609 609 extensions.loadall(lui)
610 610 exts = [ext for ext in extensions.extensions() if ext[0] not in _loaded]
611 611 # Propagate any changes to lui.__class__ by extensions
612 612 ui.__class__ = lui.__class__
613 613
614 614 # (uisetup and extsetup are handled in extensions.loadall)
615 615
616 616 for name, module in exts:
617 617 cmdtable = getattr(module, 'cmdtable', {})
618 618 overrides = [cmd for cmd in cmdtable if cmd in commands.table]
619 619 if overrides:
620 620 ui.warn(_("extension '%s' overrides commands: %s\n")
621 621 % (name, " ".join(overrides)))
622 622 commands.table.update(cmdtable)
623 623 _loaded.add(name)
624 624
625 625 # (reposetup is handled in hg.repository)
626 626
627 627 addaliases(lui, commands.table)
628 628
629 629 # check for fallback encoding
630 630 fallback = lui.config('ui', 'fallbackencoding')
631 631 if fallback:
632 632 encoding.fallbackencoding = fallback
633 633
634 634 fullargs = args
635 635 cmd, func, args, options, cmdoptions = _parse(lui, args)
636 636
637 637 if options["config"]:
638 638 raise util.Abort(_("option --config may not be abbreviated!"))
639 639 if options["cwd"]:
640 640 raise util.Abort(_("option --cwd may not be abbreviated!"))
641 641 if options["repository"]:
642 642 raise util.Abort(_(
643 643 "option -R has to be separated from other options (e.g. not -qR) "
644 644 "and --repository may only be abbreviated as --repo!"))
645 645
646 646 if options["encoding"]:
647 647 encoding.encoding = options["encoding"]
648 648 if options["encodingmode"]:
649 649 encoding.encodingmode = options["encodingmode"]
650 650 if options["time"]:
651 651 def get_times():
652 652 t = os.times()
653 653 if t[4] == 0.0: # Windows leaves this as zero, so use time.clock()
654 654 t = (t[0], t[1], t[2], t[3], time.clock())
655 655 return t
656 656 s = get_times()
657 657 def print_time():
658 658 t = get_times()
659 659 ui.warn(_("time: real %.3f secs (user %.3f+%.3f sys %.3f+%.3f)\n") %
660 660 (t[4]-s[4], t[0]-s[0], t[2]-s[2], t[1]-s[1], t[3]-s[3]))
661 661 atexit.register(print_time)
662 662
663 663 uis = set([ui, lui])
664 664
665 665 if req.repo:
666 666 uis.add(req.repo.ui)
667 667
668 668 # copy configs that were passed on the cmdline (--config) to the repo ui
669 669 for cfg in cfgs:
670 670 req.repo.ui.setconfig(*cfg)
671 671
672 672 if options['verbose'] or options['debug'] or options['quiet']:
673 673 for opt in ('verbose', 'debug', 'quiet'):
674 674 val = str(bool(options[opt]))
675 675 for ui_ in uis:
676 676 ui_.setconfig('ui', opt, val)
677 677
678 678 if options['traceback']:
679 679 for ui_ in uis:
680 680 ui_.setconfig('ui', 'traceback', 'on')
681 681
682 682 if options['noninteractive']:
683 683 for ui_ in uis:
684 684 ui_.setconfig('ui', 'interactive', 'off')
685 685
686 686 if cmdoptions.get('insecure', False):
687 687 for ui_ in uis:
688 688 ui_.setconfig('web', 'cacerts', '')
689 689
690 690 if options['version']:
691 691 return commands.version_(ui)
692 692 if options['help']:
693 693 return commands.help_(ui, cmd)
694 694 elif not cmd:
695 695 return commands.help_(ui, 'shortlist')
696 696
697 697 repo = None
698 698 cmdpats = args[:]
699 699 if cmd not in commands.norepo.split():
700 700 # use the repo from the request only if we don't have -R
701 701 if not rpath and not cwd:
702 702 repo = req.repo
703 703
704 704 if repo:
705 705 # set the descriptors of the repo ui to those of ui
706 706 repo.ui.fin = ui.fin
707 707 repo.ui.fout = ui.fout
708 708 repo.ui.ferr = ui.ferr
709 709 else:
710 710 try:
711 711 repo = hg.repository(ui, path=path)
712 712 if not repo.local():
713 713 raise util.Abort(_("repository '%s' is not local") % path)
714 714 if options['hidden']:
715 715 repo = repo.unfiltered()
716 716 repo.ui.setconfig("bundle", "mainreporoot", repo.root)
717 717 except error.RequirementError:
718 718 raise
719 719 except error.RepoError:
720 720 if cmd not in commands.optionalrepo.split():
721 721 if (cmd in commands.inferrepo.split() and
722 722 args and not path): # try to infer -R from command args
723 723 repos = map(cmdutil.findrepo, args)
724 724 guess = repos[0]
725 725 if guess and repos.count(guess) == len(repos):
726 726 req.args = ['--repository', guess] + fullargs
727 727 return _dispatch(req)
728 728 if not path:
729 729 raise error.RepoError(_("no repository found in '%s'"
730 730 " (.hg not found)")
731 731 % os.getcwd())
732 732 raise
733 733 if repo:
734 734 ui = repo.ui
735 735 args.insert(0, repo)
736 736 elif rpath:
737 737 ui.warn(_("warning: --repository ignored\n"))
738 738
739 739 msg = ' '.join(' ' in a and repr(a) or a for a in fullargs)
740 740 ui.log("command", msg + "\n")
741 741 d = lambda: util.checksignature(func)(ui, *args, **cmdoptions)
742 742 starttime = time.time()
743 743 ret = None
744 744 try:
745 745 ret = runcommand(lui, repo, cmd, fullargs, ui, options, d,
746 746 cmdpats, cmdoptions)
747 747 return ret
748 748 finally:
749 749 duration = time.time() - starttime
750 ui.log("commandfinish", _("%s exited %s after %0.2f seconds\n"),
750 ui.log("commandfinish", "%s exited %s after %0.2f seconds\n",
751 751 cmd, ret, duration)
752 752 if repo and repo != req.repo:
753 753 repo.close()
754 754
755 755 def lsprofile(ui, func, fp):
756 756 format = ui.config('profiling', 'format', default='text')
757 757 field = ui.config('profiling', 'sort', default='inlinetime')
758 758 limit = ui.configint('profiling', 'limit', default=30)
759 759 climit = ui.configint('profiling', 'nested', default=5)
760 760
761 761 if format not in ['text', 'kcachegrind']:
762 762 ui.warn(_("unrecognized profiling format '%s'"
763 763 " - Ignored\n") % format)
764 764 format = 'text'
765 765
766 766 try:
767 767 from mercurial import lsprof
768 768 except ImportError:
769 769 raise util.Abort(_(
770 770 'lsprof not available - install from '
771 771 'http://codespeak.net/svn/user/arigo/hack/misc/lsprof/'))
772 772 p = lsprof.Profiler()
773 773 p.enable(subcalls=True)
774 774 try:
775 775 return func()
776 776 finally:
777 777 p.disable()
778 778
779 779 if format == 'kcachegrind':
780 780 import lsprofcalltree
781 781 calltree = lsprofcalltree.KCacheGrind(p)
782 782 calltree.output(fp)
783 783 else:
784 784 # format == 'text'
785 785 stats = lsprof.Stats(p.getstats())
786 786 stats.sort(field)
787 787 stats.pprint(limit=limit, file=fp, climit=climit)
788 788
789 789 def statprofile(ui, func, fp):
790 790 try:
791 791 import statprof
792 792 except ImportError:
793 793 raise util.Abort(_(
794 794 'statprof not available - install using "easy_install statprof"'))
795 795
796 796 freq = ui.configint('profiling', 'freq', default=1000)
797 797 if freq > 0:
798 798 statprof.reset(freq)
799 799 else:
800 800 ui.warn(_("invalid sampling frequency '%s' - ignoring\n") % freq)
801 801
802 802 statprof.start()
803 803 try:
804 804 return func()
805 805 finally:
806 806 statprof.stop()
807 807 statprof.display(fp)
808 808
809 809 def _runcommand(ui, options, cmd, cmdfunc):
810 810 def checkargs():
811 811 try:
812 812 return cmdfunc()
813 813 except error.SignatureError:
814 814 raise error.CommandError(cmd, _("invalid arguments"))
815 815
816 816 if options['profile']:
817 817 profiler = os.getenv('HGPROF')
818 818 if profiler is None:
819 819 profiler = ui.config('profiling', 'type', default='ls')
820 820 if profiler not in ('ls', 'stat'):
821 821 ui.warn(_("unrecognized profiler '%s' - ignored\n") % profiler)
822 822 profiler = 'ls'
823 823
824 824 output = ui.config('profiling', 'output')
825 825
826 826 if output:
827 827 path = ui.expandpath(output)
828 828 fp = open(path, 'wb')
829 829 else:
830 830 fp = sys.stderr
831 831
832 832 try:
833 833 if profiler == 'ls':
834 834 return lsprofile(ui, checkargs, fp)
835 835 else:
836 836 return statprofile(ui, checkargs, fp)
837 837 finally:
838 838 if output:
839 839 fp.close()
840 840 else:
841 841 return checkargs()
@@ -1,206 +1,206
1 1 # hook.py - hook support for mercurial
2 2 #
3 3 # Copyright 2007 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 from i18n import _
9 9 import os, sys, time, types
10 10 import extensions, util, demandimport
11 11
12 12 def _pythonhook(ui, repo, name, hname, funcname, args, throw):
13 13 '''call python hook. hook is callable object, looked up as
14 14 name in python module. if callable returns "true", hook
15 15 fails, else passes. if hook raises exception, treated as
16 16 hook failure. exception propagates if throw is "true".
17 17
18 18 reason for "true" meaning "hook failed" is so that
19 19 unmodified commands (e.g. mercurial.commands.update) can
20 20 be run as hooks without wrappers to convert return values.'''
21 21
22 22 ui.note(_("calling hook %s: %s\n") % (hname, funcname))
23 23 starttime = time.time()
24 24
25 25 obj = funcname
26 26 if not util.safehasattr(obj, '__call__'):
27 27 d = funcname.rfind('.')
28 28 if d == -1:
29 29 raise util.Abort(_('%s hook is invalid ("%s" not in '
30 30 'a module)') % (hname, funcname))
31 31 modname = funcname[:d]
32 32 oldpaths = sys.path
33 33 if util.mainfrozen():
34 34 # binary installs require sys.path manipulation
35 35 modpath, modfile = os.path.split(modname)
36 36 if modpath and modfile:
37 37 sys.path = sys.path[:] + [modpath]
38 38 modname = modfile
39 39 try:
40 40 demandimport.disable()
41 41 obj = __import__(modname)
42 42 demandimport.enable()
43 43 except ImportError:
44 44 e1 = sys.exc_type, sys.exc_value, sys.exc_traceback
45 45 try:
46 46 # extensions are loaded with hgext_ prefix
47 47 obj = __import__("hgext_%s" % modname)
48 48 demandimport.enable()
49 49 except ImportError:
50 50 demandimport.enable()
51 51 e2 = sys.exc_type, sys.exc_value, sys.exc_traceback
52 52 if ui.tracebackflag:
53 53 ui.warn(_('exception from first failed import attempt:\n'))
54 54 ui.traceback(e1)
55 55 if ui.tracebackflag:
56 56 ui.warn(_('exception from second failed import attempt:\n'))
57 57 ui.traceback(e2)
58 58 raise util.Abort(_('%s hook is invalid '
59 59 '(import of "%s" failed)') %
60 60 (hname, modname))
61 61 sys.path = oldpaths
62 62 try:
63 63 for p in funcname.split('.')[1:]:
64 64 obj = getattr(obj, p)
65 65 except AttributeError:
66 66 raise util.Abort(_('%s hook is invalid '
67 67 '("%s" is not defined)') %
68 68 (hname, funcname))
69 69 if not util.safehasattr(obj, '__call__'):
70 70 raise util.Abort(_('%s hook is invalid '
71 71 '("%s" is not callable)') %
72 72 (hname, funcname))
73 73 try:
74 74 try:
75 75 # redirect IO descriptors to the ui descriptors so hooks
76 76 # that write directly to these don't mess up the command
77 77 # protocol when running through the command server
78 78 old = sys.stdout, sys.stderr, sys.stdin
79 79 sys.stdout, sys.stderr, sys.stdin = ui.fout, ui.ferr, ui.fin
80 80
81 81 r = obj(ui=ui, repo=repo, hooktype=name, **args)
82 82 except KeyboardInterrupt:
83 83 raise
84 84 except Exception, exc:
85 85 if isinstance(exc, util.Abort):
86 86 ui.warn(_('error: %s hook failed: %s\n') %
87 87 (hname, exc.args[0]))
88 88 else:
89 89 ui.warn(_('error: %s hook raised an exception: '
90 90 '%s\n') % (hname, exc))
91 91 if throw:
92 92 raise
93 93 ui.traceback()
94 94 return True
95 95 finally:
96 96 sys.stdout, sys.stderr, sys.stdin = old
97 97 duration = time.time() - starttime
98 98 readablefunc = funcname
99 99 if isinstance(funcname, types.FunctionType):
100 100 readablefunc = funcname.__module__ + "." + funcname.__name__
101 ui.log('pythonhook', _('pythonhook-%s: %s finished in %0.2f seconds\n'),
101 ui.log('pythonhook', 'pythonhook-%s: %s finished in %0.2f seconds\n',
102 102 name, readablefunc, duration)
103 103 if r:
104 104 if throw:
105 105 raise util.Abort(_('%s hook failed') % hname)
106 106 ui.warn(_('warning: %s hook failed\n') % hname)
107 107 return r
108 108
109 109 def _exthook(ui, repo, name, cmd, args, throw):
110 110 ui.note(_("running hook %s: %s\n") % (name, cmd))
111 111
112 112 starttime = time.time()
113 113 env = {}
114 114 for k, v in args.iteritems():
115 115 if util.safehasattr(v, '__call__'):
116 116 v = v()
117 117 if isinstance(v, dict):
118 118 # make the dictionary element order stable across Python
119 119 # implementations
120 120 v = ('{' +
121 121 ', '.join('%r: %r' % i for i in sorted(v.iteritems())) +
122 122 '}')
123 123 env['HG_' + k.upper()] = v
124 124
125 125 if repo:
126 126 cwd = repo.root
127 127 else:
128 128 cwd = os.getcwd()
129 129 if 'HG_URL' in env and env['HG_URL'].startswith('remote:http'):
130 130 r = util.system(cmd, environ=env, cwd=cwd, out=ui)
131 131 else:
132 132 r = util.system(cmd, environ=env, cwd=cwd, out=ui.fout)
133 133
134 134 duration = time.time() - starttime
135 ui.log('exthook', _('exthook-%s: %s finished in %0.2f seconds\n'),
135 ui.log('exthook', 'exthook-%s: %s finished in %0.2f seconds\n',
136 136 name, cmd, duration)
137 137 if r:
138 138 desc, r = util.explainexit(r)
139 139 if throw:
140 140 raise util.Abort(_('%s hook %s') % (name, desc))
141 141 ui.warn(_('warning: %s hook %s\n') % (name, desc))
142 142 return r
143 143
144 144 def _allhooks(ui):
145 145 hooks = []
146 146 for name, cmd in ui.configitems('hooks'):
147 147 if not name.startswith('priority'):
148 148 priority = ui.configint('hooks', 'priority.%s' % name, 0)
149 149 hooks.append((-priority, len(hooks), name, cmd))
150 150 return [(k, v) for p, o, k, v in sorted(hooks)]
151 151
152 152 _redirect = False
153 153 def redirect(state):
154 154 global _redirect
155 155 _redirect = state
156 156
157 157 def hook(ui, repo, name, throw=False, **args):
158 158 if not ui.callhooks:
159 159 return False
160 160
161 161 r = False
162 162 oldstdout = -1
163 163
164 164 try:
165 165 for hname, cmd in _allhooks(ui):
166 166 if hname.split('.')[0] != name or not cmd:
167 167 continue
168 168
169 169 if oldstdout == -1 and _redirect:
170 170 try:
171 171 stdoutno = sys.__stdout__.fileno()
172 172 stderrno = sys.__stderr__.fileno()
173 173 # temporarily redirect stdout to stderr, if possible
174 174 if stdoutno >= 0 and stderrno >= 0:
175 175 sys.__stdout__.flush()
176 176 oldstdout = os.dup(stdoutno)
177 177 os.dup2(stderrno, stdoutno)
178 178 except (OSError, AttributeError):
179 179 # files seem to be bogus, give up on redirecting (WSGI, etc)
180 180 pass
181 181
182 182 if util.safehasattr(cmd, '__call__'):
183 183 r = _pythonhook(ui, repo, name, hname, cmd, args, throw) or r
184 184 elif cmd.startswith('python:'):
185 185 if cmd.count(':') >= 2:
186 186 path, cmd = cmd[7:].rsplit(':', 1)
187 187 path = util.expandpath(path)
188 188 if repo:
189 189 path = os.path.join(repo.root, path)
190 190 try:
191 191 mod = extensions.loadpath(path, 'hghook.%s' % hname)
192 192 except Exception:
193 193 ui.write(_("loading %s hook failed:\n") % hname)
194 194 raise
195 195 hookfn = getattr(mod, cmd)
196 196 else:
197 197 hookfn = cmd[7:].strip()
198 198 r = _pythonhook(ui, repo, name, hname, hookfn, args, throw) or r
199 199 else:
200 200 r = _exthook(ui, repo, hname, cmd, args, throw) or r
201 201 finally:
202 202 if _redirect and oldstdout >= 0:
203 203 os.dup2(oldstdout, stdoutno)
204 204 os.close(oldstdout)
205 205
206 206 return r
@@ -1,2591 +1,2591
1 1 # localrepo.py - read/write repository class for mercurial
2 2 #
3 3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7 from node import hex, nullid, short
8 8 from i18n import _
9 9 import peer, changegroup, subrepo, discovery, pushkey, obsolete, repoview
10 10 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
11 11 import lock, transaction, store, encoding, base85
12 12 import scmutil, util, extensions, hook, error, revset
13 13 import match as matchmod
14 14 import merge as mergemod
15 15 import tags as tagsmod
16 16 from lock import release
17 17 import weakref, errno, os, time, inspect
18 18 import branchmap
19 19 propertycache = util.propertycache
20 20 filecache = scmutil.filecache
21 21
22 22 class repofilecache(filecache):
23 23 """All filecache usage on repo are done for logic that should be unfiltered
24 24 """
25 25
26 26 def __get__(self, repo, type=None):
27 27 return super(repofilecache, self).__get__(repo.unfiltered(), type)
28 28 def __set__(self, repo, value):
29 29 return super(repofilecache, self).__set__(repo.unfiltered(), value)
30 30 def __delete__(self, repo):
31 31 return super(repofilecache, self).__delete__(repo.unfiltered())
32 32
33 33 class storecache(repofilecache):
34 34 """filecache for files in the store"""
35 35 def join(self, obj, fname):
36 36 return obj.sjoin(fname)
37 37
38 38 class unfilteredpropertycache(propertycache):
39 39 """propertycache that apply to unfiltered repo only"""
40 40
41 41 def __get__(self, repo, type=None):
42 42 return super(unfilteredpropertycache, self).__get__(repo.unfiltered())
43 43
44 44 class filteredpropertycache(propertycache):
45 45 """propertycache that must take filtering in account"""
46 46
47 47 def cachevalue(self, obj, value):
48 48 object.__setattr__(obj, self.name, value)
49 49
50 50
51 51 def hasunfilteredcache(repo, name):
52 52 """check if a repo has an unfilteredpropertycache value for <name>"""
53 53 return name in vars(repo.unfiltered())
54 54
55 55 def unfilteredmethod(orig):
56 56 """decorate method that always need to be run on unfiltered version"""
57 57 def wrapper(repo, *args, **kwargs):
58 58 return orig(repo.unfiltered(), *args, **kwargs)
59 59 return wrapper
60 60
61 61 MODERNCAPS = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle'))
62 62 LEGACYCAPS = MODERNCAPS.union(set(['changegroupsubset']))
63 63
64 64 class localpeer(peer.peerrepository):
65 65 '''peer for a local repo; reflects only the most recent API'''
66 66
67 67 def __init__(self, repo, caps=MODERNCAPS):
68 68 peer.peerrepository.__init__(self)
69 69 self._repo = repo.filtered('served')
70 70 self.ui = repo.ui
71 71 self._caps = repo._restrictcapabilities(caps)
72 72 self.requirements = repo.requirements
73 73 self.supportedformats = repo.supportedformats
74 74
75 75 def close(self):
76 76 self._repo.close()
77 77
78 78 def _capabilities(self):
79 79 return self._caps
80 80
81 81 def local(self):
82 82 return self._repo
83 83
84 84 def canpush(self):
85 85 return True
86 86
87 87 def url(self):
88 88 return self._repo.url()
89 89
90 90 def lookup(self, key):
91 91 return self._repo.lookup(key)
92 92
93 93 def branchmap(self):
94 94 return self._repo.branchmap()
95 95
96 96 def heads(self):
97 97 return self._repo.heads()
98 98
99 99 def known(self, nodes):
100 100 return self._repo.known(nodes)
101 101
102 102 def getbundle(self, source, heads=None, common=None):
103 103 return self._repo.getbundle(source, heads=heads, common=common)
104 104
105 105 # TODO We might want to move the next two calls into legacypeer and add
106 106 # unbundle instead.
107 107
108 108 def lock(self):
109 109 return self._repo.lock()
110 110
111 111 def addchangegroup(self, cg, source, url):
112 112 return self._repo.addchangegroup(cg, source, url)
113 113
114 114 def pushkey(self, namespace, key, old, new):
115 115 return self._repo.pushkey(namespace, key, old, new)
116 116
117 117 def listkeys(self, namespace):
118 118 return self._repo.listkeys(namespace)
119 119
120 120 def debugwireargs(self, one, two, three=None, four=None, five=None):
121 121 '''used to test argument passing over the wire'''
122 122 return "%s %s %s %s %s" % (one, two, three, four, five)
123 123
124 124 class locallegacypeer(localpeer):
125 125 '''peer extension which implements legacy methods too; used for tests with
126 126 restricted capabilities'''
127 127
128 128 def __init__(self, repo):
129 129 localpeer.__init__(self, repo, caps=LEGACYCAPS)
130 130
131 131 def branches(self, nodes):
132 132 return self._repo.branches(nodes)
133 133
134 134 def between(self, pairs):
135 135 return self._repo.between(pairs)
136 136
137 137 def changegroup(self, basenodes, source):
138 138 return self._repo.changegroup(basenodes, source)
139 139
140 140 def changegroupsubset(self, bases, heads, source):
141 141 return self._repo.changegroupsubset(bases, heads, source)
142 142
143 143 class localrepository(object):
144 144
145 145 supportedformats = set(('revlogv1', 'generaldelta'))
146 146 supported = supportedformats | set(('store', 'fncache', 'shared',
147 147 'dotencode'))
148 148 openerreqs = set(('revlogv1', 'generaldelta'))
149 149 requirements = ['revlogv1']
150 150 filtername = None
151 151
152 152 def _baserequirements(self, create):
153 153 return self.requirements[:]
154 154
155 155 def __init__(self, baseui, path=None, create=False):
156 156 self.wvfs = scmutil.vfs(path, expand=True)
157 157 self.wopener = self.wvfs
158 158 self.root = self.wvfs.base
159 159 self.path = self.wvfs.join(".hg")
160 160 self.origroot = path
161 161 self.auditor = scmutil.pathauditor(self.root, self._checknested)
162 162 self.vfs = scmutil.vfs(self.path)
163 163 self.opener = self.vfs
164 164 self.baseui = baseui
165 165 self.ui = baseui.copy()
166 166 # A list of callback to shape the phase if no data were found.
167 167 # Callback are in the form: func(repo, roots) --> processed root.
168 168 # This list it to be filled by extension during repo setup
169 169 self._phasedefaults = []
170 170 try:
171 171 self.ui.readconfig(self.join("hgrc"), self.root)
172 172 extensions.loadall(self.ui)
173 173 except IOError:
174 174 pass
175 175
176 176 if not self.vfs.isdir():
177 177 if create:
178 178 if not self.wvfs.exists():
179 179 self.wvfs.makedirs()
180 180 self.vfs.makedir(notindexed=True)
181 181 requirements = self._baserequirements(create)
182 182 if self.ui.configbool('format', 'usestore', True):
183 183 self.vfs.mkdir("store")
184 184 requirements.append("store")
185 185 if self.ui.configbool('format', 'usefncache', True):
186 186 requirements.append("fncache")
187 187 if self.ui.configbool('format', 'dotencode', True):
188 188 requirements.append('dotencode')
189 189 # create an invalid changelog
190 190 self.vfs.append(
191 191 "00changelog.i",
192 192 '\0\0\0\2' # represents revlogv2
193 193 ' dummy changelog to prevent using the old repo layout'
194 194 )
195 195 if self.ui.configbool('format', 'generaldelta', False):
196 196 requirements.append("generaldelta")
197 197 requirements = set(requirements)
198 198 else:
199 199 raise error.RepoError(_("repository %s not found") % path)
200 200 elif create:
201 201 raise error.RepoError(_("repository %s already exists") % path)
202 202 else:
203 203 try:
204 204 requirements = scmutil.readrequires(self.vfs, self.supported)
205 205 except IOError, inst:
206 206 if inst.errno != errno.ENOENT:
207 207 raise
208 208 requirements = set()
209 209
210 210 self.sharedpath = self.path
211 211 try:
212 212 s = os.path.realpath(self.opener.read("sharedpath").rstrip('\n'))
213 213 if not os.path.exists(s):
214 214 raise error.RepoError(
215 215 _('.hg/sharedpath points to nonexistent directory %s') % s)
216 216 self.sharedpath = s
217 217 except IOError, inst:
218 218 if inst.errno != errno.ENOENT:
219 219 raise
220 220
221 221 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
222 222 self.spath = self.store.path
223 223 self.svfs = self.store.vfs
224 224 self.sopener = self.svfs
225 225 self.sjoin = self.store.join
226 226 self.vfs.createmode = self.store.createmode
227 227 self._applyrequirements(requirements)
228 228 if create:
229 229 self._writerequirements()
230 230
231 231
232 232 self._branchcaches = {}
233 233 self.filterpats = {}
234 234 self._datafilters = {}
235 235 self._transref = self._lockref = self._wlockref = None
236 236
237 237 # A cache for various files under .hg/ that tracks file changes,
238 238 # (used by the filecache decorator)
239 239 #
240 240 # Maps a property name to its util.filecacheentry
241 241 self._filecache = {}
242 242
243 243 # hold sets of revision to be filtered
244 244 # should be cleared when something might have changed the filter value:
245 245 # - new changesets,
246 246 # - phase change,
247 247 # - new obsolescence marker,
248 248 # - working directory parent change,
249 249 # - bookmark changes
250 250 self.filteredrevcache = {}
251 251
252 252 def close(self):
253 253 pass
254 254
255 255 def _restrictcapabilities(self, caps):
256 256 return caps
257 257
258 258 def _applyrequirements(self, requirements):
259 259 self.requirements = requirements
260 260 self.sopener.options = dict((r, 1) for r in requirements
261 261 if r in self.openerreqs)
262 262
263 263 def _writerequirements(self):
264 264 reqfile = self.opener("requires", "w")
265 265 for r in sorted(self.requirements):
266 266 reqfile.write("%s\n" % r)
267 267 reqfile.close()
268 268
269 269 def _checknested(self, path):
270 270 """Determine if path is a legal nested repository."""
271 271 if not path.startswith(self.root):
272 272 return False
273 273 subpath = path[len(self.root) + 1:]
274 274 normsubpath = util.pconvert(subpath)
275 275
276 276 # XXX: Checking against the current working copy is wrong in
277 277 # the sense that it can reject things like
278 278 #
279 279 # $ hg cat -r 10 sub/x.txt
280 280 #
281 281 # if sub/ is no longer a subrepository in the working copy
282 282 # parent revision.
283 283 #
284 284 # However, it can of course also allow things that would have
285 285 # been rejected before, such as the above cat command if sub/
286 286 # is a subrepository now, but was a normal directory before.
287 287 # The old path auditor would have rejected by mistake since it
288 288 # panics when it sees sub/.hg/.
289 289 #
290 290 # All in all, checking against the working copy seems sensible
291 291 # since we want to prevent access to nested repositories on
292 292 # the filesystem *now*.
293 293 ctx = self[None]
294 294 parts = util.splitpath(subpath)
295 295 while parts:
296 296 prefix = '/'.join(parts)
297 297 if prefix in ctx.substate:
298 298 if prefix == normsubpath:
299 299 return True
300 300 else:
301 301 sub = ctx.sub(prefix)
302 302 return sub.checknested(subpath[len(prefix) + 1:])
303 303 else:
304 304 parts.pop()
305 305 return False
306 306
307 307 def peer(self):
308 308 return localpeer(self) # not cached to avoid reference cycle
309 309
310 310 def unfiltered(self):
311 311 """Return unfiltered version of the repository
312 312
313 313 Intended to be overwritten by filtered repo."""
314 314 return self
315 315
316 316 def filtered(self, name):
317 317 """Return a filtered version of a repository"""
318 318 # build a new class with the mixin and the current class
319 319 # (possibly subclass of the repo)
320 320 class proxycls(repoview.repoview, self.unfiltered().__class__):
321 321 pass
322 322 return proxycls(self, name)
323 323
324 324 @repofilecache('bookmarks')
325 325 def _bookmarks(self):
326 326 return bookmarks.bmstore(self)
327 327
328 328 @repofilecache('bookmarks.current')
329 329 def _bookmarkcurrent(self):
330 330 return bookmarks.readcurrent(self)
331 331
332 332 def bookmarkheads(self, bookmark):
333 333 name = bookmark.split('@', 1)[0]
334 334 heads = []
335 335 for mark, n in self._bookmarks.iteritems():
336 336 if mark.split('@', 1)[0] == name:
337 337 heads.append(n)
338 338 return heads
339 339
340 340 @storecache('phaseroots')
341 341 def _phasecache(self):
342 342 return phases.phasecache(self, self._phasedefaults)
343 343
344 344 @storecache('obsstore')
345 345 def obsstore(self):
346 346 store = obsolete.obsstore(self.sopener)
347 347 if store and not obsolete._enabled:
348 348 # message is rare enough to not be translated
349 349 msg = 'obsolete feature not enabled but %i markers found!\n'
350 350 self.ui.warn(msg % len(list(store)))
351 351 return store
352 352
353 353 @storecache('00changelog.i')
354 354 def changelog(self):
355 355 c = changelog.changelog(self.sopener)
356 356 if 'HG_PENDING' in os.environ:
357 357 p = os.environ['HG_PENDING']
358 358 if p.startswith(self.root):
359 359 c.readpending('00changelog.i.a')
360 360 return c
361 361
362 362 @storecache('00manifest.i')
363 363 def manifest(self):
364 364 return manifest.manifest(self.sopener)
365 365
366 366 @repofilecache('dirstate')
367 367 def dirstate(self):
368 368 warned = [0]
369 369 def validate(node):
370 370 try:
371 371 self.changelog.rev(node)
372 372 return node
373 373 except error.LookupError:
374 374 if not warned[0]:
375 375 warned[0] = True
376 376 self.ui.warn(_("warning: ignoring unknown"
377 377 " working parent %s!\n") % short(node))
378 378 return nullid
379 379
380 380 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
381 381
382 382 def __getitem__(self, changeid):
383 383 if changeid is None:
384 384 return context.workingctx(self)
385 385 return context.changectx(self, changeid)
386 386
387 387 def __contains__(self, changeid):
388 388 try:
389 389 return bool(self.lookup(changeid))
390 390 except error.RepoLookupError:
391 391 return False
392 392
393 393 def __nonzero__(self):
394 394 return True
395 395
396 396 def __len__(self):
397 397 return len(self.changelog)
398 398
399 399 def __iter__(self):
400 400 return iter(self.changelog)
401 401
402 402 def revs(self, expr, *args):
403 403 '''Return a list of revisions matching the given revset'''
404 404 expr = revset.formatspec(expr, *args)
405 405 m = revset.match(None, expr)
406 406 return [r for r in m(self, list(self))]
407 407
408 408 def set(self, expr, *args):
409 409 '''
410 410 Yield a context for each matching revision, after doing arg
411 411 replacement via revset.formatspec
412 412 '''
413 413 for r in self.revs(expr, *args):
414 414 yield self[r]
415 415
416 416 def url(self):
417 417 return 'file:' + self.root
418 418
419 419 def hook(self, name, throw=False, **args):
420 420 return hook.hook(self.ui, self, name, throw, **args)
421 421
422 422 @unfilteredmethod
423 423 def _tag(self, names, node, message, local, user, date, extra={}):
424 424 if isinstance(names, str):
425 425 names = (names,)
426 426
427 427 branches = self.branchmap()
428 428 for name in names:
429 429 self.hook('pretag', throw=True, node=hex(node), tag=name,
430 430 local=local)
431 431 if name in branches:
432 432 self.ui.warn(_("warning: tag %s conflicts with existing"
433 433 " branch name\n") % name)
434 434
435 435 def writetags(fp, names, munge, prevtags):
436 436 fp.seek(0, 2)
437 437 if prevtags and prevtags[-1] != '\n':
438 438 fp.write('\n')
439 439 for name in names:
440 440 m = munge and munge(name) or name
441 441 if (self._tagscache.tagtypes and
442 442 name in self._tagscache.tagtypes):
443 443 old = self.tags().get(name, nullid)
444 444 fp.write('%s %s\n' % (hex(old), m))
445 445 fp.write('%s %s\n' % (hex(node), m))
446 446 fp.close()
447 447
448 448 prevtags = ''
449 449 if local:
450 450 try:
451 451 fp = self.opener('localtags', 'r+')
452 452 except IOError:
453 453 fp = self.opener('localtags', 'a')
454 454 else:
455 455 prevtags = fp.read()
456 456
457 457 # local tags are stored in the current charset
458 458 writetags(fp, names, None, prevtags)
459 459 for name in names:
460 460 self.hook('tag', node=hex(node), tag=name, local=local)
461 461 return
462 462
463 463 try:
464 464 fp = self.wfile('.hgtags', 'rb+')
465 465 except IOError, e:
466 466 if e.errno != errno.ENOENT:
467 467 raise
468 468 fp = self.wfile('.hgtags', 'ab')
469 469 else:
470 470 prevtags = fp.read()
471 471
472 472 # committed tags are stored in UTF-8
473 473 writetags(fp, names, encoding.fromlocal, prevtags)
474 474
475 475 fp.close()
476 476
477 477 self.invalidatecaches()
478 478
479 479 if '.hgtags' not in self.dirstate:
480 480 self[None].add(['.hgtags'])
481 481
482 482 m = matchmod.exact(self.root, '', ['.hgtags'])
483 483 tagnode = self.commit(message, user, date, extra=extra, match=m)
484 484
485 485 for name in names:
486 486 self.hook('tag', node=hex(node), tag=name, local=local)
487 487
488 488 return tagnode
489 489
490 490 def tag(self, names, node, message, local, user, date):
491 491 '''tag a revision with one or more symbolic names.
492 492
493 493 names is a list of strings or, when adding a single tag, names may be a
494 494 string.
495 495
496 496 if local is True, the tags are stored in a per-repository file.
497 497 otherwise, they are stored in the .hgtags file, and a new
498 498 changeset is committed with the change.
499 499
500 500 keyword arguments:
501 501
502 502 local: whether to store tags in non-version-controlled file
503 503 (default False)
504 504
505 505 message: commit message to use if committing
506 506
507 507 user: name of user to use if committing
508 508
509 509 date: date tuple to use if committing'''
510 510
511 511 if not local:
512 512 for x in self.status()[:5]:
513 513 if '.hgtags' in x:
514 514 raise util.Abort(_('working copy of .hgtags is changed '
515 515 '(please commit .hgtags manually)'))
516 516
517 517 self.tags() # instantiate the cache
518 518 self._tag(names, node, message, local, user, date)
519 519
520 520 @filteredpropertycache
521 521 def _tagscache(self):
522 522 '''Returns a tagscache object that contains various tags related
523 523 caches.'''
524 524
525 525 # This simplifies its cache management by having one decorated
526 526 # function (this one) and the rest simply fetch things from it.
527 527 class tagscache(object):
528 528 def __init__(self):
529 529 # These two define the set of tags for this repository. tags
530 530 # maps tag name to node; tagtypes maps tag name to 'global' or
531 531 # 'local'. (Global tags are defined by .hgtags across all
532 532 # heads, and local tags are defined in .hg/localtags.)
533 533 # They constitute the in-memory cache of tags.
534 534 self.tags = self.tagtypes = None
535 535
536 536 self.nodetagscache = self.tagslist = None
537 537
538 538 cache = tagscache()
539 539 cache.tags, cache.tagtypes = self._findtags()
540 540
541 541 return cache
542 542
543 543 def tags(self):
544 544 '''return a mapping of tag to node'''
545 545 t = {}
546 546 if self.changelog.filteredrevs:
547 547 tags, tt = self._findtags()
548 548 else:
549 549 tags = self._tagscache.tags
550 550 for k, v in tags.iteritems():
551 551 try:
552 552 # ignore tags to unknown nodes
553 553 self.changelog.rev(v)
554 554 t[k] = v
555 555 except (error.LookupError, ValueError):
556 556 pass
557 557 return t
558 558
559 559 def _findtags(self):
560 560 '''Do the hard work of finding tags. Return a pair of dicts
561 561 (tags, tagtypes) where tags maps tag name to node, and tagtypes
562 562 maps tag name to a string like \'global\' or \'local\'.
563 563 Subclasses or extensions are free to add their own tags, but
564 564 should be aware that the returned dicts will be retained for the
565 565 duration of the localrepo object.'''
566 566
567 567 # XXX what tagtype should subclasses/extensions use? Currently
568 568 # mq and bookmarks add tags, but do not set the tagtype at all.
569 569 # Should each extension invent its own tag type? Should there
570 570 # be one tagtype for all such "virtual" tags? Or is the status
571 571 # quo fine?
572 572
573 573 alltags = {} # map tag name to (node, hist)
574 574 tagtypes = {}
575 575
576 576 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
577 577 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
578 578
579 579 # Build the return dicts. Have to re-encode tag names because
580 580 # the tags module always uses UTF-8 (in order not to lose info
581 581 # writing to the cache), but the rest of Mercurial wants them in
582 582 # local encoding.
583 583 tags = {}
584 584 for (name, (node, hist)) in alltags.iteritems():
585 585 if node != nullid:
586 586 tags[encoding.tolocal(name)] = node
587 587 tags['tip'] = self.changelog.tip()
588 588 tagtypes = dict([(encoding.tolocal(name), value)
589 589 for (name, value) in tagtypes.iteritems()])
590 590 return (tags, tagtypes)
591 591
592 592 def tagtype(self, tagname):
593 593 '''
594 594 return the type of the given tag. result can be:
595 595
596 596 'local' : a local tag
597 597 'global' : a global tag
598 598 None : tag does not exist
599 599 '''
600 600
601 601 return self._tagscache.tagtypes.get(tagname)
602 602
603 603 def tagslist(self):
604 604 '''return a list of tags ordered by revision'''
605 605 if not self._tagscache.tagslist:
606 606 l = []
607 607 for t, n in self.tags().iteritems():
608 608 r = self.changelog.rev(n)
609 609 l.append((r, t, n))
610 610 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
611 611
612 612 return self._tagscache.tagslist
613 613
614 614 def nodetags(self, node):
615 615 '''return the tags associated with a node'''
616 616 if not self._tagscache.nodetagscache:
617 617 nodetagscache = {}
618 618 for t, n in self._tagscache.tags.iteritems():
619 619 nodetagscache.setdefault(n, []).append(t)
620 620 for tags in nodetagscache.itervalues():
621 621 tags.sort()
622 622 self._tagscache.nodetagscache = nodetagscache
623 623 return self._tagscache.nodetagscache.get(node, [])
624 624
625 625 def nodebookmarks(self, node):
626 626 marks = []
627 627 for bookmark, n in self._bookmarks.iteritems():
628 628 if n == node:
629 629 marks.append(bookmark)
630 630 return sorted(marks)
631 631
632 632 def branchmap(self):
633 633 '''returns a dictionary {branch: [branchheads]}'''
634 634 branchmap.updatecache(self)
635 635 return self._branchcaches[self.filtername]
636 636
637 637
638 638 def _branchtip(self, heads):
639 639 '''return the tipmost branch head in heads'''
640 640 tip = heads[-1]
641 641 for h in reversed(heads):
642 642 if not self[h].closesbranch():
643 643 tip = h
644 644 break
645 645 return tip
646 646
647 647 def branchtip(self, branch):
648 648 '''return the tip node for a given branch'''
649 649 if branch not in self.branchmap():
650 650 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
651 651 return self._branchtip(self.branchmap()[branch])
652 652
653 653 def branchtags(self):
654 654 '''return a dict where branch names map to the tipmost head of
655 655 the branch, open heads come before closed'''
656 656 bt = {}
657 657 for bn, heads in self.branchmap().iteritems():
658 658 bt[bn] = self._branchtip(heads)
659 659 return bt
660 660
661 661 def lookup(self, key):
662 662 return self[key].node()
663 663
664 664 def lookupbranch(self, key, remote=None):
665 665 repo = remote or self
666 666 if key in repo.branchmap():
667 667 return key
668 668
669 669 repo = (remote and remote.local()) and remote or self
670 670 return repo[key].branch()
671 671
672 672 def known(self, nodes):
673 673 nm = self.changelog.nodemap
674 674 pc = self._phasecache
675 675 result = []
676 676 for n in nodes:
677 677 r = nm.get(n)
678 678 resp = not (r is None or pc.phase(self, r) >= phases.secret)
679 679 result.append(resp)
680 680 return result
681 681
682 682 def local(self):
683 683 return self
684 684
685 685 def cancopy(self):
686 686 return self.local() # so statichttprepo's override of local() works
687 687
688 688 def join(self, f):
689 689 return os.path.join(self.path, f)
690 690
691 691 def wjoin(self, f):
692 692 return os.path.join(self.root, f)
693 693
694 694 def file(self, f):
695 695 if f[0] == '/':
696 696 f = f[1:]
697 697 return filelog.filelog(self.sopener, f)
698 698
699 699 def changectx(self, changeid):
700 700 return self[changeid]
701 701
702 702 def parents(self, changeid=None):
703 703 '''get list of changectxs for parents of changeid'''
704 704 return self[changeid].parents()
705 705
706 706 def setparents(self, p1, p2=nullid):
707 707 copies = self.dirstate.setparents(p1, p2)
708 708 if copies:
709 709 # Adjust copy records, the dirstate cannot do it, it
710 710 # requires access to parents manifests. Preserve them
711 711 # only for entries added to first parent.
712 712 pctx = self[p1]
713 713 for f in copies:
714 714 if f not in pctx and copies[f] in pctx:
715 715 self.dirstate.copy(copies[f], f)
716 716
717 717 def filectx(self, path, changeid=None, fileid=None):
718 718 """changeid can be a changeset revision, node, or tag.
719 719 fileid can be a file revision or node."""
720 720 return context.filectx(self, path, changeid, fileid)
721 721
722 722 def getcwd(self):
723 723 return self.dirstate.getcwd()
724 724
725 725 def pathto(self, f, cwd=None):
726 726 return self.dirstate.pathto(f, cwd)
727 727
728 728 def wfile(self, f, mode='r'):
729 729 return self.wopener(f, mode)
730 730
731 731 def _link(self, f):
732 732 return os.path.islink(self.wjoin(f))
733 733
734 734 def _loadfilter(self, filter):
735 735 if filter not in self.filterpats:
736 736 l = []
737 737 for pat, cmd in self.ui.configitems(filter):
738 738 if cmd == '!':
739 739 continue
740 740 mf = matchmod.match(self.root, '', [pat])
741 741 fn = None
742 742 params = cmd
743 743 for name, filterfn in self._datafilters.iteritems():
744 744 if cmd.startswith(name):
745 745 fn = filterfn
746 746 params = cmd[len(name):].lstrip()
747 747 break
748 748 if not fn:
749 749 fn = lambda s, c, **kwargs: util.filter(s, c)
750 750 # Wrap old filters not supporting keyword arguments
751 751 if not inspect.getargspec(fn)[2]:
752 752 oldfn = fn
753 753 fn = lambda s, c, **kwargs: oldfn(s, c)
754 754 l.append((mf, fn, params))
755 755 self.filterpats[filter] = l
756 756 return self.filterpats[filter]
757 757
758 758 def _filter(self, filterpats, filename, data):
759 759 for mf, fn, cmd in filterpats:
760 760 if mf(filename):
761 761 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
762 762 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
763 763 break
764 764
765 765 return data
766 766
767 767 @unfilteredpropertycache
768 768 def _encodefilterpats(self):
769 769 return self._loadfilter('encode')
770 770
771 771 @unfilteredpropertycache
772 772 def _decodefilterpats(self):
773 773 return self._loadfilter('decode')
774 774
775 775 def adddatafilter(self, name, filter):
776 776 self._datafilters[name] = filter
777 777
778 778 def wread(self, filename):
779 779 if self._link(filename):
780 780 data = os.readlink(self.wjoin(filename))
781 781 else:
782 782 data = self.wopener.read(filename)
783 783 return self._filter(self._encodefilterpats, filename, data)
784 784
785 785 def wwrite(self, filename, data, flags):
786 786 data = self._filter(self._decodefilterpats, filename, data)
787 787 if 'l' in flags:
788 788 self.wopener.symlink(data, filename)
789 789 else:
790 790 self.wopener.write(filename, data)
791 791 if 'x' in flags:
792 792 util.setflags(self.wjoin(filename), False, True)
793 793
794 794 def wwritedata(self, filename, data):
795 795 return self._filter(self._decodefilterpats, filename, data)
796 796
797 797 def transaction(self, desc):
798 798 tr = self._transref and self._transref() or None
799 799 if tr and tr.running():
800 800 return tr.nest()
801 801
802 802 # abort here if the journal already exists
803 803 if os.path.exists(self.sjoin("journal")):
804 804 raise error.RepoError(
805 805 _("abandoned transaction found - run hg recover"))
806 806
807 807 self._writejournal(desc)
808 808 renames = [(x, undoname(x)) for x in self._journalfiles()]
809 809
810 810 tr = transaction.transaction(self.ui.warn, self.sopener,
811 811 self.sjoin("journal"),
812 812 aftertrans(renames),
813 813 self.store.createmode)
814 814 self._transref = weakref.ref(tr)
815 815 return tr
816 816
817 817 def _journalfiles(self):
818 818 return (self.sjoin('journal'), self.join('journal.dirstate'),
819 819 self.join('journal.branch'), self.join('journal.desc'),
820 820 self.join('journal.bookmarks'),
821 821 self.sjoin('journal.phaseroots'))
822 822
823 823 def undofiles(self):
824 824 return [undoname(x) for x in self._journalfiles()]
825 825
826 826 def _writejournal(self, desc):
827 827 self.opener.write("journal.dirstate",
828 828 self.opener.tryread("dirstate"))
829 829 self.opener.write("journal.branch",
830 830 encoding.fromlocal(self.dirstate.branch()))
831 831 self.opener.write("journal.desc",
832 832 "%d\n%s\n" % (len(self), desc))
833 833 self.opener.write("journal.bookmarks",
834 834 self.opener.tryread("bookmarks"))
835 835 self.sopener.write("journal.phaseroots",
836 836 self.sopener.tryread("phaseroots"))
837 837
838 838 def recover(self):
839 839 lock = self.lock()
840 840 try:
841 841 if os.path.exists(self.sjoin("journal")):
842 842 self.ui.status(_("rolling back interrupted transaction\n"))
843 843 transaction.rollback(self.sopener, self.sjoin("journal"),
844 844 self.ui.warn)
845 845 self.invalidate()
846 846 return True
847 847 else:
848 848 self.ui.warn(_("no interrupted transaction available\n"))
849 849 return False
850 850 finally:
851 851 lock.release()
852 852
853 853 def rollback(self, dryrun=False, force=False):
854 854 wlock = lock = None
855 855 try:
856 856 wlock = self.wlock()
857 857 lock = self.lock()
858 858 if os.path.exists(self.sjoin("undo")):
859 859 return self._rollback(dryrun, force)
860 860 else:
861 861 self.ui.warn(_("no rollback information available\n"))
862 862 return 1
863 863 finally:
864 864 release(lock, wlock)
865 865
866 866 @unfilteredmethod # Until we get smarter cache management
867 867 def _rollback(self, dryrun, force):
868 868 ui = self.ui
869 869 try:
870 870 args = self.opener.read('undo.desc').splitlines()
871 871 (oldlen, desc, detail) = (int(args[0]), args[1], None)
872 872 if len(args) >= 3:
873 873 detail = args[2]
874 874 oldtip = oldlen - 1
875 875
876 876 if detail and ui.verbose:
877 877 msg = (_('repository tip rolled back to revision %s'
878 878 ' (undo %s: %s)\n')
879 879 % (oldtip, desc, detail))
880 880 else:
881 881 msg = (_('repository tip rolled back to revision %s'
882 882 ' (undo %s)\n')
883 883 % (oldtip, desc))
884 884 except IOError:
885 885 msg = _('rolling back unknown transaction\n')
886 886 desc = None
887 887
888 888 if not force and self['.'] != self['tip'] and desc == 'commit':
889 889 raise util.Abort(
890 890 _('rollback of last commit while not checked out '
891 891 'may lose data'), hint=_('use -f to force'))
892 892
893 893 ui.status(msg)
894 894 if dryrun:
895 895 return 0
896 896
897 897 parents = self.dirstate.parents()
898 898 self.destroying()
899 899 transaction.rollback(self.sopener, self.sjoin('undo'), ui.warn)
900 900 if os.path.exists(self.join('undo.bookmarks')):
901 901 util.rename(self.join('undo.bookmarks'),
902 902 self.join('bookmarks'))
903 903 if os.path.exists(self.sjoin('undo.phaseroots')):
904 904 util.rename(self.sjoin('undo.phaseroots'),
905 905 self.sjoin('phaseroots'))
906 906 self.invalidate()
907 907
908 908 parentgone = (parents[0] not in self.changelog.nodemap or
909 909 parents[1] not in self.changelog.nodemap)
910 910 if parentgone:
911 911 util.rename(self.join('undo.dirstate'), self.join('dirstate'))
912 912 try:
913 913 branch = self.opener.read('undo.branch')
914 914 self.dirstate.setbranch(encoding.tolocal(branch))
915 915 except IOError:
916 916 ui.warn(_('named branch could not be reset: '
917 917 'current branch is still \'%s\'\n')
918 918 % self.dirstate.branch())
919 919
920 920 self.dirstate.invalidate()
921 921 parents = tuple([p.rev() for p in self.parents()])
922 922 if len(parents) > 1:
923 923 ui.status(_('working directory now based on '
924 924 'revisions %d and %d\n') % parents)
925 925 else:
926 926 ui.status(_('working directory now based on '
927 927 'revision %d\n') % parents)
928 928 # TODO: if we know which new heads may result from this rollback, pass
929 929 # them to destroy(), which will prevent the branchhead cache from being
930 930 # invalidated.
931 931 self.destroyed()
932 932 return 0
933 933
934 934 def invalidatecaches(self):
935 935
936 936 if '_tagscache' in vars(self):
937 937 # can't use delattr on proxy
938 938 del self.__dict__['_tagscache']
939 939
940 940 self.unfiltered()._branchcaches.clear()
941 941 self.invalidatevolatilesets()
942 942
943 943 def invalidatevolatilesets(self):
944 944 self.filteredrevcache.clear()
945 945 obsolete.clearobscaches(self)
946 946
947 947 def invalidatedirstate(self):
948 948 '''Invalidates the dirstate, causing the next call to dirstate
949 949 to check if it was modified since the last time it was read,
950 950 rereading it if it has.
951 951
952 952 This is different to dirstate.invalidate() that it doesn't always
953 953 rereads the dirstate. Use dirstate.invalidate() if you want to
954 954 explicitly read the dirstate again (i.e. restoring it to a previous
955 955 known good state).'''
956 956 if hasunfilteredcache(self, 'dirstate'):
957 957 for k in self.dirstate._filecache:
958 958 try:
959 959 delattr(self.dirstate, k)
960 960 except AttributeError:
961 961 pass
962 962 delattr(self.unfiltered(), 'dirstate')
963 963
964 964 def invalidate(self):
965 965 unfiltered = self.unfiltered() # all file caches are stored unfiltered
966 966 for k in self._filecache:
967 967 # dirstate is invalidated separately in invalidatedirstate()
968 968 if k == 'dirstate':
969 969 continue
970 970
971 971 try:
972 972 delattr(unfiltered, k)
973 973 except AttributeError:
974 974 pass
975 975 self.invalidatecaches()
976 976
977 977 def _lock(self, lockname, wait, releasefn, acquirefn, desc):
978 978 try:
979 979 l = lock.lock(lockname, 0, releasefn, desc=desc)
980 980 except error.LockHeld, inst:
981 981 if not wait:
982 982 raise
983 983 self.ui.warn(_("waiting for lock on %s held by %r\n") %
984 984 (desc, inst.locker))
985 985 # default to 600 seconds timeout
986 986 l = lock.lock(lockname, int(self.ui.config("ui", "timeout", "600")),
987 987 releasefn, desc=desc)
988 988 if acquirefn:
989 989 acquirefn()
990 990 return l
991 991
992 992 def _afterlock(self, callback):
993 993 """add a callback to the current repository lock.
994 994
995 995 The callback will be executed on lock release."""
996 996 l = self._lockref and self._lockref()
997 997 if l:
998 998 l.postrelease.append(callback)
999 999 else:
1000 1000 callback()
1001 1001
1002 1002 def lock(self, wait=True):
1003 1003 '''Lock the repository store (.hg/store) and return a weak reference
1004 1004 to the lock. Use this before modifying the store (e.g. committing or
1005 1005 stripping). If you are opening a transaction, get a lock as well.)'''
1006 1006 l = self._lockref and self._lockref()
1007 1007 if l is not None and l.held:
1008 1008 l.lock()
1009 1009 return l
1010 1010
1011 1011 def unlock():
1012 1012 self.store.write()
1013 1013 if hasunfilteredcache(self, '_phasecache'):
1014 1014 self._phasecache.write()
1015 1015 for k, ce in self._filecache.items():
1016 1016 if k == 'dirstate' or k not in self.__dict__:
1017 1017 continue
1018 1018 ce.refresh()
1019 1019
1020 1020 l = self._lock(self.sjoin("lock"), wait, unlock,
1021 1021 self.invalidate, _('repository %s') % self.origroot)
1022 1022 self._lockref = weakref.ref(l)
1023 1023 return l
1024 1024
1025 1025 def wlock(self, wait=True):
1026 1026 '''Lock the non-store parts of the repository (everything under
1027 1027 .hg except .hg/store) and return a weak reference to the lock.
1028 1028 Use this before modifying files in .hg.'''
1029 1029 l = self._wlockref and self._wlockref()
1030 1030 if l is not None and l.held:
1031 1031 l.lock()
1032 1032 return l
1033 1033
1034 1034 def unlock():
1035 1035 self.dirstate.write()
1036 1036 self._filecache['dirstate'].refresh()
1037 1037
1038 1038 l = self._lock(self.join("wlock"), wait, unlock,
1039 1039 self.invalidatedirstate, _('working directory of %s') %
1040 1040 self.origroot)
1041 1041 self._wlockref = weakref.ref(l)
1042 1042 return l
1043 1043
1044 1044 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1045 1045 """
1046 1046 commit an individual file as part of a larger transaction
1047 1047 """
1048 1048
1049 1049 fname = fctx.path()
1050 1050 text = fctx.data()
1051 1051 flog = self.file(fname)
1052 1052 fparent1 = manifest1.get(fname, nullid)
1053 1053 fparent2 = fparent2o = manifest2.get(fname, nullid)
1054 1054
1055 1055 meta = {}
1056 1056 copy = fctx.renamed()
1057 1057 if copy and copy[0] != fname:
1058 1058 # Mark the new revision of this file as a copy of another
1059 1059 # file. This copy data will effectively act as a parent
1060 1060 # of this new revision. If this is a merge, the first
1061 1061 # parent will be the nullid (meaning "look up the copy data")
1062 1062 # and the second one will be the other parent. For example:
1063 1063 #
1064 1064 # 0 --- 1 --- 3 rev1 changes file foo
1065 1065 # \ / rev2 renames foo to bar and changes it
1066 1066 # \- 2 -/ rev3 should have bar with all changes and
1067 1067 # should record that bar descends from
1068 1068 # bar in rev2 and foo in rev1
1069 1069 #
1070 1070 # this allows this merge to succeed:
1071 1071 #
1072 1072 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1073 1073 # \ / merging rev3 and rev4 should use bar@rev2
1074 1074 # \- 2 --- 4 as the merge base
1075 1075 #
1076 1076
1077 1077 cfname = copy[0]
1078 1078 crev = manifest1.get(cfname)
1079 1079 newfparent = fparent2
1080 1080
1081 1081 if manifest2: # branch merge
1082 1082 if fparent2 == nullid or crev is None: # copied on remote side
1083 1083 if cfname in manifest2:
1084 1084 crev = manifest2[cfname]
1085 1085 newfparent = fparent1
1086 1086
1087 1087 # find source in nearest ancestor if we've lost track
1088 1088 if not crev:
1089 1089 self.ui.debug(" %s: searching for copy revision for %s\n" %
1090 1090 (fname, cfname))
1091 1091 for ancestor in self[None].ancestors():
1092 1092 if cfname in ancestor:
1093 1093 crev = ancestor[cfname].filenode()
1094 1094 break
1095 1095
1096 1096 if crev:
1097 1097 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1098 1098 meta["copy"] = cfname
1099 1099 meta["copyrev"] = hex(crev)
1100 1100 fparent1, fparent2 = nullid, newfparent
1101 1101 else:
1102 1102 self.ui.warn(_("warning: can't find ancestor for '%s' "
1103 1103 "copied from '%s'!\n") % (fname, cfname))
1104 1104
1105 1105 elif fparent2 != nullid:
1106 1106 # is one parent an ancestor of the other?
1107 1107 fparentancestor = flog.ancestor(fparent1, fparent2)
1108 1108 if fparentancestor == fparent1:
1109 1109 fparent1, fparent2 = fparent2, nullid
1110 1110 elif fparentancestor == fparent2:
1111 1111 fparent2 = nullid
1112 1112
1113 1113 # is the file changed?
1114 1114 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1115 1115 changelist.append(fname)
1116 1116 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1117 1117
1118 1118 # are just the flags changed during merge?
1119 1119 if fparent1 != fparent2o and manifest1.flags(fname) != fctx.flags():
1120 1120 changelist.append(fname)
1121 1121
1122 1122 return fparent1
1123 1123
1124 1124 @unfilteredmethod
1125 1125 def commit(self, text="", user=None, date=None, match=None, force=False,
1126 1126 editor=False, extra={}):
1127 1127 """Add a new revision to current repository.
1128 1128
1129 1129 Revision information is gathered from the working directory,
1130 1130 match can be used to filter the committed files. If editor is
1131 1131 supplied, it is called to get a commit message.
1132 1132 """
1133 1133
1134 1134 def fail(f, msg):
1135 1135 raise util.Abort('%s: %s' % (f, msg))
1136 1136
1137 1137 if not match:
1138 1138 match = matchmod.always(self.root, '')
1139 1139
1140 1140 if not force:
1141 1141 vdirs = []
1142 1142 match.dir = vdirs.append
1143 1143 match.bad = fail
1144 1144
1145 1145 wlock = self.wlock()
1146 1146 try:
1147 1147 wctx = self[None]
1148 1148 merge = len(wctx.parents()) > 1
1149 1149
1150 1150 if (not force and merge and match and
1151 1151 (match.files() or match.anypats())):
1152 1152 raise util.Abort(_('cannot partially commit a merge '
1153 1153 '(do not specify files or patterns)'))
1154 1154
1155 1155 changes = self.status(match=match, clean=force)
1156 1156 if force:
1157 1157 changes[0].extend(changes[6]) # mq may commit unchanged files
1158 1158
1159 1159 # check subrepos
1160 1160 subs = []
1161 1161 commitsubs = set()
1162 1162 newstate = wctx.substate.copy()
1163 1163 # only manage subrepos and .hgsubstate if .hgsub is present
1164 1164 if '.hgsub' in wctx:
1165 1165 # we'll decide whether to track this ourselves, thanks
1166 1166 if '.hgsubstate' in changes[0]:
1167 1167 changes[0].remove('.hgsubstate')
1168 1168 if '.hgsubstate' in changes[2]:
1169 1169 changes[2].remove('.hgsubstate')
1170 1170
1171 1171 # compare current state to last committed state
1172 1172 # build new substate based on last committed state
1173 1173 oldstate = wctx.p1().substate
1174 1174 for s in sorted(newstate.keys()):
1175 1175 if not match(s):
1176 1176 # ignore working copy, use old state if present
1177 1177 if s in oldstate:
1178 1178 newstate[s] = oldstate[s]
1179 1179 continue
1180 1180 if not force:
1181 1181 raise util.Abort(
1182 1182 _("commit with new subrepo %s excluded") % s)
1183 1183 if wctx.sub(s).dirty(True):
1184 1184 if not self.ui.configbool('ui', 'commitsubrepos'):
1185 1185 raise util.Abort(
1186 1186 _("uncommitted changes in subrepo %s") % s,
1187 1187 hint=_("use --subrepos for recursive commit"))
1188 1188 subs.append(s)
1189 1189 commitsubs.add(s)
1190 1190 else:
1191 1191 bs = wctx.sub(s).basestate()
1192 1192 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1193 1193 if oldstate.get(s, (None, None, None))[1] != bs:
1194 1194 subs.append(s)
1195 1195
1196 1196 # check for removed subrepos
1197 1197 for p in wctx.parents():
1198 1198 r = [s for s in p.substate if s not in newstate]
1199 1199 subs += [s for s in r if match(s)]
1200 1200 if subs:
1201 1201 if (not match('.hgsub') and
1202 1202 '.hgsub' in (wctx.modified() + wctx.added())):
1203 1203 raise util.Abort(
1204 1204 _("can't commit subrepos without .hgsub"))
1205 1205 changes[0].insert(0, '.hgsubstate')
1206 1206
1207 1207 elif '.hgsub' in changes[2]:
1208 1208 # clean up .hgsubstate when .hgsub is removed
1209 1209 if ('.hgsubstate' in wctx and
1210 1210 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
1211 1211 changes[2].insert(0, '.hgsubstate')
1212 1212
1213 1213 # make sure all explicit patterns are matched
1214 1214 if not force and match.files():
1215 1215 matched = set(changes[0] + changes[1] + changes[2])
1216 1216
1217 1217 for f in match.files():
1218 1218 f = self.dirstate.normalize(f)
1219 1219 if f == '.' or f in matched or f in wctx.substate:
1220 1220 continue
1221 1221 if f in changes[3]: # missing
1222 1222 fail(f, _('file not found!'))
1223 1223 if f in vdirs: # visited directory
1224 1224 d = f + '/'
1225 1225 for mf in matched:
1226 1226 if mf.startswith(d):
1227 1227 break
1228 1228 else:
1229 1229 fail(f, _("no match under directory!"))
1230 1230 elif f not in self.dirstate:
1231 1231 fail(f, _("file not tracked!"))
1232 1232
1233 1233 cctx = context.workingctx(self, text, user, date, extra, changes)
1234 1234
1235 1235 if (not force and not extra.get("close") and not merge
1236 1236 and not cctx.files()
1237 1237 and wctx.branch() == wctx.p1().branch()):
1238 1238 return None
1239 1239
1240 1240 if merge and cctx.deleted():
1241 1241 raise util.Abort(_("cannot commit merge with missing files"))
1242 1242
1243 1243 ms = mergemod.mergestate(self)
1244 1244 for f in changes[0]:
1245 1245 if f in ms and ms[f] == 'u':
1246 1246 raise util.Abort(_("unresolved merge conflicts "
1247 1247 "(see hg help resolve)"))
1248 1248
1249 1249 if editor:
1250 1250 cctx._text = editor(self, cctx, subs)
1251 1251 edited = (text != cctx._text)
1252 1252
1253 1253 # commit subs and write new state
1254 1254 if subs:
1255 1255 for s in sorted(commitsubs):
1256 1256 sub = wctx.sub(s)
1257 1257 self.ui.status(_('committing subrepository %s\n') %
1258 1258 subrepo.subrelpath(sub))
1259 1259 sr = sub.commit(cctx._text, user, date)
1260 1260 newstate[s] = (newstate[s][0], sr)
1261 1261 subrepo.writestate(self, newstate)
1262 1262
1263 1263 # Save commit message in case this transaction gets rolled back
1264 1264 # (e.g. by a pretxncommit hook). Leave the content alone on
1265 1265 # the assumption that the user will use the same editor again.
1266 1266 msgfn = self.savecommitmessage(cctx._text)
1267 1267
1268 1268 p1, p2 = self.dirstate.parents()
1269 1269 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1270 1270 try:
1271 1271 self.hook("precommit", throw=True, parent1=hookp1,
1272 1272 parent2=hookp2)
1273 1273 ret = self.commitctx(cctx, True)
1274 1274 except: # re-raises
1275 1275 if edited:
1276 1276 self.ui.write(
1277 1277 _('note: commit message saved in %s\n') % msgfn)
1278 1278 raise
1279 1279
1280 1280 # update bookmarks, dirstate and mergestate
1281 1281 bookmarks.update(self, [p1, p2], ret)
1282 1282 cctx.markcommitted(ret)
1283 1283 ms.reset()
1284 1284 finally:
1285 1285 wlock.release()
1286 1286
1287 1287 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1288 1288 self.hook("commit", node=node, parent1=parent1, parent2=parent2)
1289 1289 self._afterlock(commithook)
1290 1290 return ret
1291 1291
1292 1292 @unfilteredmethod
1293 1293 def commitctx(self, ctx, error=False):
1294 1294 """Add a new revision to current repository.
1295 1295 Revision information is passed via the context argument.
1296 1296 """
1297 1297
1298 1298 tr = lock = None
1299 1299 removed = list(ctx.removed())
1300 1300 p1, p2 = ctx.p1(), ctx.p2()
1301 1301 user = ctx.user()
1302 1302
1303 1303 lock = self.lock()
1304 1304 try:
1305 1305 tr = self.transaction("commit")
1306 1306 trp = weakref.proxy(tr)
1307 1307
1308 1308 if ctx.files():
1309 1309 m1 = p1.manifest().copy()
1310 1310 m2 = p2.manifest()
1311 1311
1312 1312 # check in files
1313 1313 new = {}
1314 1314 changed = []
1315 1315 linkrev = len(self)
1316 1316 for f in sorted(ctx.modified() + ctx.added()):
1317 1317 self.ui.note(f + "\n")
1318 1318 try:
1319 1319 fctx = ctx[f]
1320 1320 new[f] = self._filecommit(fctx, m1, m2, linkrev, trp,
1321 1321 changed)
1322 1322 m1.set(f, fctx.flags())
1323 1323 except OSError, inst:
1324 1324 self.ui.warn(_("trouble committing %s!\n") % f)
1325 1325 raise
1326 1326 except IOError, inst:
1327 1327 errcode = getattr(inst, 'errno', errno.ENOENT)
1328 1328 if error or errcode and errcode != errno.ENOENT:
1329 1329 self.ui.warn(_("trouble committing %s!\n") % f)
1330 1330 raise
1331 1331 else:
1332 1332 removed.append(f)
1333 1333
1334 1334 # update manifest
1335 1335 m1.update(new)
1336 1336 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1337 1337 drop = [f for f in removed if f in m1]
1338 1338 for f in drop:
1339 1339 del m1[f]
1340 1340 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1341 1341 p2.manifestnode(), (new, drop))
1342 1342 files = changed + removed
1343 1343 else:
1344 1344 mn = p1.manifestnode()
1345 1345 files = []
1346 1346
1347 1347 # update changelog
1348 1348 self.changelog.delayupdate()
1349 1349 n = self.changelog.add(mn, files, ctx.description(),
1350 1350 trp, p1.node(), p2.node(),
1351 1351 user, ctx.date(), ctx.extra().copy())
1352 1352 p = lambda: self.changelog.writepending() and self.root or ""
1353 1353 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1354 1354 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1355 1355 parent2=xp2, pending=p)
1356 1356 self.changelog.finalize(trp)
1357 1357 # set the new commit is proper phase
1358 1358 targetphase = phases.newcommitphase(self.ui)
1359 1359 if targetphase:
1360 1360 # retract boundary do not alter parent changeset.
1361 1361 # if a parent have higher the resulting phase will
1362 1362 # be compliant anyway
1363 1363 #
1364 1364 # if minimal phase was 0 we don't need to retract anything
1365 1365 phases.retractboundary(self, targetphase, [n])
1366 1366 tr.close()
1367 1367 branchmap.updatecache(self.filtered('served'))
1368 1368 return n
1369 1369 finally:
1370 1370 if tr:
1371 1371 tr.release()
1372 1372 lock.release()
1373 1373
1374 1374 @unfilteredmethod
1375 1375 def destroying(self):
1376 1376 '''Inform the repository that nodes are about to be destroyed.
1377 1377 Intended for use by strip and rollback, so there's a common
1378 1378 place for anything that has to be done before destroying history.
1379 1379
1380 1380 This is mostly useful for saving state that is in memory and waiting
1381 1381 to be flushed when the current lock is released. Because a call to
1382 1382 destroyed is imminent, the repo will be invalidated causing those
1383 1383 changes to stay in memory (waiting for the next unlock), or vanish
1384 1384 completely.
1385 1385 '''
1386 1386 # When using the same lock to commit and strip, the phasecache is left
1387 1387 # dirty after committing. Then when we strip, the repo is invalidated,
1388 1388 # causing those changes to disappear.
1389 1389 if '_phasecache' in vars(self):
1390 1390 self._phasecache.write()
1391 1391
1392 1392 @unfilteredmethod
1393 1393 def destroyed(self):
1394 1394 '''Inform the repository that nodes have been destroyed.
1395 1395 Intended for use by strip and rollback, so there's a common
1396 1396 place for anything that has to be done after destroying history.
1397 1397 '''
1398 1398 # When one tries to:
1399 1399 # 1) destroy nodes thus calling this method (e.g. strip)
1400 1400 # 2) use phasecache somewhere (e.g. commit)
1401 1401 #
1402 1402 # then 2) will fail because the phasecache contains nodes that were
1403 1403 # removed. We can either remove phasecache from the filecache,
1404 1404 # causing it to reload next time it is accessed, or simply filter
1405 1405 # the removed nodes now and write the updated cache.
1406 1406 if '_phasecache' in self._filecache:
1407 1407 self._phasecache.filterunknown(self)
1408 1408 self._phasecache.write()
1409 1409
1410 1410 # update the 'served' branch cache to help read only server process
1411 1411 # Thanks to branchcache collaboration this is done from the nearest
1412 1412 # filtered subset and it is expected to be fast.
1413 1413 branchmap.updatecache(self.filtered('served'))
1414 1414
1415 1415 # Ensure the persistent tag cache is updated. Doing it now
1416 1416 # means that the tag cache only has to worry about destroyed
1417 1417 # heads immediately after a strip/rollback. That in turn
1418 1418 # guarantees that "cachetip == currenttip" (comparing both rev
1419 1419 # and node) always means no nodes have been added or destroyed.
1420 1420
1421 1421 # XXX this is suboptimal when qrefresh'ing: we strip the current
1422 1422 # head, refresh the tag cache, then immediately add a new head.
1423 1423 # But I think doing it this way is necessary for the "instant
1424 1424 # tag cache retrieval" case to work.
1425 1425 self.invalidate()
1426 1426
1427 1427 def walk(self, match, node=None):
1428 1428 '''
1429 1429 walk recursively through the directory tree or a given
1430 1430 changeset, finding all files matched by the match
1431 1431 function
1432 1432 '''
1433 1433 return self[node].walk(match)
1434 1434
1435 1435 def status(self, node1='.', node2=None, match=None,
1436 1436 ignored=False, clean=False, unknown=False,
1437 1437 listsubrepos=False):
1438 1438 """return status of files between two nodes or node and working
1439 1439 directory.
1440 1440
1441 1441 If node1 is None, use the first dirstate parent instead.
1442 1442 If node2 is None, compare node1 with working directory.
1443 1443 """
1444 1444
1445 1445 def mfmatches(ctx):
1446 1446 mf = ctx.manifest().copy()
1447 1447 if match.always():
1448 1448 return mf
1449 1449 for fn in mf.keys():
1450 1450 if not match(fn):
1451 1451 del mf[fn]
1452 1452 return mf
1453 1453
1454 1454 if isinstance(node1, context.changectx):
1455 1455 ctx1 = node1
1456 1456 else:
1457 1457 ctx1 = self[node1]
1458 1458 if isinstance(node2, context.changectx):
1459 1459 ctx2 = node2
1460 1460 else:
1461 1461 ctx2 = self[node2]
1462 1462
1463 1463 working = ctx2.rev() is None
1464 1464 parentworking = working and ctx1 == self['.']
1465 1465 match = match or matchmod.always(self.root, self.getcwd())
1466 1466 listignored, listclean, listunknown = ignored, clean, unknown
1467 1467
1468 1468 # load earliest manifest first for caching reasons
1469 1469 if not working and ctx2.rev() < ctx1.rev():
1470 1470 ctx2.manifest()
1471 1471
1472 1472 if not parentworking:
1473 1473 def bad(f, msg):
1474 1474 # 'f' may be a directory pattern from 'match.files()',
1475 1475 # so 'f not in ctx1' is not enough
1476 1476 if f not in ctx1 and f not in ctx1.dirs():
1477 1477 self.ui.warn('%s: %s\n' % (self.dirstate.pathto(f), msg))
1478 1478 match.bad = bad
1479 1479
1480 1480 if working: # we need to scan the working dir
1481 1481 subrepos = []
1482 1482 if '.hgsub' in self.dirstate:
1483 1483 subrepos = sorted(ctx2.substate)
1484 1484 s = self.dirstate.status(match, subrepos, listignored,
1485 1485 listclean, listunknown)
1486 1486 cmp, modified, added, removed, deleted, unknown, ignored, clean = s
1487 1487
1488 1488 # check for any possibly clean files
1489 1489 if parentworking and cmp:
1490 1490 fixup = []
1491 1491 # do a full compare of any files that might have changed
1492 1492 for f in sorted(cmp):
1493 1493 if (f not in ctx1 or ctx2.flags(f) != ctx1.flags(f)
1494 1494 or ctx1[f].cmp(ctx2[f])):
1495 1495 modified.append(f)
1496 1496 else:
1497 1497 fixup.append(f)
1498 1498
1499 1499 # update dirstate for files that are actually clean
1500 1500 if fixup:
1501 1501 if listclean:
1502 1502 clean += fixup
1503 1503
1504 1504 try:
1505 1505 # updating the dirstate is optional
1506 1506 # so we don't wait on the lock
1507 1507 wlock = self.wlock(False)
1508 1508 try:
1509 1509 for f in fixup:
1510 1510 self.dirstate.normal(f)
1511 1511 finally:
1512 1512 wlock.release()
1513 1513 except error.LockError:
1514 1514 pass
1515 1515
1516 1516 if not parentworking:
1517 1517 mf1 = mfmatches(ctx1)
1518 1518 if working:
1519 1519 # we are comparing working dir against non-parent
1520 1520 # generate a pseudo-manifest for the working dir
1521 1521 mf2 = mfmatches(self['.'])
1522 1522 for f in cmp + modified + added:
1523 1523 mf2[f] = None
1524 1524 mf2.set(f, ctx2.flags(f))
1525 1525 for f in removed:
1526 1526 if f in mf2:
1527 1527 del mf2[f]
1528 1528 else:
1529 1529 # we are comparing two revisions
1530 1530 deleted, unknown, ignored = [], [], []
1531 1531 mf2 = mfmatches(ctx2)
1532 1532
1533 1533 modified, added, clean = [], [], []
1534 1534 withflags = mf1.withflags() | mf2.withflags()
1535 1535 for fn in mf2:
1536 1536 if fn in mf1:
1537 1537 if (fn not in deleted and
1538 1538 ((fn in withflags and mf1.flags(fn) != mf2.flags(fn)) or
1539 1539 (mf1[fn] != mf2[fn] and
1540 1540 (mf2[fn] or ctx1[fn].cmp(ctx2[fn]))))):
1541 1541 modified.append(fn)
1542 1542 elif listclean:
1543 1543 clean.append(fn)
1544 1544 del mf1[fn]
1545 1545 elif fn not in deleted:
1546 1546 added.append(fn)
1547 1547 removed = mf1.keys()
1548 1548
1549 1549 if working and modified and not self.dirstate._checklink:
1550 1550 # Symlink placeholders may get non-symlink-like contents
1551 1551 # via user error or dereferencing by NFS or Samba servers,
1552 1552 # so we filter out any placeholders that don't look like a
1553 1553 # symlink
1554 1554 sane = []
1555 1555 for f in modified:
1556 1556 if ctx2.flags(f) == 'l':
1557 1557 d = ctx2[f].data()
1558 1558 if len(d) >= 1024 or '\n' in d or util.binary(d):
1559 1559 self.ui.debug('ignoring suspect symlink placeholder'
1560 1560 ' "%s"\n' % f)
1561 1561 continue
1562 1562 sane.append(f)
1563 1563 modified = sane
1564 1564
1565 1565 r = modified, added, removed, deleted, unknown, ignored, clean
1566 1566
1567 1567 if listsubrepos:
1568 1568 for subpath, sub in subrepo.itersubrepos(ctx1, ctx2):
1569 1569 if working:
1570 1570 rev2 = None
1571 1571 else:
1572 1572 rev2 = ctx2.substate[subpath][1]
1573 1573 try:
1574 1574 submatch = matchmod.narrowmatcher(subpath, match)
1575 1575 s = sub.status(rev2, match=submatch, ignored=listignored,
1576 1576 clean=listclean, unknown=listunknown,
1577 1577 listsubrepos=True)
1578 1578 for rfiles, sfiles in zip(r, s):
1579 1579 rfiles.extend("%s/%s" % (subpath, f) for f in sfiles)
1580 1580 except error.LookupError:
1581 1581 self.ui.status(_("skipping missing subrepository: %s\n")
1582 1582 % subpath)
1583 1583
1584 1584 for l in r:
1585 1585 l.sort()
1586 1586 return r
1587 1587
1588 1588 def heads(self, start=None):
1589 1589 heads = self.changelog.heads(start)
1590 1590 # sort the output in rev descending order
1591 1591 return sorted(heads, key=self.changelog.rev, reverse=True)
1592 1592
1593 1593 def branchheads(self, branch=None, start=None, closed=False):
1594 1594 '''return a (possibly filtered) list of heads for the given branch
1595 1595
1596 1596 Heads are returned in topological order, from newest to oldest.
1597 1597 If branch is None, use the dirstate branch.
1598 1598 If start is not None, return only heads reachable from start.
1599 1599 If closed is True, return heads that are marked as closed as well.
1600 1600 '''
1601 1601 if branch is None:
1602 1602 branch = self[None].branch()
1603 1603 branches = self.branchmap()
1604 1604 if branch not in branches:
1605 1605 return []
1606 1606 # the cache returns heads ordered lowest to highest
1607 1607 bheads = list(reversed(branches[branch]))
1608 1608 if start is not None:
1609 1609 # filter out the heads that cannot be reached from startrev
1610 1610 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1611 1611 bheads = [h for h in bheads if h in fbheads]
1612 1612 if not closed:
1613 1613 bheads = [h for h in bheads if not self[h].closesbranch()]
1614 1614 return bheads
1615 1615
1616 1616 def branches(self, nodes):
1617 1617 if not nodes:
1618 1618 nodes = [self.changelog.tip()]
1619 1619 b = []
1620 1620 for n in nodes:
1621 1621 t = n
1622 1622 while True:
1623 1623 p = self.changelog.parents(n)
1624 1624 if p[1] != nullid or p[0] == nullid:
1625 1625 b.append((t, n, p[0], p[1]))
1626 1626 break
1627 1627 n = p[0]
1628 1628 return b
1629 1629
1630 1630 def between(self, pairs):
1631 1631 r = []
1632 1632
1633 1633 for top, bottom in pairs:
1634 1634 n, l, i = top, [], 0
1635 1635 f = 1
1636 1636
1637 1637 while n != bottom and n != nullid:
1638 1638 p = self.changelog.parents(n)[0]
1639 1639 if i == f:
1640 1640 l.append(n)
1641 1641 f = f * 2
1642 1642 n = p
1643 1643 i += 1
1644 1644
1645 1645 r.append(l)
1646 1646
1647 1647 return r
1648 1648
1649 1649 def pull(self, remote, heads=None, force=False):
1650 1650 # don't open transaction for nothing or you break future useful
1651 1651 # rollback call
1652 1652 tr = None
1653 1653 trname = 'pull\n' + util.hidepassword(remote.url())
1654 1654 lock = self.lock()
1655 1655 try:
1656 1656 tmp = discovery.findcommonincoming(self, remote, heads=heads,
1657 1657 force=force)
1658 1658 common, fetch, rheads = tmp
1659 1659 if not fetch:
1660 1660 self.ui.status(_("no changes found\n"))
1661 1661 added = []
1662 1662 result = 0
1663 1663 else:
1664 1664 tr = self.transaction(trname)
1665 1665 if heads is None and list(common) == [nullid]:
1666 1666 self.ui.status(_("requesting all changes\n"))
1667 1667 elif heads is None and remote.capable('changegroupsubset'):
1668 1668 # issue1320, avoid a race if remote changed after discovery
1669 1669 heads = rheads
1670 1670
1671 1671 if remote.capable('getbundle'):
1672 1672 cg = remote.getbundle('pull', common=common,
1673 1673 heads=heads or rheads)
1674 1674 elif heads is None:
1675 1675 cg = remote.changegroup(fetch, 'pull')
1676 1676 elif not remote.capable('changegroupsubset'):
1677 1677 raise util.Abort(_("partial pull cannot be done because "
1678 1678 "other repository doesn't support "
1679 1679 "changegroupsubset."))
1680 1680 else:
1681 1681 cg = remote.changegroupsubset(fetch, heads, 'pull')
1682 1682 # we use unfiltered changelog here because hidden revision must
1683 1683 # be taken in account for phase synchronization. They may
1684 1684 # becomes public and becomes visible again.
1685 1685 cl = self.unfiltered().changelog
1686 1686 clstart = len(cl)
1687 1687 result = self.addchangegroup(cg, 'pull', remote.url())
1688 1688 clend = len(cl)
1689 1689 added = [cl.node(r) for r in xrange(clstart, clend)]
1690 1690
1691 1691 # compute target subset
1692 1692 if heads is None:
1693 1693 # We pulled every thing possible
1694 1694 # sync on everything common
1695 1695 subset = common + added
1696 1696 else:
1697 1697 # We pulled a specific subset
1698 1698 # sync on this subset
1699 1699 subset = heads
1700 1700
1701 1701 # Get remote phases data from remote
1702 1702 remotephases = remote.listkeys('phases')
1703 1703 publishing = bool(remotephases.get('publishing', False))
1704 1704 if remotephases and not publishing:
1705 1705 # remote is new and unpublishing
1706 1706 pheads, _dr = phases.analyzeremotephases(self, subset,
1707 1707 remotephases)
1708 1708 phases.advanceboundary(self, phases.public, pheads)
1709 1709 phases.advanceboundary(self, phases.draft, subset)
1710 1710 else:
1711 1711 # Remote is old or publishing all common changesets
1712 1712 # should be seen as public
1713 1713 phases.advanceboundary(self, phases.public, subset)
1714 1714
1715 1715 if obsolete._enabled:
1716 1716 self.ui.debug('fetching remote obsolete markers\n')
1717 1717 remoteobs = remote.listkeys('obsolete')
1718 1718 if 'dump0' in remoteobs:
1719 1719 if tr is None:
1720 1720 tr = self.transaction(trname)
1721 1721 for key in sorted(remoteobs, reverse=True):
1722 1722 if key.startswith('dump'):
1723 1723 data = base85.b85decode(remoteobs[key])
1724 1724 self.obsstore.mergemarkers(tr, data)
1725 1725 self.invalidatevolatilesets()
1726 1726 if tr is not None:
1727 1727 tr.close()
1728 1728 finally:
1729 1729 if tr is not None:
1730 1730 tr.release()
1731 1731 lock.release()
1732 1732
1733 1733 return result
1734 1734
1735 1735 def checkpush(self, force, revs):
1736 1736 """Extensions can override this function if additional checks have
1737 1737 to be performed before pushing, or call it if they override push
1738 1738 command.
1739 1739 """
1740 1740 pass
1741 1741
1742 1742 def push(self, remote, force=False, revs=None, newbranch=False):
1743 1743 '''Push outgoing changesets (limited by revs) from the current
1744 1744 repository to remote. Return an integer:
1745 1745 - None means nothing to push
1746 1746 - 0 means HTTP error
1747 1747 - 1 means we pushed and remote head count is unchanged *or*
1748 1748 we have outgoing changesets but refused to push
1749 1749 - other values as described by addchangegroup()
1750 1750 '''
1751 1751 # there are two ways to push to remote repo:
1752 1752 #
1753 1753 # addchangegroup assumes local user can lock remote
1754 1754 # repo (local filesystem, old ssh servers).
1755 1755 #
1756 1756 # unbundle assumes local user cannot lock remote repo (new ssh
1757 1757 # servers, http servers).
1758 1758
1759 1759 if not remote.canpush():
1760 1760 raise util.Abort(_("destination does not support push"))
1761 1761 unfi = self.unfiltered()
1762 1762 # get local lock as we might write phase data
1763 1763 locallock = self.lock()
1764 1764 try:
1765 1765 self.checkpush(force, revs)
1766 1766 lock = None
1767 1767 unbundle = remote.capable('unbundle')
1768 1768 if not unbundle:
1769 1769 lock = remote.lock()
1770 1770 try:
1771 1771 # discovery
1772 1772 fci = discovery.findcommonincoming
1773 1773 commoninc = fci(unfi, remote, force=force)
1774 1774 common, inc, remoteheads = commoninc
1775 1775 fco = discovery.findcommonoutgoing
1776 1776 outgoing = fco(unfi, remote, onlyheads=revs,
1777 1777 commoninc=commoninc, force=force)
1778 1778
1779 1779
1780 1780 if not outgoing.missing:
1781 1781 # nothing to push
1782 1782 scmutil.nochangesfound(unfi.ui, unfi, outgoing.excluded)
1783 1783 ret = None
1784 1784 else:
1785 1785 # something to push
1786 1786 if not force:
1787 1787 # if self.obsstore == False --> no obsolete
1788 1788 # then, save the iteration
1789 1789 if unfi.obsstore:
1790 1790 # this message are here for 80 char limit reason
1791 1791 mso = _("push includes obsolete changeset: %s!")
1792 1792 mst = "push includes %s changeset: %s!"
1793 1793 # plain versions for i18n tool to detect them
1794 1794 _("push includes unstable changeset: %s!")
1795 1795 _("push includes bumped changeset: %s!")
1796 1796 _("push includes divergent changeset: %s!")
1797 1797 # If we are to push if there is at least one
1798 1798 # obsolete or unstable changeset in missing, at
1799 1799 # least one of the missinghead will be obsolete or
1800 1800 # unstable. So checking heads only is ok
1801 1801 for node in outgoing.missingheads:
1802 1802 ctx = unfi[node]
1803 1803 if ctx.obsolete():
1804 1804 raise util.Abort(mso % ctx)
1805 1805 elif ctx.troubled():
1806 1806 raise util.Abort(_(mst)
1807 1807 % (ctx.troubles()[0],
1808 1808 ctx))
1809 1809 discovery.checkheads(unfi, remote, outgoing,
1810 1810 remoteheads, newbranch,
1811 1811 bool(inc))
1812 1812
1813 1813 # create a changegroup from local
1814 1814 if revs is None and not outgoing.excluded:
1815 1815 # push everything,
1816 1816 # use the fast path, no race possible on push
1817 1817 cg = self._changegroup(outgoing.missing, 'push')
1818 1818 else:
1819 1819 cg = self.getlocalbundle('push', outgoing)
1820 1820
1821 1821 # apply changegroup to remote
1822 1822 if unbundle:
1823 1823 # local repo finds heads on server, finds out what
1824 1824 # revs it must push. once revs transferred, if server
1825 1825 # finds it has different heads (someone else won
1826 1826 # commit/push race), server aborts.
1827 1827 if force:
1828 1828 remoteheads = ['force']
1829 1829 # ssh: return remote's addchangegroup()
1830 1830 # http: return remote's addchangegroup() or 0 for error
1831 1831 ret = remote.unbundle(cg, remoteheads, 'push')
1832 1832 else:
1833 1833 # we return an integer indicating remote head count
1834 1834 # change
1835 1835 ret = remote.addchangegroup(cg, 'push', self.url())
1836 1836
1837 1837 if ret:
1838 1838 # push succeed, synchronize target of the push
1839 1839 cheads = outgoing.missingheads
1840 1840 elif revs is None:
1841 1841 # All out push fails. synchronize all common
1842 1842 cheads = outgoing.commonheads
1843 1843 else:
1844 1844 # I want cheads = heads(::missingheads and ::commonheads)
1845 1845 # (missingheads is revs with secret changeset filtered out)
1846 1846 #
1847 1847 # This can be expressed as:
1848 1848 # cheads = ( (missingheads and ::commonheads)
1849 1849 # + (commonheads and ::missingheads))"
1850 1850 # )
1851 1851 #
1852 1852 # while trying to push we already computed the following:
1853 1853 # common = (::commonheads)
1854 1854 # missing = ((commonheads::missingheads) - commonheads)
1855 1855 #
1856 1856 # We can pick:
1857 1857 # * missingheads part of common (::commonheads)
1858 1858 common = set(outgoing.common)
1859 1859 cheads = [node for node in revs if node in common]
1860 1860 # and
1861 1861 # * commonheads parents on missing
1862 1862 revset = unfi.set('%ln and parents(roots(%ln))',
1863 1863 outgoing.commonheads,
1864 1864 outgoing.missing)
1865 1865 cheads.extend(c.node() for c in revset)
1866 1866 # even when we don't push, exchanging phase data is useful
1867 1867 remotephases = remote.listkeys('phases')
1868 1868 if (self.ui.configbool('ui', '_usedassubrepo', False)
1869 1869 and remotephases # server supports phases
1870 1870 and ret is None # nothing was pushed
1871 1871 and remotephases.get('publishing', False)):
1872 1872 # When:
1873 1873 # - this is a subrepo push
1874 1874 # - and remote support phase
1875 1875 # - and no changeset was pushed
1876 1876 # - and remote is publishing
1877 1877 # We may be in issue 3871 case!
1878 1878 # We drop the possible phase synchronisation done by
1879 1879 # courtesy to publish changesets possibly locally draft
1880 1880 # on the remote.
1881 1881 remotephases = {'publishing': 'True'}
1882 1882 if not remotephases: # old server or public only repo
1883 1883 phases.advanceboundary(self, phases.public, cheads)
1884 1884 # don't push any phase data as there is nothing to push
1885 1885 else:
1886 1886 ana = phases.analyzeremotephases(self, cheads, remotephases)
1887 1887 pheads, droots = ana
1888 1888 ### Apply remote phase on local
1889 1889 if remotephases.get('publishing', False):
1890 1890 phases.advanceboundary(self, phases.public, cheads)
1891 1891 else: # publish = False
1892 1892 phases.advanceboundary(self, phases.public, pheads)
1893 1893 phases.advanceboundary(self, phases.draft, cheads)
1894 1894 ### Apply local phase on remote
1895 1895
1896 1896 # Get the list of all revs draft on remote by public here.
1897 1897 # XXX Beware that revset break if droots is not strictly
1898 1898 # XXX root we may want to ensure it is but it is costly
1899 1899 outdated = unfi.set('heads((%ln::%ln) and public())',
1900 1900 droots, cheads)
1901 1901 for newremotehead in outdated:
1902 1902 r = remote.pushkey('phases',
1903 1903 newremotehead.hex(),
1904 1904 str(phases.draft),
1905 1905 str(phases.public))
1906 1906 if not r:
1907 1907 self.ui.warn(_('updating %s to public failed!\n')
1908 1908 % newremotehead)
1909 1909 self.ui.debug('try to push obsolete markers to remote\n')
1910 1910 if (obsolete._enabled and self.obsstore and
1911 1911 'obsolete' in remote.listkeys('namespaces')):
1912 1912 rslts = []
1913 1913 remotedata = self.listkeys('obsolete')
1914 1914 for key in sorted(remotedata, reverse=True):
1915 1915 # reverse sort to ensure we end with dump0
1916 1916 data = remotedata[key]
1917 1917 rslts.append(remote.pushkey('obsolete', key, '', data))
1918 1918 if [r for r in rslts if not r]:
1919 1919 msg = _('failed to push some obsolete markers!\n')
1920 1920 self.ui.warn(msg)
1921 1921 finally:
1922 1922 if lock is not None:
1923 1923 lock.release()
1924 1924 finally:
1925 1925 locallock.release()
1926 1926
1927 1927 self.ui.debug("checking for updated bookmarks\n")
1928 1928 rb = remote.listkeys('bookmarks')
1929 1929 for k in rb.keys():
1930 1930 if k in unfi._bookmarks:
1931 1931 nr, nl = rb[k], hex(self._bookmarks[k])
1932 1932 if nr in unfi:
1933 1933 cr = unfi[nr]
1934 1934 cl = unfi[nl]
1935 1935 if bookmarks.validdest(unfi, cr, cl):
1936 1936 r = remote.pushkey('bookmarks', k, nr, nl)
1937 1937 if r:
1938 1938 self.ui.status(_("updating bookmark %s\n") % k)
1939 1939 else:
1940 1940 self.ui.warn(_('updating bookmark %s'
1941 1941 ' failed!\n') % k)
1942 1942
1943 1943 return ret
1944 1944
1945 1945 def changegroupinfo(self, nodes, source):
1946 1946 if self.ui.verbose or source == 'bundle':
1947 1947 self.ui.status(_("%d changesets found\n") % len(nodes))
1948 1948 if self.ui.debugflag:
1949 1949 self.ui.debug("list of changesets:\n")
1950 1950 for node in nodes:
1951 1951 self.ui.debug("%s\n" % hex(node))
1952 1952
1953 1953 def changegroupsubset(self, bases, heads, source):
1954 1954 """Compute a changegroup consisting of all the nodes that are
1955 1955 descendants of any of the bases and ancestors of any of the heads.
1956 1956 Return a chunkbuffer object whose read() method will return
1957 1957 successive changegroup chunks.
1958 1958
1959 1959 It is fairly complex as determining which filenodes and which
1960 1960 manifest nodes need to be included for the changeset to be complete
1961 1961 is non-trivial.
1962 1962
1963 1963 Another wrinkle is doing the reverse, figuring out which changeset in
1964 1964 the changegroup a particular filenode or manifestnode belongs to.
1965 1965 """
1966 1966 cl = self.changelog
1967 1967 if not bases:
1968 1968 bases = [nullid]
1969 1969 csets, bases, heads = cl.nodesbetween(bases, heads)
1970 1970 # We assume that all ancestors of bases are known
1971 1971 common = cl.ancestors([cl.rev(n) for n in bases])
1972 1972 return self._changegroupsubset(common, csets, heads, source)
1973 1973
1974 1974 def getlocalbundle(self, source, outgoing):
1975 1975 """Like getbundle, but taking a discovery.outgoing as an argument.
1976 1976
1977 1977 This is only implemented for local repos and reuses potentially
1978 1978 precomputed sets in outgoing."""
1979 1979 if not outgoing.missing:
1980 1980 return None
1981 1981 return self._changegroupsubset(outgoing.common,
1982 1982 outgoing.missing,
1983 1983 outgoing.missingheads,
1984 1984 source)
1985 1985
1986 1986 def getbundle(self, source, heads=None, common=None):
1987 1987 """Like changegroupsubset, but returns the set difference between the
1988 1988 ancestors of heads and the ancestors common.
1989 1989
1990 1990 If heads is None, use the local heads. If common is None, use [nullid].
1991 1991
1992 1992 The nodes in common might not all be known locally due to the way the
1993 1993 current discovery protocol works.
1994 1994 """
1995 1995 cl = self.changelog
1996 1996 if common:
1997 1997 hasnode = cl.hasnode
1998 1998 common = [n for n in common if hasnode(n)]
1999 1999 else:
2000 2000 common = [nullid]
2001 2001 if not heads:
2002 2002 heads = cl.heads()
2003 2003 return self.getlocalbundle(source,
2004 2004 discovery.outgoing(cl, common, heads))
2005 2005
2006 2006 @unfilteredmethod
2007 2007 def _changegroupsubset(self, commonrevs, csets, heads, source):
2008 2008
2009 2009 cl = self.changelog
2010 2010 mf = self.manifest
2011 2011 mfs = {} # needed manifests
2012 2012 fnodes = {} # needed file nodes
2013 2013 changedfiles = set()
2014 2014 fstate = ['', {}]
2015 2015 count = [0, 0]
2016 2016
2017 2017 # can we go through the fast path ?
2018 2018 heads.sort()
2019 2019 if heads == sorted(self.heads()):
2020 2020 return self._changegroup(csets, source)
2021 2021
2022 2022 # slow path
2023 2023 self.hook('preoutgoing', throw=True, source=source)
2024 2024 self.changegroupinfo(csets, source)
2025 2025
2026 2026 # filter any nodes that claim to be part of the known set
2027 2027 def prune(revlog, missing):
2028 2028 rr, rl = revlog.rev, revlog.linkrev
2029 2029 return [n for n in missing
2030 2030 if rl(rr(n)) not in commonrevs]
2031 2031
2032 2032 progress = self.ui.progress
2033 2033 _bundling = _('bundling')
2034 2034 _changesets = _('changesets')
2035 2035 _manifests = _('manifests')
2036 2036 _files = _('files')
2037 2037
2038 2038 def lookup(revlog, x):
2039 2039 if revlog == cl:
2040 2040 c = cl.read(x)
2041 2041 changedfiles.update(c[3])
2042 2042 mfs.setdefault(c[0], x)
2043 2043 count[0] += 1
2044 2044 progress(_bundling, count[0],
2045 2045 unit=_changesets, total=count[1])
2046 2046 return x
2047 2047 elif revlog == mf:
2048 2048 clnode = mfs[x]
2049 2049 mdata = mf.readfast(x)
2050 2050 for f, n in mdata.iteritems():
2051 2051 if f in changedfiles:
2052 2052 fnodes[f].setdefault(n, clnode)
2053 2053 count[0] += 1
2054 2054 progress(_bundling, count[0],
2055 2055 unit=_manifests, total=count[1])
2056 2056 return clnode
2057 2057 else:
2058 2058 progress(_bundling, count[0], item=fstate[0],
2059 2059 unit=_files, total=count[1])
2060 2060 return fstate[1][x]
2061 2061
2062 2062 bundler = changegroup.bundle10(lookup)
2063 2063 reorder = self.ui.config('bundle', 'reorder', 'auto')
2064 2064 if reorder == 'auto':
2065 2065 reorder = None
2066 2066 else:
2067 2067 reorder = util.parsebool(reorder)
2068 2068
2069 2069 def gengroup():
2070 2070 # Create a changenode group generator that will call our functions
2071 2071 # back to lookup the owning changenode and collect information.
2072 2072 count[:] = [0, len(csets)]
2073 2073 for chunk in cl.group(csets, bundler, reorder=reorder):
2074 2074 yield chunk
2075 2075 progress(_bundling, None)
2076 2076
2077 2077 # Create a generator for the manifestnodes that calls our lookup
2078 2078 # and data collection functions back.
2079 2079 for f in changedfiles:
2080 2080 fnodes[f] = {}
2081 2081 count[:] = [0, len(mfs)]
2082 2082 for chunk in mf.group(prune(mf, mfs), bundler, reorder=reorder):
2083 2083 yield chunk
2084 2084 progress(_bundling, None)
2085 2085
2086 2086 mfs.clear()
2087 2087
2088 2088 # Go through all our files in order sorted by name.
2089 2089 count[:] = [0, len(changedfiles)]
2090 2090 for fname in sorted(changedfiles):
2091 2091 filerevlog = self.file(fname)
2092 2092 if not len(filerevlog):
2093 2093 raise util.Abort(_("empty or missing revlog for %s")
2094 2094 % fname)
2095 2095 fstate[0] = fname
2096 2096 fstate[1] = fnodes.pop(fname, {})
2097 2097
2098 2098 nodelist = prune(filerevlog, fstate[1])
2099 2099 if nodelist:
2100 2100 count[0] += 1
2101 2101 yield bundler.fileheader(fname)
2102 2102 for chunk in filerevlog.group(nodelist, bundler, reorder):
2103 2103 yield chunk
2104 2104
2105 2105 # Signal that no more groups are left.
2106 2106 yield bundler.close()
2107 2107 progress(_bundling, None)
2108 2108
2109 2109 if csets:
2110 2110 self.hook('outgoing', node=hex(csets[0]), source=source)
2111 2111
2112 2112 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2113 2113
2114 2114 def changegroup(self, basenodes, source):
2115 2115 # to avoid a race we use changegroupsubset() (issue1320)
2116 2116 return self.changegroupsubset(basenodes, self.heads(), source)
2117 2117
2118 2118 @unfilteredmethod
2119 2119 def _changegroup(self, nodes, source):
2120 2120 """Compute the changegroup of all nodes that we have that a recipient
2121 2121 doesn't. Return a chunkbuffer object whose read() method will return
2122 2122 successive changegroup chunks.
2123 2123
2124 2124 This is much easier than the previous function as we can assume that
2125 2125 the recipient has any changenode we aren't sending them.
2126 2126
2127 2127 nodes is the set of nodes to send"""
2128 2128
2129 2129 cl = self.changelog
2130 2130 mf = self.manifest
2131 2131 mfs = {}
2132 2132 changedfiles = set()
2133 2133 fstate = ['']
2134 2134 count = [0, 0]
2135 2135
2136 2136 self.hook('preoutgoing', throw=True, source=source)
2137 2137 self.changegroupinfo(nodes, source)
2138 2138
2139 2139 revset = set([cl.rev(n) for n in nodes])
2140 2140
2141 2141 def gennodelst(log):
2142 2142 ln, llr = log.node, log.linkrev
2143 2143 return [ln(r) for r in log if llr(r) in revset]
2144 2144
2145 2145 progress = self.ui.progress
2146 2146 _bundling = _('bundling')
2147 2147 _changesets = _('changesets')
2148 2148 _manifests = _('manifests')
2149 2149 _files = _('files')
2150 2150
2151 2151 def lookup(revlog, x):
2152 2152 if revlog == cl:
2153 2153 c = cl.read(x)
2154 2154 changedfiles.update(c[3])
2155 2155 mfs.setdefault(c[0], x)
2156 2156 count[0] += 1
2157 2157 progress(_bundling, count[0],
2158 2158 unit=_changesets, total=count[1])
2159 2159 return x
2160 2160 elif revlog == mf:
2161 2161 count[0] += 1
2162 2162 progress(_bundling, count[0],
2163 2163 unit=_manifests, total=count[1])
2164 2164 return cl.node(revlog.linkrev(revlog.rev(x)))
2165 2165 else:
2166 2166 progress(_bundling, count[0], item=fstate[0],
2167 2167 total=count[1], unit=_files)
2168 2168 return cl.node(revlog.linkrev(revlog.rev(x)))
2169 2169
2170 2170 bundler = changegroup.bundle10(lookup)
2171 2171 reorder = self.ui.config('bundle', 'reorder', 'auto')
2172 2172 if reorder == 'auto':
2173 2173 reorder = None
2174 2174 else:
2175 2175 reorder = util.parsebool(reorder)
2176 2176
2177 2177 def gengroup():
2178 2178 '''yield a sequence of changegroup chunks (strings)'''
2179 2179 # construct a list of all changed files
2180 2180
2181 2181 count[:] = [0, len(nodes)]
2182 2182 for chunk in cl.group(nodes, bundler, reorder=reorder):
2183 2183 yield chunk
2184 2184 progress(_bundling, None)
2185 2185
2186 2186 count[:] = [0, len(mfs)]
2187 2187 for chunk in mf.group(gennodelst(mf), bundler, reorder=reorder):
2188 2188 yield chunk
2189 2189 progress(_bundling, None)
2190 2190
2191 2191 count[:] = [0, len(changedfiles)]
2192 2192 for fname in sorted(changedfiles):
2193 2193 filerevlog = self.file(fname)
2194 2194 if not len(filerevlog):
2195 2195 raise util.Abort(_("empty or missing revlog for %s")
2196 2196 % fname)
2197 2197 fstate[0] = fname
2198 2198 nodelist = gennodelst(filerevlog)
2199 2199 if nodelist:
2200 2200 count[0] += 1
2201 2201 yield bundler.fileheader(fname)
2202 2202 for chunk in filerevlog.group(nodelist, bundler, reorder):
2203 2203 yield chunk
2204 2204 yield bundler.close()
2205 2205 progress(_bundling, None)
2206 2206
2207 2207 if nodes:
2208 2208 self.hook('outgoing', node=hex(nodes[0]), source=source)
2209 2209
2210 2210 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2211 2211
2212 2212 @unfilteredmethod
2213 2213 def addchangegroup(self, source, srctype, url, emptyok=False):
2214 2214 """Add the changegroup returned by source.read() to this repo.
2215 2215 srctype is a string like 'push', 'pull', or 'unbundle'. url is
2216 2216 the URL of the repo where this changegroup is coming from.
2217 2217
2218 2218 Return an integer summarizing the change to this repo:
2219 2219 - nothing changed or no source: 0
2220 2220 - more heads than before: 1+added heads (2..n)
2221 2221 - fewer heads than before: -1-removed heads (-2..-n)
2222 2222 - number of heads stays the same: 1
2223 2223 """
2224 2224 def csmap(x):
2225 2225 self.ui.debug("add changeset %s\n" % short(x))
2226 2226 return len(cl)
2227 2227
2228 2228 def revmap(x):
2229 2229 return cl.rev(x)
2230 2230
2231 2231 if not source:
2232 2232 return 0
2233 2233
2234 2234 self.hook('prechangegroup', throw=True, source=srctype, url=url)
2235 2235
2236 2236 changesets = files = revisions = 0
2237 2237 efiles = set()
2238 2238
2239 2239 # write changelog data to temp files so concurrent readers will not see
2240 2240 # inconsistent view
2241 2241 cl = self.changelog
2242 2242 cl.delayupdate()
2243 2243 oldheads = cl.heads()
2244 2244
2245 2245 tr = self.transaction("\n".join([srctype, util.hidepassword(url)]))
2246 2246 try:
2247 2247 trp = weakref.proxy(tr)
2248 2248 # pull off the changeset group
2249 2249 self.ui.status(_("adding changesets\n"))
2250 2250 clstart = len(cl)
2251 2251 class prog(object):
2252 2252 step = _('changesets')
2253 2253 count = 1
2254 2254 ui = self.ui
2255 2255 total = None
2256 2256 def __call__(self):
2257 2257 self.ui.progress(self.step, self.count, unit=_('chunks'),
2258 2258 total=self.total)
2259 2259 self.count += 1
2260 2260 pr = prog()
2261 2261 source.callback = pr
2262 2262
2263 2263 source.changelogheader()
2264 2264 srccontent = cl.addgroup(source, csmap, trp)
2265 2265 if not (srccontent or emptyok):
2266 2266 raise util.Abort(_("received changelog group is empty"))
2267 2267 clend = len(cl)
2268 2268 changesets = clend - clstart
2269 2269 for c in xrange(clstart, clend):
2270 2270 efiles.update(self[c].files())
2271 2271 efiles = len(efiles)
2272 2272 self.ui.progress(_('changesets'), None)
2273 2273
2274 2274 # pull off the manifest group
2275 2275 self.ui.status(_("adding manifests\n"))
2276 2276 pr.step = _('manifests')
2277 2277 pr.count = 1
2278 2278 pr.total = changesets # manifests <= changesets
2279 2279 # no need to check for empty manifest group here:
2280 2280 # if the result of the merge of 1 and 2 is the same in 3 and 4,
2281 2281 # no new manifest will be created and the manifest group will
2282 2282 # be empty during the pull
2283 2283 source.manifestheader()
2284 2284 self.manifest.addgroup(source, revmap, trp)
2285 2285 self.ui.progress(_('manifests'), None)
2286 2286
2287 2287 needfiles = {}
2288 2288 if self.ui.configbool('server', 'validate', default=False):
2289 2289 # validate incoming csets have their manifests
2290 2290 for cset in xrange(clstart, clend):
2291 2291 mfest = self.changelog.read(self.changelog.node(cset))[0]
2292 2292 mfest = self.manifest.readdelta(mfest)
2293 2293 # store file nodes we must see
2294 2294 for f, n in mfest.iteritems():
2295 2295 needfiles.setdefault(f, set()).add(n)
2296 2296
2297 2297 # process the files
2298 2298 self.ui.status(_("adding file changes\n"))
2299 2299 pr.step = _('files')
2300 2300 pr.count = 1
2301 2301 pr.total = efiles
2302 2302 source.callback = None
2303 2303
2304 2304 while True:
2305 2305 chunkdata = source.filelogheader()
2306 2306 if not chunkdata:
2307 2307 break
2308 2308 f = chunkdata["filename"]
2309 2309 self.ui.debug("adding %s revisions\n" % f)
2310 2310 pr()
2311 2311 fl = self.file(f)
2312 2312 o = len(fl)
2313 2313 if not fl.addgroup(source, revmap, trp):
2314 2314 raise util.Abort(_("received file revlog group is empty"))
2315 2315 revisions += len(fl) - o
2316 2316 files += 1
2317 2317 if f in needfiles:
2318 2318 needs = needfiles[f]
2319 2319 for new in xrange(o, len(fl)):
2320 2320 n = fl.node(new)
2321 2321 if n in needs:
2322 2322 needs.remove(n)
2323 2323 else:
2324 2324 raise util.Abort(
2325 2325 _("received spurious file revlog entry"))
2326 2326 if not needs:
2327 2327 del needfiles[f]
2328 2328 self.ui.progress(_('files'), None)
2329 2329
2330 2330 for f, needs in needfiles.iteritems():
2331 2331 fl = self.file(f)
2332 2332 for n in needs:
2333 2333 try:
2334 2334 fl.rev(n)
2335 2335 except error.LookupError:
2336 2336 raise util.Abort(
2337 2337 _('missing file data for %s:%s - run hg verify') %
2338 2338 (f, hex(n)))
2339 2339
2340 2340 dh = 0
2341 2341 if oldheads:
2342 2342 heads = cl.heads()
2343 2343 dh = len(heads) - len(oldheads)
2344 2344 for h in heads:
2345 2345 if h not in oldheads and self[h].closesbranch():
2346 2346 dh -= 1
2347 2347 htext = ""
2348 2348 if dh:
2349 2349 htext = _(" (%+d heads)") % dh
2350 2350
2351 2351 self.ui.status(_("added %d changesets"
2352 2352 " with %d changes to %d files%s\n")
2353 2353 % (changesets, revisions, files, htext))
2354 2354 self.invalidatevolatilesets()
2355 2355
2356 2356 if changesets > 0:
2357 2357 p = lambda: cl.writepending() and self.root or ""
2358 2358 self.hook('pretxnchangegroup', throw=True,
2359 2359 node=hex(cl.node(clstart)), source=srctype,
2360 2360 url=url, pending=p)
2361 2361
2362 2362 added = [cl.node(r) for r in xrange(clstart, clend)]
2363 2363 publishing = self.ui.configbool('phases', 'publish', True)
2364 2364 if srctype == 'push':
2365 2365 # Old server can not push the boundary themself.
2366 2366 # New server won't push the boundary if changeset already
2367 2367 # existed locally as secrete
2368 2368 #
2369 2369 # We should not use added here but the list of all change in
2370 2370 # the bundle
2371 2371 if publishing:
2372 2372 phases.advanceboundary(self, phases.public, srccontent)
2373 2373 else:
2374 2374 phases.advanceboundary(self, phases.draft, srccontent)
2375 2375 phases.retractboundary(self, phases.draft, added)
2376 2376 elif srctype != 'strip':
2377 2377 # publishing only alter behavior during push
2378 2378 #
2379 2379 # strip should not touch boundary at all
2380 2380 phases.retractboundary(self, phases.draft, added)
2381 2381
2382 2382 # make changelog see real files again
2383 2383 cl.finalize(trp)
2384 2384
2385 2385 tr.close()
2386 2386
2387 2387 if changesets > 0:
2388 2388 if srctype != 'strip':
2389 2389 # During strip, branchcache is invalid but coming call to
2390 2390 # `destroyed` will repair it.
2391 2391 # In other case we can safely update cache on disk.
2392 2392 branchmap.updatecache(self.filtered('served'))
2393 2393 def runhooks():
2394 2394 # forcefully update the on-disk branch cache
2395 2395 self.ui.debug("updating the branch cache\n")
2396 2396 self.hook("changegroup", node=hex(cl.node(clstart)),
2397 2397 source=srctype, url=url)
2398 2398
2399 2399 for n in added:
2400 2400 self.hook("incoming", node=hex(n), source=srctype,
2401 2401 url=url)
2402 2402
2403 2403 newheads = [h for h in self.heads() if h not in oldheads]
2404 2404 self.ui.log("incoming",
2405 _("%s incoming changes - new heads: %s\n"),
2405 "%s incoming changes - new heads: %s\n",
2406 2406 len(added),
2407 2407 ', '.join([hex(c[:6]) for c in newheads]))
2408 2408 self._afterlock(runhooks)
2409 2409
2410 2410 finally:
2411 2411 tr.release()
2412 2412 # never return 0 here:
2413 2413 if dh < 0:
2414 2414 return dh - 1
2415 2415 else:
2416 2416 return dh + 1
2417 2417
2418 2418 def stream_in(self, remote, requirements):
2419 2419 lock = self.lock()
2420 2420 try:
2421 2421 # Save remote branchmap. We will use it later
2422 2422 # to speed up branchcache creation
2423 2423 rbranchmap = None
2424 2424 if remote.capable("branchmap"):
2425 2425 rbranchmap = remote.branchmap()
2426 2426
2427 2427 fp = remote.stream_out()
2428 2428 l = fp.readline()
2429 2429 try:
2430 2430 resp = int(l)
2431 2431 except ValueError:
2432 2432 raise error.ResponseError(
2433 2433 _('unexpected response from remote server:'), l)
2434 2434 if resp == 1:
2435 2435 raise util.Abort(_('operation forbidden by server'))
2436 2436 elif resp == 2:
2437 2437 raise util.Abort(_('locking the remote repository failed'))
2438 2438 elif resp != 0:
2439 2439 raise util.Abort(_('the server sent an unknown error code'))
2440 2440 self.ui.status(_('streaming all changes\n'))
2441 2441 l = fp.readline()
2442 2442 try:
2443 2443 total_files, total_bytes = map(int, l.split(' ', 1))
2444 2444 except (ValueError, TypeError):
2445 2445 raise error.ResponseError(
2446 2446 _('unexpected response from remote server:'), l)
2447 2447 self.ui.status(_('%d files to transfer, %s of data\n') %
2448 2448 (total_files, util.bytecount(total_bytes)))
2449 2449 handled_bytes = 0
2450 2450 self.ui.progress(_('clone'), 0, total=total_bytes)
2451 2451 start = time.time()
2452 2452 for i in xrange(total_files):
2453 2453 # XXX doesn't support '\n' or '\r' in filenames
2454 2454 l = fp.readline()
2455 2455 try:
2456 2456 name, size = l.split('\0', 1)
2457 2457 size = int(size)
2458 2458 except (ValueError, TypeError):
2459 2459 raise error.ResponseError(
2460 2460 _('unexpected response from remote server:'), l)
2461 2461 if self.ui.debugflag:
2462 2462 self.ui.debug('adding %s (%s)\n' %
2463 2463 (name, util.bytecount(size)))
2464 2464 # for backwards compat, name was partially encoded
2465 2465 ofp = self.sopener(store.decodedir(name), 'w')
2466 2466 for chunk in util.filechunkiter(fp, limit=size):
2467 2467 handled_bytes += len(chunk)
2468 2468 self.ui.progress(_('clone'), handled_bytes,
2469 2469 total=total_bytes)
2470 2470 ofp.write(chunk)
2471 2471 ofp.close()
2472 2472 elapsed = time.time() - start
2473 2473 if elapsed <= 0:
2474 2474 elapsed = 0.001
2475 2475 self.ui.progress(_('clone'), None)
2476 2476 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
2477 2477 (util.bytecount(total_bytes), elapsed,
2478 2478 util.bytecount(total_bytes / elapsed)))
2479 2479
2480 2480 # new requirements = old non-format requirements +
2481 2481 # new format-related
2482 2482 # requirements from the streamed-in repository
2483 2483 requirements.update(set(self.requirements) - self.supportedformats)
2484 2484 self._applyrequirements(requirements)
2485 2485 self._writerequirements()
2486 2486
2487 2487 if rbranchmap:
2488 2488 rbheads = []
2489 2489 for bheads in rbranchmap.itervalues():
2490 2490 rbheads.extend(bheads)
2491 2491
2492 2492 if rbheads:
2493 2493 rtiprev = max((int(self.changelog.rev(node))
2494 2494 for node in rbheads))
2495 2495 cache = branchmap.branchcache(rbranchmap,
2496 2496 self[rtiprev].node(),
2497 2497 rtiprev)
2498 2498 # Try to stick it as low as possible
2499 2499 # filter above served are unlikely to be fetch from a clone
2500 2500 for candidate in ('base', 'immutable', 'served'):
2501 2501 rview = self.filtered(candidate)
2502 2502 if cache.validfor(rview):
2503 2503 self._branchcaches[candidate] = cache
2504 2504 cache.write(rview)
2505 2505 break
2506 2506 self.invalidate()
2507 2507 return len(self.heads()) + 1
2508 2508 finally:
2509 2509 lock.release()
2510 2510
2511 2511 def clone(self, remote, heads=[], stream=False):
2512 2512 '''clone remote repository.
2513 2513
2514 2514 keyword arguments:
2515 2515 heads: list of revs to clone (forces use of pull)
2516 2516 stream: use streaming clone if possible'''
2517 2517
2518 2518 # now, all clients that can request uncompressed clones can
2519 2519 # read repo formats supported by all servers that can serve
2520 2520 # them.
2521 2521
2522 2522 # if revlog format changes, client will have to check version
2523 2523 # and format flags on "stream" capability, and use
2524 2524 # uncompressed only if compatible.
2525 2525
2526 2526 if not stream:
2527 2527 # if the server explicitly prefers to stream (for fast LANs)
2528 2528 stream = remote.capable('stream-preferred')
2529 2529
2530 2530 if stream and not heads:
2531 2531 # 'stream' means remote revlog format is revlogv1 only
2532 2532 if remote.capable('stream'):
2533 2533 return self.stream_in(remote, set(('revlogv1',)))
2534 2534 # otherwise, 'streamreqs' contains the remote revlog format
2535 2535 streamreqs = remote.capable('streamreqs')
2536 2536 if streamreqs:
2537 2537 streamreqs = set(streamreqs.split(','))
2538 2538 # if we support it, stream in and adjust our requirements
2539 2539 if not streamreqs - self.supportedformats:
2540 2540 return self.stream_in(remote, streamreqs)
2541 2541 return self.pull(remote, heads)
2542 2542
2543 2543 def pushkey(self, namespace, key, old, new):
2544 2544 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
2545 2545 old=old, new=new)
2546 2546 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
2547 2547 ret = pushkey.push(self, namespace, key, old, new)
2548 2548 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
2549 2549 ret=ret)
2550 2550 return ret
2551 2551
2552 2552 def listkeys(self, namespace):
2553 2553 self.hook('prelistkeys', throw=True, namespace=namespace)
2554 2554 self.ui.debug('listing keys for "%s"\n' % namespace)
2555 2555 values = pushkey.list(self, namespace)
2556 2556 self.hook('listkeys', namespace=namespace, values=values)
2557 2557 return values
2558 2558
2559 2559 def debugwireargs(self, one, two, three=None, four=None, five=None):
2560 2560 '''used to test argument passing over the wire'''
2561 2561 return "%s %s %s %s %s" % (one, two, three, four, five)
2562 2562
2563 2563 def savecommitmessage(self, text):
2564 2564 fp = self.opener('last-message.txt', 'wb')
2565 2565 try:
2566 2566 fp.write(text)
2567 2567 finally:
2568 2568 fp.close()
2569 2569 return self.pathto(fp.name[len(self.root) + 1:])
2570 2570
2571 2571 # used to avoid circular references so destructors work
2572 2572 def aftertrans(files):
2573 2573 renamefiles = [tuple(t) for t in files]
2574 2574 def a():
2575 2575 for src, dest in renamefiles:
2576 2576 try:
2577 2577 util.rename(src, dest)
2578 2578 except OSError: # journal file does not yet exist
2579 2579 pass
2580 2580 return a
2581 2581
2582 2582 def undoname(fn):
2583 2583 base, name = os.path.split(fn)
2584 2584 assert name.startswith('journal')
2585 2585 return os.path.join(base, name.replace('journal', 'undo', 1))
2586 2586
2587 2587 def instance(ui, path, create):
2588 2588 return localrepository(ui, util.urllocalpath(path), create)
2589 2589
2590 2590 def islocal(path):
2591 2591 return True
General Comments 0
You need to be logged in to leave comments. Login now