##// END OF EJS Templates
patch: include newline at EOF in help text for interactive patch...
Martin von Zweigbergk -
r42153:95e4ae86 default
parent child Browse files
Show More
@@ -1,2850 +1,2851
1 1 # patch.py - patch file parsing routines
2 2 #
3 3 # Copyright 2006 Brendan Cully <brendan@kublai.com>
4 4 # Copyright 2007 Chris Mason <chris.mason@oracle.com>
5 5 #
6 6 # This software may be used and distributed according to the terms of the
7 7 # GNU General Public License version 2 or any later version.
8 8
9 9 from __future__ import absolute_import, print_function
10 10
11 11 import collections
12 12 import contextlib
13 13 import copy
14 14 import email
15 15 import errno
16 16 import hashlib
17 17 import os
18 18 import re
19 19 import shutil
20 20 import zlib
21 21
22 22 from .i18n import _
23 23 from .node import (
24 24 hex,
25 25 short,
26 26 )
27 27 from . import (
28 28 copies,
29 29 diffhelper,
30 30 diffutil,
31 31 encoding,
32 32 error,
33 33 mail,
34 34 mdiff,
35 35 pathutil,
36 36 pycompat,
37 37 scmutil,
38 38 similar,
39 39 util,
40 40 vfs as vfsmod,
41 41 )
42 42 from .utils import (
43 43 dateutil,
44 44 procutil,
45 45 stringutil,
46 46 )
47 47
48 48 stringio = util.stringio
49 49
50 50 gitre = re.compile(br'diff --git a/(.*) b/(.*)')
51 51 tabsplitter = re.compile(br'(\t+|[^\t]+)')
52 52 wordsplitter = re.compile(br'(\t+| +|[a-zA-Z0-9_\x80-\xff]+|'
53 53 b'[^ \ta-zA-Z0-9_\x80-\xff])')
54 54
55 55 PatchError = error.PatchError
56 56
57 57 # public functions
58 58
59 59 def split(stream):
60 60 '''return an iterator of individual patches from a stream'''
61 61 def isheader(line, inheader):
62 62 if inheader and line.startswith((' ', '\t')):
63 63 # continuation
64 64 return True
65 65 if line.startswith((' ', '-', '+')):
66 66 # diff line - don't check for header pattern in there
67 67 return False
68 68 l = line.split(': ', 1)
69 69 return len(l) == 2 and ' ' not in l[0]
70 70
71 71 def chunk(lines):
72 72 return stringio(''.join(lines))
73 73
74 74 def hgsplit(stream, cur):
75 75 inheader = True
76 76
77 77 for line in stream:
78 78 if not line.strip():
79 79 inheader = False
80 80 if not inheader and line.startswith('# HG changeset patch'):
81 81 yield chunk(cur)
82 82 cur = []
83 83 inheader = True
84 84
85 85 cur.append(line)
86 86
87 87 if cur:
88 88 yield chunk(cur)
89 89
90 90 def mboxsplit(stream, cur):
91 91 for line in stream:
92 92 if line.startswith('From '):
93 93 for c in split(chunk(cur[1:])):
94 94 yield c
95 95 cur = []
96 96
97 97 cur.append(line)
98 98
99 99 if cur:
100 100 for c in split(chunk(cur[1:])):
101 101 yield c
102 102
103 103 def mimesplit(stream, cur):
104 104 def msgfp(m):
105 105 fp = stringio()
106 106 g = email.Generator.Generator(fp, mangle_from_=False)
107 107 g.flatten(m)
108 108 fp.seek(0)
109 109 return fp
110 110
111 111 for line in stream:
112 112 cur.append(line)
113 113 c = chunk(cur)
114 114
115 115 m = mail.parse(c)
116 116 if not m.is_multipart():
117 117 yield msgfp(m)
118 118 else:
119 119 ok_types = ('text/plain', 'text/x-diff', 'text/x-patch')
120 120 for part in m.walk():
121 121 ct = part.get_content_type()
122 122 if ct not in ok_types:
123 123 continue
124 124 yield msgfp(part)
125 125
126 126 def headersplit(stream, cur):
127 127 inheader = False
128 128
129 129 for line in stream:
130 130 if not inheader and isheader(line, inheader):
131 131 yield chunk(cur)
132 132 cur = []
133 133 inheader = True
134 134 if inheader and not isheader(line, inheader):
135 135 inheader = False
136 136
137 137 cur.append(line)
138 138
139 139 if cur:
140 140 yield chunk(cur)
141 141
142 142 def remainder(cur):
143 143 yield chunk(cur)
144 144
145 145 class fiter(object):
146 146 def __init__(self, fp):
147 147 self.fp = fp
148 148
149 149 def __iter__(self):
150 150 return self
151 151
152 152 def next(self):
153 153 l = self.fp.readline()
154 154 if not l:
155 155 raise StopIteration
156 156 return l
157 157
158 158 __next__ = next
159 159
160 160 inheader = False
161 161 cur = []
162 162
163 163 mimeheaders = ['content-type']
164 164
165 165 if not util.safehasattr(stream, 'next'):
166 166 # http responses, for example, have readline but not next
167 167 stream = fiter(stream)
168 168
169 169 for line in stream:
170 170 cur.append(line)
171 171 if line.startswith('# HG changeset patch'):
172 172 return hgsplit(stream, cur)
173 173 elif line.startswith('From '):
174 174 return mboxsplit(stream, cur)
175 175 elif isheader(line, inheader):
176 176 inheader = True
177 177 if line.split(':', 1)[0].lower() in mimeheaders:
178 178 # let email parser handle this
179 179 return mimesplit(stream, cur)
180 180 elif line.startswith('--- ') and inheader:
181 181 # No evil headers seen by diff start, split by hand
182 182 return headersplit(stream, cur)
183 183 # Not enough info, keep reading
184 184
185 185 # if we are here, we have a very plain patch
186 186 return remainder(cur)
187 187
188 188 ## Some facility for extensible patch parsing:
189 189 # list of pairs ("header to match", "data key")
190 190 patchheadermap = [('Date', 'date'),
191 191 ('Branch', 'branch'),
192 192 ('Node ID', 'nodeid'),
193 193 ]
194 194
195 195 @contextlib.contextmanager
196 196 def extract(ui, fileobj):
197 197 '''extract patch from data read from fileobj.
198 198
199 199 patch can be a normal patch or contained in an email message.
200 200
201 201 return a dictionary. Standard keys are:
202 202 - filename,
203 203 - message,
204 204 - user,
205 205 - date,
206 206 - branch,
207 207 - node,
208 208 - p1,
209 209 - p2.
210 210 Any item can be missing from the dictionary. If filename is missing,
211 211 fileobj did not contain a patch. Caller must unlink filename when done.'''
212 212
213 213 fd, tmpname = pycompat.mkstemp(prefix='hg-patch-')
214 214 tmpfp = os.fdopen(fd, r'wb')
215 215 try:
216 216 yield _extract(ui, fileobj, tmpname, tmpfp)
217 217 finally:
218 218 tmpfp.close()
219 219 os.unlink(tmpname)
220 220
221 221 def _extract(ui, fileobj, tmpname, tmpfp):
222 222
223 223 # attempt to detect the start of a patch
224 224 # (this heuristic is borrowed from quilt)
225 225 diffre = re.compile(br'^(?:Index:[ \t]|diff[ \t]-|RCS file: |'
226 226 br'retrieving revision [0-9]+(\.[0-9]+)*$|'
227 227 br'---[ \t].*?^\+\+\+[ \t]|'
228 228 br'\*\*\*[ \t].*?^---[ \t])',
229 229 re.MULTILINE | re.DOTALL)
230 230
231 231 data = {}
232 232
233 233 msg = mail.parse(fileobj)
234 234
235 235 subject = msg[r'Subject'] and mail.headdecode(msg[r'Subject'])
236 236 data['user'] = msg[r'From'] and mail.headdecode(msg[r'From'])
237 237 if not subject and not data['user']:
238 238 # Not an email, restore parsed headers if any
239 239 subject = '\n'.join(': '.join(map(encoding.strtolocal, h))
240 240 for h in msg.items()) + '\n'
241 241
242 242 # should try to parse msg['Date']
243 243 parents = []
244 244
245 245 if subject:
246 246 if subject.startswith('[PATCH'):
247 247 pend = subject.find(']')
248 248 if pend >= 0:
249 249 subject = subject[pend + 1:].lstrip()
250 250 subject = re.sub(br'\n[ \t]+', ' ', subject)
251 251 ui.debug('Subject: %s\n' % subject)
252 252 if data['user']:
253 253 ui.debug('From: %s\n' % data['user'])
254 254 diffs_seen = 0
255 255 ok_types = ('text/plain', 'text/x-diff', 'text/x-patch')
256 256 message = ''
257 257 for part in msg.walk():
258 258 content_type = pycompat.bytestr(part.get_content_type())
259 259 ui.debug('Content-Type: %s\n' % content_type)
260 260 if content_type not in ok_types:
261 261 continue
262 262 payload = part.get_payload(decode=True)
263 263 m = diffre.search(payload)
264 264 if m:
265 265 hgpatch = False
266 266 hgpatchheader = False
267 267 ignoretext = False
268 268
269 269 ui.debug('found patch at byte %d\n' % m.start(0))
270 270 diffs_seen += 1
271 271 cfp = stringio()
272 272 for line in payload[:m.start(0)].splitlines():
273 273 if line.startswith('# HG changeset patch') and not hgpatch:
274 274 ui.debug('patch generated by hg export\n')
275 275 hgpatch = True
276 276 hgpatchheader = True
277 277 # drop earlier commit message content
278 278 cfp.seek(0)
279 279 cfp.truncate()
280 280 subject = None
281 281 elif hgpatchheader:
282 282 if line.startswith('# User '):
283 283 data['user'] = line[7:]
284 284 ui.debug('From: %s\n' % data['user'])
285 285 elif line.startswith("# Parent "):
286 286 parents.append(line[9:].lstrip())
287 287 elif line.startswith("# "):
288 288 for header, key in patchheadermap:
289 289 prefix = '# %s ' % header
290 290 if line.startswith(prefix):
291 291 data[key] = line[len(prefix):]
292 292 else:
293 293 hgpatchheader = False
294 294 elif line == '---':
295 295 ignoretext = True
296 296 if not hgpatchheader and not ignoretext:
297 297 cfp.write(line)
298 298 cfp.write('\n')
299 299 message = cfp.getvalue()
300 300 if tmpfp:
301 301 tmpfp.write(payload)
302 302 if not payload.endswith('\n'):
303 303 tmpfp.write('\n')
304 304 elif not diffs_seen and message and content_type == 'text/plain':
305 305 message += '\n' + payload
306 306
307 307 if subject and not message.startswith(subject):
308 308 message = '%s\n%s' % (subject, message)
309 309 data['message'] = message
310 310 tmpfp.close()
311 311 if parents:
312 312 data['p1'] = parents.pop(0)
313 313 if parents:
314 314 data['p2'] = parents.pop(0)
315 315
316 316 if diffs_seen:
317 317 data['filename'] = tmpname
318 318
319 319 return data
320 320
321 321 class patchmeta(object):
322 322 """Patched file metadata
323 323
324 324 'op' is the performed operation within ADD, DELETE, RENAME, MODIFY
325 325 or COPY. 'path' is patched file path. 'oldpath' is set to the
326 326 origin file when 'op' is either COPY or RENAME, None otherwise. If
327 327 file mode is changed, 'mode' is a tuple (islink, isexec) where
328 328 'islink' is True if the file is a symlink and 'isexec' is True if
329 329 the file is executable. Otherwise, 'mode' is None.
330 330 """
331 331 def __init__(self, path):
332 332 self.path = path
333 333 self.oldpath = None
334 334 self.mode = None
335 335 self.op = 'MODIFY'
336 336 self.binary = False
337 337
338 338 def setmode(self, mode):
339 339 islink = mode & 0o20000
340 340 isexec = mode & 0o100
341 341 self.mode = (islink, isexec)
342 342
343 343 def copy(self):
344 344 other = patchmeta(self.path)
345 345 other.oldpath = self.oldpath
346 346 other.mode = self.mode
347 347 other.op = self.op
348 348 other.binary = self.binary
349 349 return other
350 350
351 351 def _ispatchinga(self, afile):
352 352 if afile == '/dev/null':
353 353 return self.op == 'ADD'
354 354 return afile == 'a/' + (self.oldpath or self.path)
355 355
356 356 def _ispatchingb(self, bfile):
357 357 if bfile == '/dev/null':
358 358 return self.op == 'DELETE'
359 359 return bfile == 'b/' + self.path
360 360
361 361 def ispatching(self, afile, bfile):
362 362 return self._ispatchinga(afile) and self._ispatchingb(bfile)
363 363
364 364 def __repr__(self):
365 365 return r"<patchmeta %s %r>" % (self.op, self.path)
366 366
367 367 def readgitpatch(lr):
368 368 """extract git-style metadata about patches from <patchname>"""
369 369
370 370 # Filter patch for git information
371 371 gp = None
372 372 gitpatches = []
373 373 for line in lr:
374 374 line = line.rstrip(' \r\n')
375 375 if line.startswith('diff --git a/'):
376 376 m = gitre.match(line)
377 377 if m:
378 378 if gp:
379 379 gitpatches.append(gp)
380 380 dst = m.group(2)
381 381 gp = patchmeta(dst)
382 382 elif gp:
383 383 if line.startswith('--- '):
384 384 gitpatches.append(gp)
385 385 gp = None
386 386 continue
387 387 if line.startswith('rename from '):
388 388 gp.op = 'RENAME'
389 389 gp.oldpath = line[12:]
390 390 elif line.startswith('rename to '):
391 391 gp.path = line[10:]
392 392 elif line.startswith('copy from '):
393 393 gp.op = 'COPY'
394 394 gp.oldpath = line[10:]
395 395 elif line.startswith('copy to '):
396 396 gp.path = line[8:]
397 397 elif line.startswith('deleted file'):
398 398 gp.op = 'DELETE'
399 399 elif line.startswith('new file mode '):
400 400 gp.op = 'ADD'
401 401 gp.setmode(int(line[-6:], 8))
402 402 elif line.startswith('new mode '):
403 403 gp.setmode(int(line[-6:], 8))
404 404 elif line.startswith('GIT binary patch'):
405 405 gp.binary = True
406 406 if gp:
407 407 gitpatches.append(gp)
408 408
409 409 return gitpatches
410 410
411 411 class linereader(object):
412 412 # simple class to allow pushing lines back into the input stream
413 413 def __init__(self, fp):
414 414 self.fp = fp
415 415 self.buf = []
416 416
417 417 def push(self, line):
418 418 if line is not None:
419 419 self.buf.append(line)
420 420
421 421 def readline(self):
422 422 if self.buf:
423 423 l = self.buf[0]
424 424 del self.buf[0]
425 425 return l
426 426 return self.fp.readline()
427 427
428 428 def __iter__(self):
429 429 return iter(self.readline, '')
430 430
431 431 class abstractbackend(object):
432 432 def __init__(self, ui):
433 433 self.ui = ui
434 434
435 435 def getfile(self, fname):
436 436 """Return target file data and flags as a (data, (islink,
437 437 isexec)) tuple. Data is None if file is missing/deleted.
438 438 """
439 439 raise NotImplementedError
440 440
441 441 def setfile(self, fname, data, mode, copysource):
442 442 """Write data to target file fname and set its mode. mode is a
443 443 (islink, isexec) tuple. If data is None, the file content should
444 444 be left unchanged. If the file is modified after being copied,
445 445 copysource is set to the original file name.
446 446 """
447 447 raise NotImplementedError
448 448
449 449 def unlink(self, fname):
450 450 """Unlink target file."""
451 451 raise NotImplementedError
452 452
453 453 def writerej(self, fname, failed, total, lines):
454 454 """Write rejected lines for fname. total is the number of hunks
455 455 which failed to apply and total the total number of hunks for this
456 456 files.
457 457 """
458 458
459 459 def exists(self, fname):
460 460 raise NotImplementedError
461 461
462 462 def close(self):
463 463 raise NotImplementedError
464 464
465 465 class fsbackend(abstractbackend):
466 466 def __init__(self, ui, basedir):
467 467 super(fsbackend, self).__init__(ui)
468 468 self.opener = vfsmod.vfs(basedir)
469 469
470 470 def getfile(self, fname):
471 471 if self.opener.islink(fname):
472 472 return (self.opener.readlink(fname), (True, False))
473 473
474 474 isexec = False
475 475 try:
476 476 isexec = self.opener.lstat(fname).st_mode & 0o100 != 0
477 477 except OSError as e:
478 478 if e.errno != errno.ENOENT:
479 479 raise
480 480 try:
481 481 return (self.opener.read(fname), (False, isexec))
482 482 except IOError as e:
483 483 if e.errno != errno.ENOENT:
484 484 raise
485 485 return None, None
486 486
487 487 def setfile(self, fname, data, mode, copysource):
488 488 islink, isexec = mode
489 489 if data is None:
490 490 self.opener.setflags(fname, islink, isexec)
491 491 return
492 492 if islink:
493 493 self.opener.symlink(data, fname)
494 494 else:
495 495 self.opener.write(fname, data)
496 496 if isexec:
497 497 self.opener.setflags(fname, False, True)
498 498
499 499 def unlink(self, fname):
500 500 rmdir = self.ui.configbool('experimental', 'removeemptydirs')
501 501 self.opener.unlinkpath(fname, ignoremissing=True, rmdir=rmdir)
502 502
503 503 def writerej(self, fname, failed, total, lines):
504 504 fname = fname + ".rej"
505 505 self.ui.warn(
506 506 _("%d out of %d hunks FAILED -- saving rejects to file %s\n") %
507 507 (failed, total, fname))
508 508 fp = self.opener(fname, 'w')
509 509 fp.writelines(lines)
510 510 fp.close()
511 511
512 512 def exists(self, fname):
513 513 return self.opener.lexists(fname)
514 514
515 515 class workingbackend(fsbackend):
516 516 def __init__(self, ui, repo, similarity):
517 517 super(workingbackend, self).__init__(ui, repo.root)
518 518 self.repo = repo
519 519 self.similarity = similarity
520 520 self.removed = set()
521 521 self.changed = set()
522 522 self.copied = []
523 523
524 524 def _checkknown(self, fname):
525 525 if self.repo.dirstate[fname] == '?' and self.exists(fname):
526 526 raise PatchError(_('cannot patch %s: file is not tracked') % fname)
527 527
528 528 def setfile(self, fname, data, mode, copysource):
529 529 self._checkknown(fname)
530 530 super(workingbackend, self).setfile(fname, data, mode, copysource)
531 531 if copysource is not None:
532 532 self.copied.append((copysource, fname))
533 533 self.changed.add(fname)
534 534
535 535 def unlink(self, fname):
536 536 self._checkknown(fname)
537 537 super(workingbackend, self).unlink(fname)
538 538 self.removed.add(fname)
539 539 self.changed.add(fname)
540 540
541 541 def close(self):
542 542 wctx = self.repo[None]
543 543 changed = set(self.changed)
544 544 for src, dst in self.copied:
545 545 scmutil.dirstatecopy(self.ui, self.repo, wctx, src, dst)
546 546 if self.removed:
547 547 wctx.forget(sorted(self.removed))
548 548 for f in self.removed:
549 549 if f not in self.repo.dirstate:
550 550 # File was deleted and no longer belongs to the
551 551 # dirstate, it was probably marked added then
552 552 # deleted, and should not be considered by
553 553 # marktouched().
554 554 changed.discard(f)
555 555 if changed:
556 556 scmutil.marktouched(self.repo, changed, self.similarity)
557 557 return sorted(self.changed)
558 558
559 559 class filestore(object):
560 560 def __init__(self, maxsize=None):
561 561 self.opener = None
562 562 self.files = {}
563 563 self.created = 0
564 564 self.maxsize = maxsize
565 565 if self.maxsize is None:
566 566 self.maxsize = 4*(2**20)
567 567 self.size = 0
568 568 self.data = {}
569 569
570 570 def setfile(self, fname, data, mode, copied=None):
571 571 if self.maxsize < 0 or (len(data) + self.size) <= self.maxsize:
572 572 self.data[fname] = (data, mode, copied)
573 573 self.size += len(data)
574 574 else:
575 575 if self.opener is None:
576 576 root = pycompat.mkdtemp(prefix='hg-patch-')
577 577 self.opener = vfsmod.vfs(root)
578 578 # Avoid filename issues with these simple names
579 579 fn = '%d' % self.created
580 580 self.opener.write(fn, data)
581 581 self.created += 1
582 582 self.files[fname] = (fn, mode, copied)
583 583
584 584 def getfile(self, fname):
585 585 if fname in self.data:
586 586 return self.data[fname]
587 587 if not self.opener or fname not in self.files:
588 588 return None, None, None
589 589 fn, mode, copied = self.files[fname]
590 590 return self.opener.read(fn), mode, copied
591 591
592 592 def close(self):
593 593 if self.opener:
594 594 shutil.rmtree(self.opener.base)
595 595
596 596 class repobackend(abstractbackend):
597 597 def __init__(self, ui, repo, ctx, store):
598 598 super(repobackend, self).__init__(ui)
599 599 self.repo = repo
600 600 self.ctx = ctx
601 601 self.store = store
602 602 self.changed = set()
603 603 self.removed = set()
604 604 self.copied = {}
605 605
606 606 def _checkknown(self, fname):
607 607 if fname not in self.ctx:
608 608 raise PatchError(_('cannot patch %s: file is not tracked') % fname)
609 609
610 610 def getfile(self, fname):
611 611 try:
612 612 fctx = self.ctx[fname]
613 613 except error.LookupError:
614 614 return None, None
615 615 flags = fctx.flags()
616 616 return fctx.data(), ('l' in flags, 'x' in flags)
617 617
618 618 def setfile(self, fname, data, mode, copysource):
619 619 if copysource:
620 620 self._checkknown(copysource)
621 621 if data is None:
622 622 data = self.ctx[fname].data()
623 623 self.store.setfile(fname, data, mode, copysource)
624 624 self.changed.add(fname)
625 625 if copysource:
626 626 self.copied[fname] = copysource
627 627
628 628 def unlink(self, fname):
629 629 self._checkknown(fname)
630 630 self.removed.add(fname)
631 631
632 632 def exists(self, fname):
633 633 return fname in self.ctx
634 634
635 635 def close(self):
636 636 return self.changed | self.removed
637 637
638 638 # @@ -start,len +start,len @@ or @@ -start +start @@ if len is 1
639 639 unidesc = re.compile(br'@@ -(\d+)(?:,(\d+))? \+(\d+)(?:,(\d+))? @@')
640 640 contextdesc = re.compile(br'(?:---|\*\*\*) (\d+)(?:,(\d+))? (?:---|\*\*\*)')
641 641 eolmodes = ['strict', 'crlf', 'lf', 'auto']
642 642
643 643 class patchfile(object):
644 644 def __init__(self, ui, gp, backend, store, eolmode='strict'):
645 645 self.fname = gp.path
646 646 self.eolmode = eolmode
647 647 self.eol = None
648 648 self.backend = backend
649 649 self.ui = ui
650 650 self.lines = []
651 651 self.exists = False
652 652 self.missing = True
653 653 self.mode = gp.mode
654 654 self.copysource = gp.oldpath
655 655 self.create = gp.op in ('ADD', 'COPY', 'RENAME')
656 656 self.remove = gp.op == 'DELETE'
657 657 if self.copysource is None:
658 658 data, mode = backend.getfile(self.fname)
659 659 else:
660 660 data, mode = store.getfile(self.copysource)[:2]
661 661 if data is not None:
662 662 self.exists = self.copysource is None or backend.exists(self.fname)
663 663 self.missing = False
664 664 if data:
665 665 self.lines = mdiff.splitnewlines(data)
666 666 if self.mode is None:
667 667 self.mode = mode
668 668 if self.lines:
669 669 # Normalize line endings
670 670 if self.lines[0].endswith('\r\n'):
671 671 self.eol = '\r\n'
672 672 elif self.lines[0].endswith('\n'):
673 673 self.eol = '\n'
674 674 if eolmode != 'strict':
675 675 nlines = []
676 676 for l in self.lines:
677 677 if l.endswith('\r\n'):
678 678 l = l[:-2] + '\n'
679 679 nlines.append(l)
680 680 self.lines = nlines
681 681 else:
682 682 if self.create:
683 683 self.missing = False
684 684 if self.mode is None:
685 685 self.mode = (False, False)
686 686 if self.missing:
687 687 self.ui.warn(_("unable to find '%s' for patching\n") % self.fname)
688 688 self.ui.warn(_("(use '--prefix' to apply patch relative to the "
689 689 "current directory)\n"))
690 690
691 691 self.hash = {}
692 692 self.dirty = 0
693 693 self.offset = 0
694 694 self.skew = 0
695 695 self.rej = []
696 696 self.fileprinted = False
697 697 self.printfile(False)
698 698 self.hunks = 0
699 699
700 700 def writelines(self, fname, lines, mode):
701 701 if self.eolmode == 'auto':
702 702 eol = self.eol
703 703 elif self.eolmode == 'crlf':
704 704 eol = '\r\n'
705 705 else:
706 706 eol = '\n'
707 707
708 708 if self.eolmode != 'strict' and eol and eol != '\n':
709 709 rawlines = []
710 710 for l in lines:
711 711 if l and l.endswith('\n'):
712 712 l = l[:-1] + eol
713 713 rawlines.append(l)
714 714 lines = rawlines
715 715
716 716 self.backend.setfile(fname, ''.join(lines), mode, self.copysource)
717 717
718 718 def printfile(self, warn):
719 719 if self.fileprinted:
720 720 return
721 721 if warn or self.ui.verbose:
722 722 self.fileprinted = True
723 723 s = _("patching file %s\n") % self.fname
724 724 if warn:
725 725 self.ui.warn(s)
726 726 else:
727 727 self.ui.note(s)
728 728
729 729
730 730 def findlines(self, l, linenum):
731 731 # looks through the hash and finds candidate lines. The
732 732 # result is a list of line numbers sorted based on distance
733 733 # from linenum
734 734
735 735 cand = self.hash.get(l, [])
736 736 if len(cand) > 1:
737 737 # resort our list of potentials forward then back.
738 738 cand.sort(key=lambda x: abs(x - linenum))
739 739 return cand
740 740
741 741 def write_rej(self):
742 742 # our rejects are a little different from patch(1). This always
743 743 # creates rejects in the same form as the original patch. A file
744 744 # header is inserted so that you can run the reject through patch again
745 745 # without having to type the filename.
746 746 if not self.rej:
747 747 return
748 748 base = os.path.basename(self.fname)
749 749 lines = ["--- %s\n+++ %s\n" % (base, base)]
750 750 for x in self.rej:
751 751 for l in x.hunk:
752 752 lines.append(l)
753 753 if l[-1:] != '\n':
754 754 lines.append("\n\\ No newline at end of file\n")
755 755 self.backend.writerej(self.fname, len(self.rej), self.hunks, lines)
756 756
757 757 def apply(self, h):
758 758 if not h.complete():
759 759 raise PatchError(_("bad hunk #%d %s (%d %d %d %d)") %
760 760 (h.number, h.desc, len(h.a), h.lena, len(h.b),
761 761 h.lenb))
762 762
763 763 self.hunks += 1
764 764
765 765 if self.missing:
766 766 self.rej.append(h)
767 767 return -1
768 768
769 769 if self.exists and self.create:
770 770 if self.copysource:
771 771 self.ui.warn(_("cannot create %s: destination already "
772 772 "exists\n") % self.fname)
773 773 else:
774 774 self.ui.warn(_("file %s already exists\n") % self.fname)
775 775 self.rej.append(h)
776 776 return -1
777 777
778 778 if isinstance(h, binhunk):
779 779 if self.remove:
780 780 self.backend.unlink(self.fname)
781 781 else:
782 782 l = h.new(self.lines)
783 783 self.lines[:] = l
784 784 self.offset += len(l)
785 785 self.dirty = True
786 786 return 0
787 787
788 788 horig = h
789 789 if (self.eolmode in ('crlf', 'lf')
790 790 or self.eolmode == 'auto' and self.eol):
791 791 # If new eols are going to be normalized, then normalize
792 792 # hunk data before patching. Otherwise, preserve input
793 793 # line-endings.
794 794 h = h.getnormalized()
795 795
796 796 # fast case first, no offsets, no fuzz
797 797 old, oldstart, new, newstart = h.fuzzit(0, False)
798 798 oldstart += self.offset
799 799 orig_start = oldstart
800 800 # if there's skew we want to emit the "(offset %d lines)" even
801 801 # when the hunk cleanly applies at start + skew, so skip the
802 802 # fast case code
803 803 if self.skew == 0 and diffhelper.testhunk(old, self.lines, oldstart):
804 804 if self.remove:
805 805 self.backend.unlink(self.fname)
806 806 else:
807 807 self.lines[oldstart:oldstart + len(old)] = new
808 808 self.offset += len(new) - len(old)
809 809 self.dirty = True
810 810 return 0
811 811
812 812 # ok, we couldn't match the hunk. Lets look for offsets and fuzz it
813 813 self.hash = {}
814 814 for x, s in enumerate(self.lines):
815 815 self.hash.setdefault(s, []).append(x)
816 816
817 817 for fuzzlen in pycompat.xrange(self.ui.configint("patch", "fuzz") + 1):
818 818 for toponly in [True, False]:
819 819 old, oldstart, new, newstart = h.fuzzit(fuzzlen, toponly)
820 820 oldstart = oldstart + self.offset + self.skew
821 821 oldstart = min(oldstart, len(self.lines))
822 822 if old:
823 823 cand = self.findlines(old[0][1:], oldstart)
824 824 else:
825 825 # Only adding lines with no or fuzzed context, just
826 826 # take the skew in account
827 827 cand = [oldstart]
828 828
829 829 for l in cand:
830 830 if not old or diffhelper.testhunk(old, self.lines, l):
831 831 self.lines[l : l + len(old)] = new
832 832 self.offset += len(new) - len(old)
833 833 self.skew = l - orig_start
834 834 self.dirty = True
835 835 offset = l - orig_start - fuzzlen
836 836 if fuzzlen:
837 837 msg = _("Hunk #%d succeeded at %d "
838 838 "with fuzz %d "
839 839 "(offset %d lines).\n")
840 840 self.printfile(True)
841 841 self.ui.warn(msg %
842 842 (h.number, l + 1, fuzzlen, offset))
843 843 else:
844 844 msg = _("Hunk #%d succeeded at %d "
845 845 "(offset %d lines).\n")
846 846 self.ui.note(msg % (h.number, l + 1, offset))
847 847 return fuzzlen
848 848 self.printfile(True)
849 849 self.ui.warn(_("Hunk #%d FAILED at %d\n") % (h.number, orig_start))
850 850 self.rej.append(horig)
851 851 return -1
852 852
853 853 def close(self):
854 854 if self.dirty:
855 855 self.writelines(self.fname, self.lines, self.mode)
856 856 self.write_rej()
857 857 return len(self.rej)
858 858
859 859 class header(object):
860 860 """patch header
861 861 """
862 862 diffgit_re = re.compile('diff --git a/(.*) b/(.*)$')
863 863 diff_re = re.compile('diff -r .* (.*)$')
864 864 allhunks_re = re.compile('(?:index|deleted file) ')
865 865 pretty_re = re.compile('(?:new file|deleted file) ')
866 866 special_re = re.compile('(?:index|deleted|copy|rename|new mode) ')
867 867 newfile_re = re.compile('(?:new file)')
868 868
869 869 def __init__(self, header):
870 870 self.header = header
871 871 self.hunks = []
872 872
873 873 def binary(self):
874 874 return any(h.startswith('index ') for h in self.header)
875 875
876 876 def pretty(self, fp):
877 877 for h in self.header:
878 878 if h.startswith('index '):
879 879 fp.write(_('this modifies a binary file (all or nothing)\n'))
880 880 break
881 881 if self.pretty_re.match(h):
882 882 fp.write(h)
883 883 if self.binary():
884 884 fp.write(_('this is a binary file\n'))
885 885 break
886 886 if h.startswith('---'):
887 887 fp.write(_('%d hunks, %d lines changed\n') %
888 888 (len(self.hunks),
889 889 sum([max(h.added, h.removed) for h in self.hunks])))
890 890 break
891 891 fp.write(h)
892 892
893 893 def write(self, fp):
894 894 fp.write(''.join(self.header))
895 895
896 896 def allhunks(self):
897 897 return any(self.allhunks_re.match(h) for h in self.header)
898 898
899 899 def files(self):
900 900 match = self.diffgit_re.match(self.header[0])
901 901 if match:
902 902 fromfile, tofile = match.groups()
903 903 if fromfile == tofile:
904 904 return [fromfile]
905 905 return [fromfile, tofile]
906 906 else:
907 907 return self.diff_re.match(self.header[0]).groups()
908 908
909 909 def filename(self):
910 910 return self.files()[-1]
911 911
912 912 def __repr__(self):
913 913 return '<header %s>' % (' '.join(map(repr, self.files())))
914 914
915 915 def isnewfile(self):
916 916 return any(self.newfile_re.match(h) for h in self.header)
917 917
918 918 def special(self):
919 919 # Special files are shown only at the header level and not at the hunk
920 920 # level for example a file that has been deleted is a special file.
921 921 # The user cannot change the content of the operation, in the case of
922 922 # the deleted file he has to take the deletion or not take it, he
923 923 # cannot take some of it.
924 924 # Newly added files are special if they are empty, they are not special
925 925 # if they have some content as we want to be able to change it
926 926 nocontent = len(self.header) == 2
927 927 emptynewfile = self.isnewfile() and nocontent
928 928 return (emptynewfile
929 929 or any(self.special_re.match(h) for h in self.header))
930 930
931 931 class recordhunk(object):
932 932 """patch hunk
933 933
934 934 XXX shouldn't we merge this with the other hunk class?
935 935 """
936 936
937 937 def __init__(self, header, fromline, toline, proc, before, hunk, after,
938 938 maxcontext=None):
939 939 def trimcontext(lines, reverse=False):
940 940 if maxcontext is not None:
941 941 delta = len(lines) - maxcontext
942 942 if delta > 0:
943 943 if reverse:
944 944 return delta, lines[delta:]
945 945 else:
946 946 return delta, lines[:maxcontext]
947 947 return 0, lines
948 948
949 949 self.header = header
950 950 trimedbefore, self.before = trimcontext(before, True)
951 951 self.fromline = fromline + trimedbefore
952 952 self.toline = toline + trimedbefore
953 953 _trimedafter, self.after = trimcontext(after, False)
954 954 self.proc = proc
955 955 self.hunk = hunk
956 956 self.added, self.removed = self.countchanges(self.hunk)
957 957
958 958 def __eq__(self, v):
959 959 if not isinstance(v, recordhunk):
960 960 return False
961 961
962 962 return ((v.hunk == self.hunk) and
963 963 (v.proc == self.proc) and
964 964 (self.fromline == v.fromline) and
965 965 (self.header.files() == v.header.files()))
966 966
967 967 def __hash__(self):
968 968 return hash((tuple(self.hunk),
969 969 tuple(self.header.files()),
970 970 self.fromline,
971 971 self.proc))
972 972
973 973 def countchanges(self, hunk):
974 974 """hunk -> (n+,n-)"""
975 975 add = len([h for h in hunk if h.startswith('+')])
976 976 rem = len([h for h in hunk if h.startswith('-')])
977 977 return add, rem
978 978
979 979 def reversehunk(self):
980 980 """return another recordhunk which is the reverse of the hunk
981 981
982 982 If this hunk is diff(A, B), the returned hunk is diff(B, A). To do
983 983 that, swap fromline/toline and +/- signs while keep other things
984 984 unchanged.
985 985 """
986 986 m = {'+': '-', '-': '+', '\\': '\\'}
987 987 hunk = ['%s%s' % (m[l[0:1]], l[1:]) for l in self.hunk]
988 988 return recordhunk(self.header, self.toline, self.fromline, self.proc,
989 989 self.before, hunk, self.after)
990 990
991 991 def write(self, fp):
992 992 delta = len(self.before) + len(self.after)
993 993 if self.after and self.after[-1] == '\\ No newline at end of file\n':
994 994 delta -= 1
995 995 fromlen = delta + self.removed
996 996 tolen = delta + self.added
997 997 fp.write('@@ -%d,%d +%d,%d @@%s\n' %
998 998 (self.fromline, fromlen, self.toline, tolen,
999 999 self.proc and (' ' + self.proc)))
1000 1000 fp.write(''.join(self.before + self.hunk + self.after))
1001 1001
1002 1002 pretty = write
1003 1003
1004 1004 def filename(self):
1005 1005 return self.header.filename()
1006 1006
1007 1007 def __repr__(self):
1008 1008 return '<hunk %r@%d>' % (self.filename(), self.fromline)
1009 1009
1010 1010 def getmessages():
1011 1011 return {
1012 1012 'multiple': {
1013 1013 'apply': _("apply change %d/%d to '%s'?"),
1014 1014 'discard': _("discard change %d/%d to '%s'?"),
1015 1015 'record': _("record change %d/%d to '%s'?"),
1016 1016 },
1017 1017 'single': {
1018 1018 'apply': _("apply this change to '%s'?"),
1019 1019 'discard': _("discard this change to '%s'?"),
1020 1020 'record': _("record this change to '%s'?"),
1021 1021 },
1022 1022 'help': {
1023 1023 'apply': _('[Ynesfdaq?]'
1024 1024 '$$ &Yes, apply this change'
1025 1025 '$$ &No, skip this change'
1026 1026 '$$ &Edit this change manually'
1027 1027 '$$ &Skip remaining changes to this file'
1028 1028 '$$ Apply remaining changes to this &file'
1029 1029 '$$ &Done, skip remaining changes and files'
1030 1030 '$$ Apply &all changes to all remaining files'
1031 1031 '$$ &Quit, applying no changes'
1032 1032 '$$ &? (display help)'),
1033 1033 'discard': _('[Ynesfdaq?]'
1034 1034 '$$ &Yes, discard this change'
1035 1035 '$$ &No, skip this change'
1036 1036 '$$ &Edit this change manually'
1037 1037 '$$ &Skip remaining changes to this file'
1038 1038 '$$ Discard remaining changes to this &file'
1039 1039 '$$ &Done, skip remaining changes and files'
1040 1040 '$$ Discard &all changes to all remaining files'
1041 1041 '$$ &Quit, discarding no changes'
1042 1042 '$$ &? (display help)'),
1043 1043 'record': _('[Ynesfdaq?]'
1044 1044 '$$ &Yes, record this change'
1045 1045 '$$ &No, skip this change'
1046 1046 '$$ &Edit this change manually'
1047 1047 '$$ &Skip remaining changes to this file'
1048 1048 '$$ Record remaining changes to this &file'
1049 1049 '$$ &Done, skip remaining changes and files'
1050 1050 '$$ Record &all changes to all remaining files'
1051 1051 '$$ &Quit, recording no changes'
1052 1052 '$$ &? (display help)'),
1053 1053 }
1054 1054 }
1055 1055
1056 1056 def filterpatch(ui, headers, operation=None):
1057 1057 """Interactively filter patch chunks into applied-only chunks"""
1058 1058 messages = getmessages()
1059 1059
1060 1060 if operation is None:
1061 1061 operation = 'record'
1062 1062
1063 1063 def prompt(skipfile, skipall, query, chunk):
1064 1064 """prompt query, and process base inputs
1065 1065
1066 1066 - y/n for the rest of file
1067 1067 - y/n for the rest
1068 1068 - ? (help)
1069 1069 - q (quit)
1070 1070
1071 1071 Return True/False and possibly updated skipfile and skipall.
1072 1072 """
1073 1073 newpatches = None
1074 1074 if skipall is not None:
1075 1075 return skipall, skipfile, skipall, newpatches
1076 1076 if skipfile is not None:
1077 1077 return skipfile, skipfile, skipall, newpatches
1078 1078 while True:
1079 1079 resps = messages['help'][operation]
1080 1080 r = ui.promptchoice("%s %s" % (query, resps))
1081 1081 ui.write("\n")
1082 1082 if r == 8: # ?
1083 1083 for c, t in ui.extractchoices(resps)[1]:
1084 1084 ui.write('%s - %s\n' % (c, encoding.lower(t)))
1085 1085 continue
1086 1086 elif r == 0: # yes
1087 1087 ret = True
1088 1088 elif r == 1: # no
1089 1089 ret = False
1090 1090 elif r == 2: # Edit patch
1091 1091 if chunk is None:
1092 1092 ui.write(_('cannot edit patch for whole file'))
1093 1093 ui.write("\n")
1094 1094 continue
1095 1095 if chunk.header.binary():
1096 1096 ui.write(_('cannot edit patch for binary file'))
1097 1097 ui.write("\n")
1098 1098 continue
1099 1099 # Patch comment based on the Git one (based on comment at end of
1100 1100 # https://mercurial-scm.org/wiki/RecordExtension)
1101 1101 phelp = '---' + _("""
1102 1102 To remove '-' lines, make them ' ' lines (context).
1103 1103 To remove '+' lines, delete them.
1104 1104 Lines starting with # will be removed from the patch.
1105 1105
1106 1106 If the patch applies cleanly, the edited hunk will immediately be
1107 1107 added to the record list. If it does not apply cleanly, a rejects
1108 1108 file will be generated: you can use that when you try again. If
1109 1109 all lines of the hunk are removed, then the edit is aborted and
1110 1110 the hunk is left unchanged.
1111 1111 """)
1112 1112 (patchfd, patchfn) = pycompat.mkstemp(prefix="hg-editor-",
1113 1113 suffix=".diff")
1114 1114 ncpatchfp = None
1115 1115 try:
1116 1116 # Write the initial patch
1117 1117 f = util.nativeeolwriter(os.fdopen(patchfd, r'wb'))
1118 1118 chunk.header.write(f)
1119 1119 chunk.write(f)
1120 f.write('\n'.join(['# ' + i for i in phelp.splitlines()]))
1120 f.write(''.join(['# ' + i + '\n'
1121 for i in phelp.splitlines()]))
1121 1122 f.close()
1122 1123 # Start the editor and wait for it to complete
1123 1124 editor = ui.geteditor()
1124 1125 ret = ui.system("%s \"%s\"" % (editor, patchfn),
1125 1126 environ={'HGUSER': ui.username()},
1126 1127 blockedtag='filterpatch')
1127 1128 if ret != 0:
1128 1129 ui.warn(_("editor exited with exit code %d\n") % ret)
1129 1130 continue
1130 1131 # Remove comment lines
1131 1132 patchfp = open(patchfn, r'rb')
1132 1133 ncpatchfp = stringio()
1133 1134 for line in util.iterfile(patchfp):
1134 1135 line = util.fromnativeeol(line)
1135 1136 if not line.startswith('#'):
1136 1137 ncpatchfp.write(line)
1137 1138 patchfp.close()
1138 1139 ncpatchfp.seek(0)
1139 1140 newpatches = parsepatch(ncpatchfp)
1140 1141 finally:
1141 1142 os.unlink(patchfn)
1142 1143 del ncpatchfp
1143 1144 # Signal that the chunk shouldn't be applied as-is, but
1144 1145 # provide the new patch to be used instead.
1145 1146 ret = False
1146 1147 elif r == 3: # Skip
1147 1148 ret = skipfile = False
1148 1149 elif r == 4: # file (Record remaining)
1149 1150 ret = skipfile = True
1150 1151 elif r == 5: # done, skip remaining
1151 1152 ret = skipall = False
1152 1153 elif r == 6: # all
1153 1154 ret = skipall = True
1154 1155 elif r == 7: # quit
1155 1156 raise error.Abort(_('user quit'))
1156 1157 return ret, skipfile, skipall, newpatches
1157 1158
1158 1159 seen = set()
1159 1160 applied = {} # 'filename' -> [] of chunks
1160 1161 skipfile, skipall = None, None
1161 1162 pos, total = 1, sum(len(h.hunks) for h in headers)
1162 1163 for h in headers:
1163 1164 pos += len(h.hunks)
1164 1165 skipfile = None
1165 1166 fixoffset = 0
1166 1167 hdr = ''.join(h.header)
1167 1168 if hdr in seen:
1168 1169 continue
1169 1170 seen.add(hdr)
1170 1171 if skipall is None:
1171 1172 h.pretty(ui)
1172 1173 msg = (_('examine changes to %s?') %
1173 1174 _(' and ').join("'%s'" % f for f in h.files()))
1174 1175 r, skipfile, skipall, np = prompt(skipfile, skipall, msg, None)
1175 1176 if not r:
1176 1177 continue
1177 1178 applied[h.filename()] = [h]
1178 1179 if h.allhunks():
1179 1180 applied[h.filename()] += h.hunks
1180 1181 continue
1181 1182 for i, chunk in enumerate(h.hunks):
1182 1183 if skipfile is None and skipall is None:
1183 1184 chunk.pretty(ui)
1184 1185 if total == 1:
1185 1186 msg = messages['single'][operation] % chunk.filename()
1186 1187 else:
1187 1188 idx = pos - len(h.hunks) + i
1188 1189 msg = messages['multiple'][operation] % (idx, total,
1189 1190 chunk.filename())
1190 1191 r, skipfile, skipall, newpatches = prompt(skipfile,
1191 1192 skipall, msg, chunk)
1192 1193 if r:
1193 1194 if fixoffset:
1194 1195 chunk = copy.copy(chunk)
1195 1196 chunk.toline += fixoffset
1196 1197 applied[chunk.filename()].append(chunk)
1197 1198 elif newpatches is not None:
1198 1199 for newpatch in newpatches:
1199 1200 for newhunk in newpatch.hunks:
1200 1201 if fixoffset:
1201 1202 newhunk.toline += fixoffset
1202 1203 applied[newhunk.filename()].append(newhunk)
1203 1204 else:
1204 1205 fixoffset += chunk.removed - chunk.added
1205 1206 return (sum([h for h in applied.itervalues()
1206 1207 if h[0].special() or len(h) > 1], []), {})
1207 1208 class hunk(object):
1208 1209 def __init__(self, desc, num, lr, context):
1209 1210 self.number = num
1210 1211 self.desc = desc
1211 1212 self.hunk = [desc]
1212 1213 self.a = []
1213 1214 self.b = []
1214 1215 self.starta = self.lena = None
1215 1216 self.startb = self.lenb = None
1216 1217 if lr is not None:
1217 1218 if context:
1218 1219 self.read_context_hunk(lr)
1219 1220 else:
1220 1221 self.read_unified_hunk(lr)
1221 1222
1222 1223 def getnormalized(self):
1223 1224 """Return a copy with line endings normalized to LF."""
1224 1225
1225 1226 def normalize(lines):
1226 1227 nlines = []
1227 1228 for line in lines:
1228 1229 if line.endswith('\r\n'):
1229 1230 line = line[:-2] + '\n'
1230 1231 nlines.append(line)
1231 1232 return nlines
1232 1233
1233 1234 # Dummy object, it is rebuilt manually
1234 1235 nh = hunk(self.desc, self.number, None, None)
1235 1236 nh.number = self.number
1236 1237 nh.desc = self.desc
1237 1238 nh.hunk = self.hunk
1238 1239 nh.a = normalize(self.a)
1239 1240 nh.b = normalize(self.b)
1240 1241 nh.starta = self.starta
1241 1242 nh.startb = self.startb
1242 1243 nh.lena = self.lena
1243 1244 nh.lenb = self.lenb
1244 1245 return nh
1245 1246
1246 1247 def read_unified_hunk(self, lr):
1247 1248 m = unidesc.match(self.desc)
1248 1249 if not m:
1249 1250 raise PatchError(_("bad hunk #%d") % self.number)
1250 1251 self.starta, self.lena, self.startb, self.lenb = m.groups()
1251 1252 if self.lena is None:
1252 1253 self.lena = 1
1253 1254 else:
1254 1255 self.lena = int(self.lena)
1255 1256 if self.lenb is None:
1256 1257 self.lenb = 1
1257 1258 else:
1258 1259 self.lenb = int(self.lenb)
1259 1260 self.starta = int(self.starta)
1260 1261 self.startb = int(self.startb)
1261 1262 try:
1262 1263 diffhelper.addlines(lr, self.hunk, self.lena, self.lenb,
1263 1264 self.a, self.b)
1264 1265 except error.ParseError as e:
1265 1266 raise PatchError(_("bad hunk #%d: %s") % (self.number, e))
1266 1267 # if we hit eof before finishing out the hunk, the last line will
1267 1268 # be zero length. Lets try to fix it up.
1268 1269 while len(self.hunk[-1]) == 0:
1269 1270 del self.hunk[-1]
1270 1271 del self.a[-1]
1271 1272 del self.b[-1]
1272 1273 self.lena -= 1
1273 1274 self.lenb -= 1
1274 1275 self._fixnewline(lr)
1275 1276
1276 1277 def read_context_hunk(self, lr):
1277 1278 self.desc = lr.readline()
1278 1279 m = contextdesc.match(self.desc)
1279 1280 if not m:
1280 1281 raise PatchError(_("bad hunk #%d") % self.number)
1281 1282 self.starta, aend = m.groups()
1282 1283 self.starta = int(self.starta)
1283 1284 if aend is None:
1284 1285 aend = self.starta
1285 1286 self.lena = int(aend) - self.starta
1286 1287 if self.starta:
1287 1288 self.lena += 1
1288 1289 for x in pycompat.xrange(self.lena):
1289 1290 l = lr.readline()
1290 1291 if l.startswith('---'):
1291 1292 # lines addition, old block is empty
1292 1293 lr.push(l)
1293 1294 break
1294 1295 s = l[2:]
1295 1296 if l.startswith('- ') or l.startswith('! '):
1296 1297 u = '-' + s
1297 1298 elif l.startswith(' '):
1298 1299 u = ' ' + s
1299 1300 else:
1300 1301 raise PatchError(_("bad hunk #%d old text line %d") %
1301 1302 (self.number, x))
1302 1303 self.a.append(u)
1303 1304 self.hunk.append(u)
1304 1305
1305 1306 l = lr.readline()
1306 1307 if l.startswith(br'\ '):
1307 1308 s = self.a[-1][:-1]
1308 1309 self.a[-1] = s
1309 1310 self.hunk[-1] = s
1310 1311 l = lr.readline()
1311 1312 m = contextdesc.match(l)
1312 1313 if not m:
1313 1314 raise PatchError(_("bad hunk #%d") % self.number)
1314 1315 self.startb, bend = m.groups()
1315 1316 self.startb = int(self.startb)
1316 1317 if bend is None:
1317 1318 bend = self.startb
1318 1319 self.lenb = int(bend) - self.startb
1319 1320 if self.startb:
1320 1321 self.lenb += 1
1321 1322 hunki = 1
1322 1323 for x in pycompat.xrange(self.lenb):
1323 1324 l = lr.readline()
1324 1325 if l.startswith(br'\ '):
1325 1326 # XXX: the only way to hit this is with an invalid line range.
1326 1327 # The no-eol marker is not counted in the line range, but I
1327 1328 # guess there are diff(1) out there which behave differently.
1328 1329 s = self.b[-1][:-1]
1329 1330 self.b[-1] = s
1330 1331 self.hunk[hunki - 1] = s
1331 1332 continue
1332 1333 if not l:
1333 1334 # line deletions, new block is empty and we hit EOF
1334 1335 lr.push(l)
1335 1336 break
1336 1337 s = l[2:]
1337 1338 if l.startswith('+ ') or l.startswith('! '):
1338 1339 u = '+' + s
1339 1340 elif l.startswith(' '):
1340 1341 u = ' ' + s
1341 1342 elif len(self.b) == 0:
1342 1343 # line deletions, new block is empty
1343 1344 lr.push(l)
1344 1345 break
1345 1346 else:
1346 1347 raise PatchError(_("bad hunk #%d old text line %d") %
1347 1348 (self.number, x))
1348 1349 self.b.append(s)
1349 1350 while True:
1350 1351 if hunki >= len(self.hunk):
1351 1352 h = ""
1352 1353 else:
1353 1354 h = self.hunk[hunki]
1354 1355 hunki += 1
1355 1356 if h == u:
1356 1357 break
1357 1358 elif h.startswith('-'):
1358 1359 continue
1359 1360 else:
1360 1361 self.hunk.insert(hunki - 1, u)
1361 1362 break
1362 1363
1363 1364 if not self.a:
1364 1365 # this happens when lines were only added to the hunk
1365 1366 for x in self.hunk:
1366 1367 if x.startswith('-') or x.startswith(' '):
1367 1368 self.a.append(x)
1368 1369 if not self.b:
1369 1370 # this happens when lines were only deleted from the hunk
1370 1371 for x in self.hunk:
1371 1372 if x.startswith('+') or x.startswith(' '):
1372 1373 self.b.append(x[1:])
1373 1374 # @@ -start,len +start,len @@
1374 1375 self.desc = "@@ -%d,%d +%d,%d @@\n" % (self.starta, self.lena,
1375 1376 self.startb, self.lenb)
1376 1377 self.hunk[0] = self.desc
1377 1378 self._fixnewline(lr)
1378 1379
1379 1380 def _fixnewline(self, lr):
1380 1381 l = lr.readline()
1381 1382 if l.startswith(br'\ '):
1382 1383 diffhelper.fixnewline(self.hunk, self.a, self.b)
1383 1384 else:
1384 1385 lr.push(l)
1385 1386
1386 1387 def complete(self):
1387 1388 return len(self.a) == self.lena and len(self.b) == self.lenb
1388 1389
1389 1390 def _fuzzit(self, old, new, fuzz, toponly):
1390 1391 # this removes context lines from the top and bottom of list 'l'. It
1391 1392 # checks the hunk to make sure only context lines are removed, and then
1392 1393 # returns a new shortened list of lines.
1393 1394 fuzz = min(fuzz, len(old))
1394 1395 if fuzz:
1395 1396 top = 0
1396 1397 bot = 0
1397 1398 hlen = len(self.hunk)
1398 1399 for x in pycompat.xrange(hlen - 1):
1399 1400 # the hunk starts with the @@ line, so use x+1
1400 1401 if self.hunk[x + 1].startswith(' '):
1401 1402 top += 1
1402 1403 else:
1403 1404 break
1404 1405 if not toponly:
1405 1406 for x in pycompat.xrange(hlen - 1):
1406 1407 if self.hunk[hlen - bot - 1].startswith(' '):
1407 1408 bot += 1
1408 1409 else:
1409 1410 break
1410 1411
1411 1412 bot = min(fuzz, bot)
1412 1413 top = min(fuzz, top)
1413 1414 return old[top:len(old) - bot], new[top:len(new) - bot], top
1414 1415 return old, new, 0
1415 1416
1416 1417 def fuzzit(self, fuzz, toponly):
1417 1418 old, new, top = self._fuzzit(self.a, self.b, fuzz, toponly)
1418 1419 oldstart = self.starta + top
1419 1420 newstart = self.startb + top
1420 1421 # zero length hunk ranges already have their start decremented
1421 1422 if self.lena and oldstart > 0:
1422 1423 oldstart -= 1
1423 1424 if self.lenb and newstart > 0:
1424 1425 newstart -= 1
1425 1426 return old, oldstart, new, newstart
1426 1427
1427 1428 class binhunk(object):
1428 1429 'A binary patch file.'
1429 1430 def __init__(self, lr, fname):
1430 1431 self.text = None
1431 1432 self.delta = False
1432 1433 self.hunk = ['GIT binary patch\n']
1433 1434 self._fname = fname
1434 1435 self._read(lr)
1435 1436
1436 1437 def complete(self):
1437 1438 return self.text is not None
1438 1439
1439 1440 def new(self, lines):
1440 1441 if self.delta:
1441 1442 return [applybindelta(self.text, ''.join(lines))]
1442 1443 return [self.text]
1443 1444
1444 1445 def _read(self, lr):
1445 1446 def getline(lr, hunk):
1446 1447 l = lr.readline()
1447 1448 hunk.append(l)
1448 1449 return l.rstrip('\r\n')
1449 1450
1450 1451 while True:
1451 1452 line = getline(lr, self.hunk)
1452 1453 if not line:
1453 1454 raise PatchError(_('could not extract "%s" binary data')
1454 1455 % self._fname)
1455 1456 if line.startswith('literal '):
1456 1457 size = int(line[8:].rstrip())
1457 1458 break
1458 1459 if line.startswith('delta '):
1459 1460 size = int(line[6:].rstrip())
1460 1461 self.delta = True
1461 1462 break
1462 1463 dec = []
1463 1464 line = getline(lr, self.hunk)
1464 1465 while len(line) > 1:
1465 1466 l = line[0:1]
1466 1467 if l <= 'Z' and l >= 'A':
1467 1468 l = ord(l) - ord('A') + 1
1468 1469 else:
1469 1470 l = ord(l) - ord('a') + 27
1470 1471 try:
1471 1472 dec.append(util.b85decode(line[1:])[:l])
1472 1473 except ValueError as e:
1473 1474 raise PatchError(_('could not decode "%s" binary patch: %s')
1474 1475 % (self._fname, stringutil.forcebytestr(e)))
1475 1476 line = getline(lr, self.hunk)
1476 1477 text = zlib.decompress(''.join(dec))
1477 1478 if len(text) != size:
1478 1479 raise PatchError(_('"%s" length is %d bytes, should be %d')
1479 1480 % (self._fname, len(text), size))
1480 1481 self.text = text
1481 1482
1482 1483 def parsefilename(str):
1483 1484 # --- filename \t|space stuff
1484 1485 s = str[4:].rstrip('\r\n')
1485 1486 i = s.find('\t')
1486 1487 if i < 0:
1487 1488 i = s.find(' ')
1488 1489 if i < 0:
1489 1490 return s
1490 1491 return s[:i]
1491 1492
1492 1493 def reversehunks(hunks):
1493 1494 '''reverse the signs in the hunks given as argument
1494 1495
1495 1496 This function operates on hunks coming out of patch.filterpatch, that is
1496 1497 a list of the form: [header1, hunk1, hunk2, header2...]. Example usage:
1497 1498
1498 1499 >>> rawpatch = b"""diff --git a/folder1/g b/folder1/g
1499 1500 ... --- a/folder1/g
1500 1501 ... +++ b/folder1/g
1501 1502 ... @@ -1,7 +1,7 @@
1502 1503 ... +firstline
1503 1504 ... c
1504 1505 ... 1
1505 1506 ... 2
1506 1507 ... + 3
1507 1508 ... -4
1508 1509 ... 5
1509 1510 ... d
1510 1511 ... +lastline"""
1511 1512 >>> hunks = parsepatch([rawpatch])
1512 1513 >>> hunkscomingfromfilterpatch = []
1513 1514 >>> for h in hunks:
1514 1515 ... hunkscomingfromfilterpatch.append(h)
1515 1516 ... hunkscomingfromfilterpatch.extend(h.hunks)
1516 1517
1517 1518 >>> reversedhunks = reversehunks(hunkscomingfromfilterpatch)
1518 1519 >>> from . import util
1519 1520 >>> fp = util.stringio()
1520 1521 >>> for c in reversedhunks:
1521 1522 ... c.write(fp)
1522 1523 >>> fp.seek(0) or None
1523 1524 >>> reversedpatch = fp.read()
1524 1525 >>> print(pycompat.sysstr(reversedpatch))
1525 1526 diff --git a/folder1/g b/folder1/g
1526 1527 --- a/folder1/g
1527 1528 +++ b/folder1/g
1528 1529 @@ -1,4 +1,3 @@
1529 1530 -firstline
1530 1531 c
1531 1532 1
1532 1533 2
1533 1534 @@ -2,6 +1,6 @@
1534 1535 c
1535 1536 1
1536 1537 2
1537 1538 - 3
1538 1539 +4
1539 1540 5
1540 1541 d
1541 1542 @@ -6,3 +5,2 @@
1542 1543 5
1543 1544 d
1544 1545 -lastline
1545 1546
1546 1547 '''
1547 1548
1548 1549 newhunks = []
1549 1550 for c in hunks:
1550 1551 if util.safehasattr(c, 'reversehunk'):
1551 1552 c = c.reversehunk()
1552 1553 newhunks.append(c)
1553 1554 return newhunks
1554 1555
1555 1556 def parsepatch(originalchunks, maxcontext=None):
1556 1557 """patch -> [] of headers -> [] of hunks
1557 1558
1558 1559 If maxcontext is not None, trim context lines if necessary.
1559 1560
1560 1561 >>> rawpatch = b'''diff --git a/folder1/g b/folder1/g
1561 1562 ... --- a/folder1/g
1562 1563 ... +++ b/folder1/g
1563 1564 ... @@ -1,8 +1,10 @@
1564 1565 ... 1
1565 1566 ... 2
1566 1567 ... -3
1567 1568 ... 4
1568 1569 ... 5
1569 1570 ... 6
1570 1571 ... +6.1
1571 1572 ... +6.2
1572 1573 ... 7
1573 1574 ... 8
1574 1575 ... +9'''
1575 1576 >>> out = util.stringio()
1576 1577 >>> headers = parsepatch([rawpatch], maxcontext=1)
1577 1578 >>> for header in headers:
1578 1579 ... header.write(out)
1579 1580 ... for hunk in header.hunks:
1580 1581 ... hunk.write(out)
1581 1582 >>> print(pycompat.sysstr(out.getvalue()))
1582 1583 diff --git a/folder1/g b/folder1/g
1583 1584 --- a/folder1/g
1584 1585 +++ b/folder1/g
1585 1586 @@ -2,3 +2,2 @@
1586 1587 2
1587 1588 -3
1588 1589 4
1589 1590 @@ -6,2 +5,4 @@
1590 1591 6
1591 1592 +6.1
1592 1593 +6.2
1593 1594 7
1594 1595 @@ -8,1 +9,2 @@
1595 1596 8
1596 1597 +9
1597 1598 """
1598 1599 class parser(object):
1599 1600 """patch parsing state machine"""
1600 1601 def __init__(self):
1601 1602 self.fromline = 0
1602 1603 self.toline = 0
1603 1604 self.proc = ''
1604 1605 self.header = None
1605 1606 self.context = []
1606 1607 self.before = []
1607 1608 self.hunk = []
1608 1609 self.headers = []
1609 1610
1610 1611 def addrange(self, limits):
1611 1612 self.addcontext([])
1612 1613 fromstart, fromend, tostart, toend, proc = limits
1613 1614 self.fromline = int(fromstart)
1614 1615 self.toline = int(tostart)
1615 1616 self.proc = proc
1616 1617
1617 1618 def addcontext(self, context):
1618 1619 if self.hunk:
1619 1620 h = recordhunk(self.header, self.fromline, self.toline,
1620 1621 self.proc, self.before, self.hunk, context, maxcontext)
1621 1622 self.header.hunks.append(h)
1622 1623 self.fromline += len(self.before) + h.removed
1623 1624 self.toline += len(self.before) + h.added
1624 1625 self.before = []
1625 1626 self.hunk = []
1626 1627 self.context = context
1627 1628
1628 1629 def addhunk(self, hunk):
1629 1630 if self.context:
1630 1631 self.before = self.context
1631 1632 self.context = []
1632 1633 if self.hunk:
1633 1634 self.addcontext([])
1634 1635 self.hunk = hunk
1635 1636
1636 1637 def newfile(self, hdr):
1637 1638 self.addcontext([])
1638 1639 h = header(hdr)
1639 1640 self.headers.append(h)
1640 1641 self.header = h
1641 1642
1642 1643 def addother(self, line):
1643 1644 pass # 'other' lines are ignored
1644 1645
1645 1646 def finished(self):
1646 1647 self.addcontext([])
1647 1648 return self.headers
1648 1649
1649 1650 transitions = {
1650 1651 'file': {'context': addcontext,
1651 1652 'file': newfile,
1652 1653 'hunk': addhunk,
1653 1654 'range': addrange},
1654 1655 'context': {'file': newfile,
1655 1656 'hunk': addhunk,
1656 1657 'range': addrange,
1657 1658 'other': addother},
1658 1659 'hunk': {'context': addcontext,
1659 1660 'file': newfile,
1660 1661 'range': addrange},
1661 1662 'range': {'context': addcontext,
1662 1663 'hunk': addhunk},
1663 1664 'other': {'other': addother},
1664 1665 }
1665 1666
1666 1667 p = parser()
1667 1668 fp = stringio()
1668 1669 fp.write(''.join(originalchunks))
1669 1670 fp.seek(0)
1670 1671
1671 1672 state = 'context'
1672 1673 for newstate, data in scanpatch(fp):
1673 1674 try:
1674 1675 p.transitions[state][newstate](p, data)
1675 1676 except KeyError:
1676 1677 raise PatchError('unhandled transition: %s -> %s' %
1677 1678 (state, newstate))
1678 1679 state = newstate
1679 1680 del fp
1680 1681 return p.finished()
1681 1682
1682 1683 def pathtransform(path, strip, prefix):
1683 1684 '''turn a path from a patch into a path suitable for the repository
1684 1685
1685 1686 prefix, if not empty, is expected to be normalized with a / at the end.
1686 1687
1687 1688 Returns (stripped components, path in repository).
1688 1689
1689 1690 >>> pathtransform(b'a/b/c', 0, b'')
1690 1691 ('', 'a/b/c')
1691 1692 >>> pathtransform(b' a/b/c ', 0, b'')
1692 1693 ('', ' a/b/c')
1693 1694 >>> pathtransform(b' a/b/c ', 2, b'')
1694 1695 ('a/b/', 'c')
1695 1696 >>> pathtransform(b'a/b/c', 0, b'd/e/')
1696 1697 ('', 'd/e/a/b/c')
1697 1698 >>> pathtransform(b' a//b/c ', 2, b'd/e/')
1698 1699 ('a//b/', 'd/e/c')
1699 1700 >>> pathtransform(b'a/b/c', 3, b'')
1700 1701 Traceback (most recent call last):
1701 1702 PatchError: unable to strip away 1 of 3 dirs from a/b/c
1702 1703 '''
1703 1704 pathlen = len(path)
1704 1705 i = 0
1705 1706 if strip == 0:
1706 1707 return '', prefix + path.rstrip()
1707 1708 count = strip
1708 1709 while count > 0:
1709 1710 i = path.find('/', i)
1710 1711 if i == -1:
1711 1712 raise PatchError(_("unable to strip away %d of %d dirs from %s") %
1712 1713 (count, strip, path))
1713 1714 i += 1
1714 1715 # consume '//' in the path
1715 1716 while i < pathlen - 1 and path[i:i + 1] == '/':
1716 1717 i += 1
1717 1718 count -= 1
1718 1719 return path[:i].lstrip(), prefix + path[i:].rstrip()
1719 1720
1720 1721 def makepatchmeta(backend, afile_orig, bfile_orig, hunk, strip, prefix):
1721 1722 nulla = afile_orig == "/dev/null"
1722 1723 nullb = bfile_orig == "/dev/null"
1723 1724 create = nulla and hunk.starta == 0 and hunk.lena == 0
1724 1725 remove = nullb and hunk.startb == 0 and hunk.lenb == 0
1725 1726 abase, afile = pathtransform(afile_orig, strip, prefix)
1726 1727 gooda = not nulla and backend.exists(afile)
1727 1728 bbase, bfile = pathtransform(bfile_orig, strip, prefix)
1728 1729 if afile == bfile:
1729 1730 goodb = gooda
1730 1731 else:
1731 1732 goodb = not nullb and backend.exists(bfile)
1732 1733 missing = not goodb and not gooda and not create
1733 1734
1734 1735 # some diff programs apparently produce patches where the afile is
1735 1736 # not /dev/null, but afile starts with bfile
1736 1737 abasedir = afile[:afile.rfind('/') + 1]
1737 1738 bbasedir = bfile[:bfile.rfind('/') + 1]
1738 1739 if (missing and abasedir == bbasedir and afile.startswith(bfile)
1739 1740 and hunk.starta == 0 and hunk.lena == 0):
1740 1741 create = True
1741 1742 missing = False
1742 1743
1743 1744 # If afile is "a/b/foo" and bfile is "a/b/foo.orig" we assume the
1744 1745 # diff is between a file and its backup. In this case, the original
1745 1746 # file should be patched (see original mpatch code).
1746 1747 isbackup = (abase == bbase and bfile.startswith(afile))
1747 1748 fname = None
1748 1749 if not missing:
1749 1750 if gooda and goodb:
1750 1751 if isbackup:
1751 1752 fname = afile
1752 1753 else:
1753 1754 fname = bfile
1754 1755 elif gooda:
1755 1756 fname = afile
1756 1757
1757 1758 if not fname:
1758 1759 if not nullb:
1759 1760 if isbackup:
1760 1761 fname = afile
1761 1762 else:
1762 1763 fname = bfile
1763 1764 elif not nulla:
1764 1765 fname = afile
1765 1766 else:
1766 1767 raise PatchError(_("undefined source and destination files"))
1767 1768
1768 1769 gp = patchmeta(fname)
1769 1770 if create:
1770 1771 gp.op = 'ADD'
1771 1772 elif remove:
1772 1773 gp.op = 'DELETE'
1773 1774 return gp
1774 1775
1775 1776 def scanpatch(fp):
1776 1777 """like patch.iterhunks, but yield different events
1777 1778
1778 1779 - ('file', [header_lines + fromfile + tofile])
1779 1780 - ('context', [context_lines])
1780 1781 - ('hunk', [hunk_lines])
1781 1782 - ('range', (-start,len, +start,len, proc))
1782 1783 """
1783 1784 lines_re = re.compile(br'@@ -(\d+),(\d+) \+(\d+),(\d+) @@\s*(.*)')
1784 1785 lr = linereader(fp)
1785 1786
1786 1787 def scanwhile(first, p):
1787 1788 """scan lr while predicate holds"""
1788 1789 lines = [first]
1789 1790 for line in iter(lr.readline, ''):
1790 1791 if p(line):
1791 1792 lines.append(line)
1792 1793 else:
1793 1794 lr.push(line)
1794 1795 break
1795 1796 return lines
1796 1797
1797 1798 for line in iter(lr.readline, ''):
1798 1799 if line.startswith('diff --git a/') or line.startswith('diff -r '):
1799 1800 def notheader(line):
1800 1801 s = line.split(None, 1)
1801 1802 return not s or s[0] not in ('---', 'diff')
1802 1803 header = scanwhile(line, notheader)
1803 1804 fromfile = lr.readline()
1804 1805 if fromfile.startswith('---'):
1805 1806 tofile = lr.readline()
1806 1807 header += [fromfile, tofile]
1807 1808 else:
1808 1809 lr.push(fromfile)
1809 1810 yield 'file', header
1810 1811 elif line.startswith(' '):
1811 1812 cs = (' ', '\\')
1812 1813 yield 'context', scanwhile(line, lambda l: l.startswith(cs))
1813 1814 elif line.startswith(('-', '+')):
1814 1815 cs = ('-', '+', '\\')
1815 1816 yield 'hunk', scanwhile(line, lambda l: l.startswith(cs))
1816 1817 else:
1817 1818 m = lines_re.match(line)
1818 1819 if m:
1819 1820 yield 'range', m.groups()
1820 1821 else:
1821 1822 yield 'other', line
1822 1823
1823 1824 def scangitpatch(lr, firstline):
1824 1825 """
1825 1826 Git patches can emit:
1826 1827 - rename a to b
1827 1828 - change b
1828 1829 - copy a to c
1829 1830 - change c
1830 1831
1831 1832 We cannot apply this sequence as-is, the renamed 'a' could not be
1832 1833 found for it would have been renamed already. And we cannot copy
1833 1834 from 'b' instead because 'b' would have been changed already. So
1834 1835 we scan the git patch for copy and rename commands so we can
1835 1836 perform the copies ahead of time.
1836 1837 """
1837 1838 pos = 0
1838 1839 try:
1839 1840 pos = lr.fp.tell()
1840 1841 fp = lr.fp
1841 1842 except IOError:
1842 1843 fp = stringio(lr.fp.read())
1843 1844 gitlr = linereader(fp)
1844 1845 gitlr.push(firstline)
1845 1846 gitpatches = readgitpatch(gitlr)
1846 1847 fp.seek(pos)
1847 1848 return gitpatches
1848 1849
1849 1850 def iterhunks(fp):
1850 1851 """Read a patch and yield the following events:
1851 1852 - ("file", afile, bfile, firsthunk): select a new target file.
1852 1853 - ("hunk", hunk): a new hunk is ready to be applied, follows a
1853 1854 "file" event.
1854 1855 - ("git", gitchanges): current diff is in git format, gitchanges
1855 1856 maps filenames to gitpatch records. Unique event.
1856 1857 """
1857 1858 afile = ""
1858 1859 bfile = ""
1859 1860 state = None
1860 1861 hunknum = 0
1861 1862 emitfile = newfile = False
1862 1863 gitpatches = None
1863 1864
1864 1865 # our states
1865 1866 BFILE = 1
1866 1867 context = None
1867 1868 lr = linereader(fp)
1868 1869
1869 1870 for x in iter(lr.readline, ''):
1870 1871 if state == BFILE and (
1871 1872 (not context and x.startswith('@'))
1872 1873 or (context is not False and x.startswith('***************'))
1873 1874 or x.startswith('GIT binary patch')):
1874 1875 gp = None
1875 1876 if (gitpatches and
1876 1877 gitpatches[-1].ispatching(afile, bfile)):
1877 1878 gp = gitpatches.pop()
1878 1879 if x.startswith('GIT binary patch'):
1879 1880 h = binhunk(lr, gp.path)
1880 1881 else:
1881 1882 if context is None and x.startswith('***************'):
1882 1883 context = True
1883 1884 h = hunk(x, hunknum + 1, lr, context)
1884 1885 hunknum += 1
1885 1886 if emitfile:
1886 1887 emitfile = False
1887 1888 yield 'file', (afile, bfile, h, gp and gp.copy() or None)
1888 1889 yield 'hunk', h
1889 1890 elif x.startswith('diff --git a/'):
1890 1891 m = gitre.match(x.rstrip(' \r\n'))
1891 1892 if not m:
1892 1893 continue
1893 1894 if gitpatches is None:
1894 1895 # scan whole input for git metadata
1895 1896 gitpatches = scangitpatch(lr, x)
1896 1897 yield 'git', [g.copy() for g in gitpatches
1897 1898 if g.op in ('COPY', 'RENAME')]
1898 1899 gitpatches.reverse()
1899 1900 afile = 'a/' + m.group(1)
1900 1901 bfile = 'b/' + m.group(2)
1901 1902 while gitpatches and not gitpatches[-1].ispatching(afile, bfile):
1902 1903 gp = gitpatches.pop()
1903 1904 yield 'file', ('a/' + gp.path, 'b/' + gp.path, None, gp.copy())
1904 1905 if not gitpatches:
1905 1906 raise PatchError(_('failed to synchronize metadata for "%s"')
1906 1907 % afile[2:])
1907 1908 newfile = True
1908 1909 elif x.startswith('---'):
1909 1910 # check for a unified diff
1910 1911 l2 = lr.readline()
1911 1912 if not l2.startswith('+++'):
1912 1913 lr.push(l2)
1913 1914 continue
1914 1915 newfile = True
1915 1916 context = False
1916 1917 afile = parsefilename(x)
1917 1918 bfile = parsefilename(l2)
1918 1919 elif x.startswith('***'):
1919 1920 # check for a context diff
1920 1921 l2 = lr.readline()
1921 1922 if not l2.startswith('---'):
1922 1923 lr.push(l2)
1923 1924 continue
1924 1925 l3 = lr.readline()
1925 1926 lr.push(l3)
1926 1927 if not l3.startswith("***************"):
1927 1928 lr.push(l2)
1928 1929 continue
1929 1930 newfile = True
1930 1931 context = True
1931 1932 afile = parsefilename(x)
1932 1933 bfile = parsefilename(l2)
1933 1934
1934 1935 if newfile:
1935 1936 newfile = False
1936 1937 emitfile = True
1937 1938 state = BFILE
1938 1939 hunknum = 0
1939 1940
1940 1941 while gitpatches:
1941 1942 gp = gitpatches.pop()
1942 1943 yield 'file', ('a/' + gp.path, 'b/' + gp.path, None, gp.copy())
1943 1944
1944 1945 def applybindelta(binchunk, data):
1945 1946 """Apply a binary delta hunk
1946 1947 The algorithm used is the algorithm from git's patch-delta.c
1947 1948 """
1948 1949 def deltahead(binchunk):
1949 1950 i = 0
1950 1951 for c in pycompat.bytestr(binchunk):
1951 1952 i += 1
1952 1953 if not (ord(c) & 0x80):
1953 1954 return i
1954 1955 return i
1955 1956 out = ""
1956 1957 s = deltahead(binchunk)
1957 1958 binchunk = binchunk[s:]
1958 1959 s = deltahead(binchunk)
1959 1960 binchunk = binchunk[s:]
1960 1961 i = 0
1961 1962 while i < len(binchunk):
1962 1963 cmd = ord(binchunk[i:i + 1])
1963 1964 i += 1
1964 1965 if (cmd & 0x80):
1965 1966 offset = 0
1966 1967 size = 0
1967 1968 if (cmd & 0x01):
1968 1969 offset = ord(binchunk[i:i + 1])
1969 1970 i += 1
1970 1971 if (cmd & 0x02):
1971 1972 offset |= ord(binchunk[i:i + 1]) << 8
1972 1973 i += 1
1973 1974 if (cmd & 0x04):
1974 1975 offset |= ord(binchunk[i:i + 1]) << 16
1975 1976 i += 1
1976 1977 if (cmd & 0x08):
1977 1978 offset |= ord(binchunk[i:i + 1]) << 24
1978 1979 i += 1
1979 1980 if (cmd & 0x10):
1980 1981 size = ord(binchunk[i:i + 1])
1981 1982 i += 1
1982 1983 if (cmd & 0x20):
1983 1984 size |= ord(binchunk[i:i + 1]) << 8
1984 1985 i += 1
1985 1986 if (cmd & 0x40):
1986 1987 size |= ord(binchunk[i:i + 1]) << 16
1987 1988 i += 1
1988 1989 if size == 0:
1989 1990 size = 0x10000
1990 1991 offset_end = offset + size
1991 1992 out += data[offset:offset_end]
1992 1993 elif cmd != 0:
1993 1994 offset_end = i + cmd
1994 1995 out += binchunk[i:offset_end]
1995 1996 i += cmd
1996 1997 else:
1997 1998 raise PatchError(_('unexpected delta opcode 0'))
1998 1999 return out
1999 2000
2000 2001 def applydiff(ui, fp, backend, store, strip=1, prefix='', eolmode='strict'):
2001 2002 """Reads a patch from fp and tries to apply it.
2002 2003
2003 2004 Returns 0 for a clean patch, -1 if any rejects were found and 1 if
2004 2005 there was any fuzz.
2005 2006
2006 2007 If 'eolmode' is 'strict', the patch content and patched file are
2007 2008 read in binary mode. Otherwise, line endings are ignored when
2008 2009 patching then normalized according to 'eolmode'.
2009 2010 """
2010 2011 return _applydiff(ui, fp, patchfile, backend, store, strip=strip,
2011 2012 prefix=prefix, eolmode=eolmode)
2012 2013
2013 2014 def _canonprefix(repo, prefix):
2014 2015 if prefix:
2015 2016 prefix = pathutil.canonpath(repo.root, repo.getcwd(), prefix)
2016 2017 if prefix != '':
2017 2018 prefix += '/'
2018 2019 return prefix
2019 2020
2020 2021 def _applydiff(ui, fp, patcher, backend, store, strip=1, prefix='',
2021 2022 eolmode='strict'):
2022 2023 prefix = _canonprefix(backend.repo, prefix)
2023 2024 def pstrip(p):
2024 2025 return pathtransform(p, strip - 1, prefix)[1]
2025 2026
2026 2027 rejects = 0
2027 2028 err = 0
2028 2029 current_file = None
2029 2030
2030 2031 for state, values in iterhunks(fp):
2031 2032 if state == 'hunk':
2032 2033 if not current_file:
2033 2034 continue
2034 2035 ret = current_file.apply(values)
2035 2036 if ret > 0:
2036 2037 err = 1
2037 2038 elif state == 'file':
2038 2039 if current_file:
2039 2040 rejects += current_file.close()
2040 2041 current_file = None
2041 2042 afile, bfile, first_hunk, gp = values
2042 2043 if gp:
2043 2044 gp.path = pstrip(gp.path)
2044 2045 if gp.oldpath:
2045 2046 gp.oldpath = pstrip(gp.oldpath)
2046 2047 else:
2047 2048 gp = makepatchmeta(backend, afile, bfile, first_hunk, strip,
2048 2049 prefix)
2049 2050 if gp.op == 'RENAME':
2050 2051 backend.unlink(gp.oldpath)
2051 2052 if not first_hunk:
2052 2053 if gp.op == 'DELETE':
2053 2054 backend.unlink(gp.path)
2054 2055 continue
2055 2056 data, mode = None, None
2056 2057 if gp.op in ('RENAME', 'COPY'):
2057 2058 data, mode = store.getfile(gp.oldpath)[:2]
2058 2059 if data is None:
2059 2060 # This means that the old path does not exist
2060 2061 raise PatchError(_("source file '%s' does not exist")
2061 2062 % gp.oldpath)
2062 2063 if gp.mode:
2063 2064 mode = gp.mode
2064 2065 if gp.op == 'ADD':
2065 2066 # Added files without content have no hunk and
2066 2067 # must be created
2067 2068 data = ''
2068 2069 if data or mode:
2069 2070 if (gp.op in ('ADD', 'RENAME', 'COPY')
2070 2071 and backend.exists(gp.path)):
2071 2072 raise PatchError(_("cannot create %s: destination "
2072 2073 "already exists") % gp.path)
2073 2074 backend.setfile(gp.path, data, mode, gp.oldpath)
2074 2075 continue
2075 2076 try:
2076 2077 current_file = patcher(ui, gp, backend, store,
2077 2078 eolmode=eolmode)
2078 2079 except PatchError as inst:
2079 2080 ui.warn(str(inst) + '\n')
2080 2081 current_file = None
2081 2082 rejects += 1
2082 2083 continue
2083 2084 elif state == 'git':
2084 2085 for gp in values:
2085 2086 path = pstrip(gp.oldpath)
2086 2087 data, mode = backend.getfile(path)
2087 2088 if data is None:
2088 2089 # The error ignored here will trigger a getfile()
2089 2090 # error in a place more appropriate for error
2090 2091 # handling, and will not interrupt the patching
2091 2092 # process.
2092 2093 pass
2093 2094 else:
2094 2095 store.setfile(path, data, mode)
2095 2096 else:
2096 2097 raise error.Abort(_('unsupported parser state: %s') % state)
2097 2098
2098 2099 if current_file:
2099 2100 rejects += current_file.close()
2100 2101
2101 2102 if rejects:
2102 2103 return -1
2103 2104 return err
2104 2105
2105 2106 def _externalpatch(ui, repo, patcher, patchname, strip, files,
2106 2107 similarity):
2107 2108 """use <patcher> to apply <patchname> to the working directory.
2108 2109 returns whether patch was applied with fuzz factor."""
2109 2110
2110 2111 fuzz = False
2111 2112 args = []
2112 2113 cwd = repo.root
2113 2114 if cwd:
2114 2115 args.append('-d %s' % procutil.shellquote(cwd))
2115 2116 cmd = ('%s %s -p%d < %s'
2116 2117 % (patcher, ' '.join(args), strip, procutil.shellquote(patchname)))
2117 2118 ui.debug('Using external patch tool: %s\n' % cmd)
2118 2119 fp = procutil.popen(cmd, 'rb')
2119 2120 try:
2120 2121 for line in util.iterfile(fp):
2121 2122 line = line.rstrip()
2122 2123 ui.note(line + '\n')
2123 2124 if line.startswith('patching file '):
2124 2125 pf = util.parsepatchoutput(line)
2125 2126 printed_file = False
2126 2127 files.add(pf)
2127 2128 elif line.find('with fuzz') >= 0:
2128 2129 fuzz = True
2129 2130 if not printed_file:
2130 2131 ui.warn(pf + '\n')
2131 2132 printed_file = True
2132 2133 ui.warn(line + '\n')
2133 2134 elif line.find('saving rejects to file') >= 0:
2134 2135 ui.warn(line + '\n')
2135 2136 elif line.find('FAILED') >= 0:
2136 2137 if not printed_file:
2137 2138 ui.warn(pf + '\n')
2138 2139 printed_file = True
2139 2140 ui.warn(line + '\n')
2140 2141 finally:
2141 2142 if files:
2142 2143 scmutil.marktouched(repo, files, similarity)
2143 2144 code = fp.close()
2144 2145 if code:
2145 2146 raise PatchError(_("patch command failed: %s") %
2146 2147 procutil.explainexit(code))
2147 2148 return fuzz
2148 2149
2149 2150 def patchbackend(ui, backend, patchobj, strip, prefix, files=None,
2150 2151 eolmode='strict'):
2151 2152 if files is None:
2152 2153 files = set()
2153 2154 if eolmode is None:
2154 2155 eolmode = ui.config('patch', 'eol')
2155 2156 if eolmode.lower() not in eolmodes:
2156 2157 raise error.Abort(_('unsupported line endings type: %s') % eolmode)
2157 2158 eolmode = eolmode.lower()
2158 2159
2159 2160 store = filestore()
2160 2161 try:
2161 2162 fp = open(patchobj, 'rb')
2162 2163 except TypeError:
2163 2164 fp = patchobj
2164 2165 try:
2165 2166 ret = applydiff(ui, fp, backend, store, strip=strip, prefix=prefix,
2166 2167 eolmode=eolmode)
2167 2168 finally:
2168 2169 if fp != patchobj:
2169 2170 fp.close()
2170 2171 files.update(backend.close())
2171 2172 store.close()
2172 2173 if ret < 0:
2173 2174 raise PatchError(_('patch failed to apply'))
2174 2175 return ret > 0
2175 2176
2176 2177 def internalpatch(ui, repo, patchobj, strip, prefix='', files=None,
2177 2178 eolmode='strict', similarity=0):
2178 2179 """use builtin patch to apply <patchobj> to the working directory.
2179 2180 returns whether patch was applied with fuzz factor."""
2180 2181 backend = workingbackend(ui, repo, similarity)
2181 2182 return patchbackend(ui, backend, patchobj, strip, prefix, files, eolmode)
2182 2183
2183 2184 def patchrepo(ui, repo, ctx, store, patchobj, strip, prefix, files=None,
2184 2185 eolmode='strict'):
2185 2186 backend = repobackend(ui, repo, ctx, store)
2186 2187 return patchbackend(ui, backend, patchobj, strip, prefix, files, eolmode)
2187 2188
2188 2189 def patch(ui, repo, patchname, strip=1, prefix='', files=None, eolmode='strict',
2189 2190 similarity=0):
2190 2191 """Apply <patchname> to the working directory.
2191 2192
2192 2193 'eolmode' specifies how end of lines should be handled. It can be:
2193 2194 - 'strict': inputs are read in binary mode, EOLs are preserved
2194 2195 - 'crlf': EOLs are ignored when patching and reset to CRLF
2195 2196 - 'lf': EOLs are ignored when patching and reset to LF
2196 2197 - None: get it from user settings, default to 'strict'
2197 2198 'eolmode' is ignored when using an external patcher program.
2198 2199
2199 2200 Returns whether patch was applied with fuzz factor.
2200 2201 """
2201 2202 patcher = ui.config('ui', 'patch')
2202 2203 if files is None:
2203 2204 files = set()
2204 2205 if patcher:
2205 2206 return _externalpatch(ui, repo, patcher, patchname, strip,
2206 2207 files, similarity)
2207 2208 return internalpatch(ui, repo, patchname, strip, prefix, files, eolmode,
2208 2209 similarity)
2209 2210
2210 2211 def changedfiles(ui, repo, patchpath, strip=1, prefix=''):
2211 2212 backend = fsbackend(ui, repo.root)
2212 2213 prefix = _canonprefix(repo, prefix)
2213 2214 with open(patchpath, 'rb') as fp:
2214 2215 changed = set()
2215 2216 for state, values in iterhunks(fp):
2216 2217 if state == 'file':
2217 2218 afile, bfile, first_hunk, gp = values
2218 2219 if gp:
2219 2220 gp.path = pathtransform(gp.path, strip - 1, prefix)[1]
2220 2221 if gp.oldpath:
2221 2222 gp.oldpath = pathtransform(gp.oldpath, strip - 1,
2222 2223 prefix)[1]
2223 2224 else:
2224 2225 gp = makepatchmeta(backend, afile, bfile, first_hunk, strip,
2225 2226 prefix)
2226 2227 changed.add(gp.path)
2227 2228 if gp.op == 'RENAME':
2228 2229 changed.add(gp.oldpath)
2229 2230 elif state not in ('hunk', 'git'):
2230 2231 raise error.Abort(_('unsupported parser state: %s') % state)
2231 2232 return changed
2232 2233
2233 2234 class GitDiffRequired(Exception):
2234 2235 pass
2235 2236
2236 2237 diffopts = diffutil.diffallopts
2237 2238 diffallopts = diffutil.diffallopts
2238 2239 difffeatureopts = diffutil.difffeatureopts
2239 2240
2240 2241 def diff(repo, node1=None, node2=None, match=None, changes=None,
2241 2242 opts=None, losedatafn=None, pathfn=None, copy=None,
2242 2243 copysourcematch=None, hunksfilterfn=None):
2243 2244 '''yields diff of changes to files between two nodes, or node and
2244 2245 working directory.
2245 2246
2246 2247 if node1 is None, use first dirstate parent instead.
2247 2248 if node2 is None, compare node1 with working directory.
2248 2249
2249 2250 losedatafn(**kwarg) is a callable run when opts.upgrade=True and
2250 2251 every time some change cannot be represented with the current
2251 2252 patch format. Return False to upgrade to git patch format, True to
2252 2253 accept the loss or raise an exception to abort the diff. It is
2253 2254 called with the name of current file being diffed as 'fn'. If set
2254 2255 to None, patches will always be upgraded to git format when
2255 2256 necessary.
2256 2257
2257 2258 prefix is a filename prefix that is prepended to all filenames on
2258 2259 display (used for subrepos).
2259 2260
2260 2261 relroot, if not empty, must be normalized with a trailing /. Any match
2261 2262 patterns that fall outside it will be ignored.
2262 2263
2263 2264 copy, if not empty, should contain mappings {dst@y: src@x} of copy
2264 2265 information.
2265 2266
2266 2267 if copysourcematch is not None, then copy sources will be filtered by this
2267 2268 matcher
2268 2269
2269 2270 hunksfilterfn, if not None, should be a function taking a filectx and
2270 2271 hunks generator that may yield filtered hunks.
2271 2272 '''
2272 2273 if not node1 and not node2:
2273 2274 node1 = repo.dirstate.p1()
2274 2275
2275 2276 ctx1 = repo[node1]
2276 2277 ctx2 = repo[node2]
2277 2278
2278 2279 for fctx1, fctx2, hdr, hunks in diffhunks(
2279 2280 repo, ctx1=ctx1, ctx2=ctx2, match=match, changes=changes, opts=opts,
2280 2281 losedatafn=losedatafn, pathfn=pathfn, copy=copy,
2281 2282 copysourcematch=copysourcematch):
2282 2283 if hunksfilterfn is not None:
2283 2284 # If the file has been removed, fctx2 is None; but this should
2284 2285 # not occur here since we catch removed files early in
2285 2286 # logcmdutil.getlinerangerevs() for 'hg log -L'.
2286 2287 assert fctx2 is not None, (
2287 2288 'fctx2 unexpectly None in diff hunks filtering')
2288 2289 hunks = hunksfilterfn(fctx2, hunks)
2289 2290 text = ''.join(sum((list(hlines) for hrange, hlines in hunks), []))
2290 2291 if hdr and (text or len(hdr) > 1):
2291 2292 yield '\n'.join(hdr) + '\n'
2292 2293 if text:
2293 2294 yield text
2294 2295
2295 2296 def diffhunks(repo, ctx1, ctx2, match=None, changes=None, opts=None,
2296 2297 losedatafn=None, pathfn=None, copy=None, copysourcematch=None):
2297 2298 """Yield diff of changes to files in the form of (`header`, `hunks`) tuples
2298 2299 where `header` is a list of diff headers and `hunks` is an iterable of
2299 2300 (`hunkrange`, `hunklines`) tuples.
2300 2301
2301 2302 See diff() for the meaning of parameters.
2302 2303 """
2303 2304
2304 2305 if opts is None:
2305 2306 opts = mdiff.defaultopts
2306 2307
2307 2308 def lrugetfilectx():
2308 2309 cache = {}
2309 2310 order = collections.deque()
2310 2311 def getfilectx(f, ctx):
2311 2312 fctx = ctx.filectx(f, filelog=cache.get(f))
2312 2313 if f not in cache:
2313 2314 if len(cache) > 20:
2314 2315 del cache[order.popleft()]
2315 2316 cache[f] = fctx.filelog()
2316 2317 else:
2317 2318 order.remove(f)
2318 2319 order.append(f)
2319 2320 return fctx
2320 2321 return getfilectx
2321 2322 getfilectx = lrugetfilectx()
2322 2323
2323 2324 if not changes:
2324 2325 changes = ctx1.status(ctx2, match=match)
2325 2326 modified, added, removed = changes[:3]
2326 2327
2327 2328 if not modified and not added and not removed:
2328 2329 return []
2329 2330
2330 2331 if repo.ui.debugflag:
2331 2332 hexfunc = hex
2332 2333 else:
2333 2334 hexfunc = short
2334 2335 revs = [hexfunc(node) for node in [ctx1.node(), ctx2.node()] if node]
2335 2336
2336 2337 if copy is None:
2337 2338 copy = {}
2338 2339 if opts.git or opts.upgrade:
2339 2340 copy = copies.pathcopies(ctx1, ctx2, match=match)
2340 2341
2341 2342 if copysourcematch:
2342 2343 # filter out copies where source side isn't inside the matcher
2343 2344 # (copies.pathcopies() already filtered out the destination)
2344 2345 copy = {dst: src for dst, src in copy.iteritems()
2345 2346 if copysourcematch(src)}
2346 2347
2347 2348 modifiedset = set(modified)
2348 2349 addedset = set(added)
2349 2350 removedset = set(removed)
2350 2351 for f in modified:
2351 2352 if f not in ctx1:
2352 2353 # Fix up added, since merged-in additions appear as
2353 2354 # modifications during merges
2354 2355 modifiedset.remove(f)
2355 2356 addedset.add(f)
2356 2357 for f in removed:
2357 2358 if f not in ctx1:
2358 2359 # Merged-in additions that are then removed are reported as removed.
2359 2360 # They are not in ctx1, so We don't want to show them in the diff.
2360 2361 removedset.remove(f)
2361 2362 modified = sorted(modifiedset)
2362 2363 added = sorted(addedset)
2363 2364 removed = sorted(removedset)
2364 2365 for dst, src in list(copy.items()):
2365 2366 if src not in ctx1:
2366 2367 # Files merged in during a merge and then copied/renamed are
2367 2368 # reported as copies. We want to show them in the diff as additions.
2368 2369 del copy[dst]
2369 2370
2370 2371 prefetchmatch = scmutil.matchfiles(
2371 2372 repo, list(modifiedset | addedset | removedset))
2372 2373 scmutil.prefetchfiles(repo, [ctx1.rev(), ctx2.rev()], prefetchmatch)
2373 2374
2374 2375 def difffn(opts, losedata):
2375 2376 return trydiff(repo, revs, ctx1, ctx2, modified, added, removed,
2376 2377 copy, getfilectx, opts, losedata, pathfn)
2377 2378 if opts.upgrade and not opts.git:
2378 2379 try:
2379 2380 def losedata(fn):
2380 2381 if not losedatafn or not losedatafn(fn=fn):
2381 2382 raise GitDiffRequired
2382 2383 # Buffer the whole output until we are sure it can be generated
2383 2384 return list(difffn(opts.copy(git=False), losedata))
2384 2385 except GitDiffRequired:
2385 2386 return difffn(opts.copy(git=True), None)
2386 2387 else:
2387 2388 return difffn(opts, None)
2388 2389
2389 2390 def diffsinglehunk(hunklines):
2390 2391 """yield tokens for a list of lines in a single hunk"""
2391 2392 for line in hunklines:
2392 2393 # chomp
2393 2394 chompline = line.rstrip('\r\n')
2394 2395 # highlight tabs and trailing whitespace
2395 2396 stripline = chompline.rstrip()
2396 2397 if line.startswith('-'):
2397 2398 label = 'diff.deleted'
2398 2399 elif line.startswith('+'):
2399 2400 label = 'diff.inserted'
2400 2401 else:
2401 2402 raise error.ProgrammingError('unexpected hunk line: %s' % line)
2402 2403 for token in tabsplitter.findall(stripline):
2403 2404 if token.startswith('\t'):
2404 2405 yield (token, 'diff.tab')
2405 2406 else:
2406 2407 yield (token, label)
2407 2408
2408 2409 if chompline != stripline:
2409 2410 yield (chompline[len(stripline):], 'diff.trailingwhitespace')
2410 2411 if chompline != line:
2411 2412 yield (line[len(chompline):], '')
2412 2413
2413 2414 def diffsinglehunkinline(hunklines):
2414 2415 """yield tokens for a list of lines in a single hunk, with inline colors"""
2415 2416 # prepare deleted, and inserted content
2416 2417 a = ''
2417 2418 b = ''
2418 2419 for line in hunklines:
2419 2420 if line[0:1] == '-':
2420 2421 a += line[1:]
2421 2422 elif line[0:1] == '+':
2422 2423 b += line[1:]
2423 2424 else:
2424 2425 raise error.ProgrammingError('unexpected hunk line: %s' % line)
2425 2426 # fast path: if either side is empty, use diffsinglehunk
2426 2427 if not a or not b:
2427 2428 for t in diffsinglehunk(hunklines):
2428 2429 yield t
2429 2430 return
2430 2431 # re-split the content into words
2431 2432 al = wordsplitter.findall(a)
2432 2433 bl = wordsplitter.findall(b)
2433 2434 # re-arrange the words to lines since the diff algorithm is line-based
2434 2435 aln = [s if s == '\n' else s + '\n' for s in al]
2435 2436 bln = [s if s == '\n' else s + '\n' for s in bl]
2436 2437 an = ''.join(aln)
2437 2438 bn = ''.join(bln)
2438 2439 # run the diff algorithm, prepare atokens and btokens
2439 2440 atokens = []
2440 2441 btokens = []
2441 2442 blocks = mdiff.allblocks(an, bn, lines1=aln, lines2=bln)
2442 2443 for (a1, a2, b1, b2), btype in blocks:
2443 2444 changed = btype == '!'
2444 2445 for token in mdiff.splitnewlines(''.join(al[a1:a2])):
2445 2446 atokens.append((changed, token))
2446 2447 for token in mdiff.splitnewlines(''.join(bl[b1:b2])):
2447 2448 btokens.append((changed, token))
2448 2449
2449 2450 # yield deleted tokens, then inserted ones
2450 2451 for prefix, label, tokens in [('-', 'diff.deleted', atokens),
2451 2452 ('+', 'diff.inserted', btokens)]:
2452 2453 nextisnewline = True
2453 2454 for changed, token in tokens:
2454 2455 if nextisnewline:
2455 2456 yield (prefix, label)
2456 2457 nextisnewline = False
2457 2458 # special handling line end
2458 2459 isendofline = token.endswith('\n')
2459 2460 if isendofline:
2460 2461 chomp = token[:-1] # chomp
2461 2462 if chomp.endswith('\r'):
2462 2463 chomp = chomp[:-1]
2463 2464 endofline = token[len(chomp):]
2464 2465 token = chomp.rstrip() # detect spaces at the end
2465 2466 endspaces = chomp[len(token):]
2466 2467 # scan tabs
2467 2468 for maybetab in tabsplitter.findall(token):
2468 2469 if b'\t' == maybetab[0:1]:
2469 2470 currentlabel = 'diff.tab'
2470 2471 else:
2471 2472 if changed:
2472 2473 currentlabel = label + '.changed'
2473 2474 else:
2474 2475 currentlabel = label + '.unchanged'
2475 2476 yield (maybetab, currentlabel)
2476 2477 if isendofline:
2477 2478 if endspaces:
2478 2479 yield (endspaces, 'diff.trailingwhitespace')
2479 2480 yield (endofline, '')
2480 2481 nextisnewline = True
2481 2482
2482 2483 def difflabel(func, *args, **kw):
2483 2484 '''yields 2-tuples of (output, label) based on the output of func()'''
2484 2485 if kw.get(r'opts') and kw[r'opts'].worddiff:
2485 2486 dodiffhunk = diffsinglehunkinline
2486 2487 else:
2487 2488 dodiffhunk = diffsinglehunk
2488 2489 headprefixes = [('diff', 'diff.diffline'),
2489 2490 ('copy', 'diff.extended'),
2490 2491 ('rename', 'diff.extended'),
2491 2492 ('old', 'diff.extended'),
2492 2493 ('new', 'diff.extended'),
2493 2494 ('deleted', 'diff.extended'),
2494 2495 ('index', 'diff.extended'),
2495 2496 ('similarity', 'diff.extended'),
2496 2497 ('---', 'diff.file_a'),
2497 2498 ('+++', 'diff.file_b')]
2498 2499 textprefixes = [('@', 'diff.hunk'),
2499 2500 # - and + are handled by diffsinglehunk
2500 2501 ]
2501 2502 head = False
2502 2503
2503 2504 # buffers a hunk, i.e. adjacent "-", "+" lines without other changes.
2504 2505 hunkbuffer = []
2505 2506 def consumehunkbuffer():
2506 2507 if hunkbuffer:
2507 2508 for token in dodiffhunk(hunkbuffer):
2508 2509 yield token
2509 2510 hunkbuffer[:] = []
2510 2511
2511 2512 for chunk in func(*args, **kw):
2512 2513 lines = chunk.split('\n')
2513 2514 linecount = len(lines)
2514 2515 for i, line in enumerate(lines):
2515 2516 if head:
2516 2517 if line.startswith('@'):
2517 2518 head = False
2518 2519 else:
2519 2520 if line and not line.startswith((' ', '+', '-', '@', '\\')):
2520 2521 head = True
2521 2522 diffline = False
2522 2523 if not head and line and line.startswith(('+', '-')):
2523 2524 diffline = True
2524 2525
2525 2526 prefixes = textprefixes
2526 2527 if head:
2527 2528 prefixes = headprefixes
2528 2529 if diffline:
2529 2530 # buffered
2530 2531 bufferedline = line
2531 2532 if i + 1 < linecount:
2532 2533 bufferedline += "\n"
2533 2534 hunkbuffer.append(bufferedline)
2534 2535 else:
2535 2536 # unbuffered
2536 2537 for token in consumehunkbuffer():
2537 2538 yield token
2538 2539 stripline = line.rstrip()
2539 2540 for prefix, label in prefixes:
2540 2541 if stripline.startswith(prefix):
2541 2542 yield (stripline, label)
2542 2543 if line != stripline:
2543 2544 yield (line[len(stripline):],
2544 2545 'diff.trailingwhitespace')
2545 2546 break
2546 2547 else:
2547 2548 yield (line, '')
2548 2549 if i + 1 < linecount:
2549 2550 yield ('\n', '')
2550 2551 for token in consumehunkbuffer():
2551 2552 yield token
2552 2553
2553 2554 def diffui(*args, **kw):
2554 2555 '''like diff(), but yields 2-tuples of (output, label) for ui.write()'''
2555 2556 return difflabel(diff, *args, **kw)
2556 2557
2557 2558 def _filepairs(modified, added, removed, copy, opts):
2558 2559 '''generates tuples (f1, f2, copyop), where f1 is the name of the file
2559 2560 before and f2 is the the name after. For added files, f1 will be None,
2560 2561 and for removed files, f2 will be None. copyop may be set to None, 'copy'
2561 2562 or 'rename' (the latter two only if opts.git is set).'''
2562 2563 gone = set()
2563 2564
2564 2565 copyto = dict([(v, k) for k, v in copy.items()])
2565 2566
2566 2567 addedset, removedset = set(added), set(removed)
2567 2568
2568 2569 for f in sorted(modified + added + removed):
2569 2570 copyop = None
2570 2571 f1, f2 = f, f
2571 2572 if f in addedset:
2572 2573 f1 = None
2573 2574 if f in copy:
2574 2575 if opts.git:
2575 2576 f1 = copy[f]
2576 2577 if f1 in removedset and f1 not in gone:
2577 2578 copyop = 'rename'
2578 2579 gone.add(f1)
2579 2580 else:
2580 2581 copyop = 'copy'
2581 2582 elif f in removedset:
2582 2583 f2 = None
2583 2584 if opts.git:
2584 2585 # have we already reported a copy above?
2585 2586 if (f in copyto and copyto[f] in addedset
2586 2587 and copy[copyto[f]] == f):
2587 2588 continue
2588 2589 yield f1, f2, copyop
2589 2590
2590 2591 def trydiff(repo, revs, ctx1, ctx2, modified, added, removed,
2591 2592 copy, getfilectx, opts, losedatafn, pathfn):
2592 2593 '''given input data, generate a diff and yield it in blocks
2593 2594
2594 2595 If generating a diff would lose data like flags or binary data and
2595 2596 losedatafn is not None, it will be called.
2596 2597
2597 2598 pathfn is applied to every path in the diff output.
2598 2599 '''
2599 2600
2600 2601 def gitindex(text):
2601 2602 if not text:
2602 2603 text = ""
2603 2604 l = len(text)
2604 2605 s = hashlib.sha1('blob %d\0' % l)
2605 2606 s.update(text)
2606 2607 return hex(s.digest())
2607 2608
2608 2609 if opts.noprefix:
2609 2610 aprefix = bprefix = ''
2610 2611 else:
2611 2612 aprefix = 'a/'
2612 2613 bprefix = 'b/'
2613 2614
2614 2615 def diffline(f, revs):
2615 2616 revinfo = ' '.join(["-r %s" % rev for rev in revs])
2616 2617 return 'diff %s %s' % (revinfo, f)
2617 2618
2618 2619 def isempty(fctx):
2619 2620 return fctx is None or fctx.size() == 0
2620 2621
2621 2622 date1 = dateutil.datestr(ctx1.date())
2622 2623 date2 = dateutil.datestr(ctx2.date())
2623 2624
2624 2625 gitmode = {'l': '120000', 'x': '100755', '': '100644'}
2625 2626
2626 2627 if not pathfn:
2627 2628 pathfn = lambda f: f
2628 2629
2629 2630 for f1, f2, copyop in _filepairs(modified, added, removed, copy, opts):
2630 2631 content1 = None
2631 2632 content2 = None
2632 2633 fctx1 = None
2633 2634 fctx2 = None
2634 2635 flag1 = None
2635 2636 flag2 = None
2636 2637 if f1:
2637 2638 fctx1 = getfilectx(f1, ctx1)
2638 2639 if opts.git or losedatafn:
2639 2640 flag1 = ctx1.flags(f1)
2640 2641 if f2:
2641 2642 fctx2 = getfilectx(f2, ctx2)
2642 2643 if opts.git or losedatafn:
2643 2644 flag2 = ctx2.flags(f2)
2644 2645 # if binary is True, output "summary" or "base85", but not "text diff"
2645 2646 if opts.text:
2646 2647 binary = False
2647 2648 else:
2648 2649 binary = any(f.isbinary() for f in [fctx1, fctx2] if f is not None)
2649 2650
2650 2651 if losedatafn and not opts.git:
2651 2652 if (binary or
2652 2653 # copy/rename
2653 2654 f2 in copy or
2654 2655 # empty file creation
2655 2656 (not f1 and isempty(fctx2)) or
2656 2657 # empty file deletion
2657 2658 (isempty(fctx1) and not f2) or
2658 2659 # create with flags
2659 2660 (not f1 and flag2) or
2660 2661 # change flags
2661 2662 (f1 and f2 and flag1 != flag2)):
2662 2663 losedatafn(f2 or f1)
2663 2664
2664 2665 path1 = pathfn(f1 or f2)
2665 2666 path2 = pathfn(f2 or f1)
2666 2667 header = []
2667 2668 if opts.git:
2668 2669 header.append('diff --git %s%s %s%s' %
2669 2670 (aprefix, path1, bprefix, path2))
2670 2671 if not f1: # added
2671 2672 header.append('new file mode %s' % gitmode[flag2])
2672 2673 elif not f2: # removed
2673 2674 header.append('deleted file mode %s' % gitmode[flag1])
2674 2675 else: # modified/copied/renamed
2675 2676 mode1, mode2 = gitmode[flag1], gitmode[flag2]
2676 2677 if mode1 != mode2:
2677 2678 header.append('old mode %s' % mode1)
2678 2679 header.append('new mode %s' % mode2)
2679 2680 if copyop is not None:
2680 2681 if opts.showsimilarity:
2681 2682 sim = similar.score(ctx1[path1], ctx2[path2]) * 100
2682 2683 header.append('similarity index %d%%' % sim)
2683 2684 header.append('%s from %s' % (copyop, path1))
2684 2685 header.append('%s to %s' % (copyop, path2))
2685 2686 elif revs:
2686 2687 header.append(diffline(path1, revs))
2687 2688
2688 2689 # fctx.is | diffopts | what to | is fctx.data()
2689 2690 # binary() | text nobinary git index | output? | outputted?
2690 2691 # ------------------------------------|----------------------------
2691 2692 # yes | no no no * | summary | no
2692 2693 # yes | no no yes * | base85 | yes
2693 2694 # yes | no yes no * | summary | no
2694 2695 # yes | no yes yes 0 | summary | no
2695 2696 # yes | no yes yes >0 | summary | semi [1]
2696 2697 # yes | yes * * * | text diff | yes
2697 2698 # no | * * * * | text diff | yes
2698 2699 # [1]: hash(fctx.data()) is outputted. so fctx.data() cannot be faked
2699 2700 if binary and (not opts.git or (opts.git and opts.nobinary and not
2700 2701 opts.index)):
2701 2702 # fast path: no binary content will be displayed, content1 and
2702 2703 # content2 are only used for equivalent test. cmp() could have a
2703 2704 # fast path.
2704 2705 if fctx1 is not None:
2705 2706 content1 = b'\0'
2706 2707 if fctx2 is not None:
2707 2708 if fctx1 is not None and not fctx1.cmp(fctx2):
2708 2709 content2 = b'\0' # not different
2709 2710 else:
2710 2711 content2 = b'\0\0'
2711 2712 else:
2712 2713 # normal path: load contents
2713 2714 if fctx1 is not None:
2714 2715 content1 = fctx1.data()
2715 2716 if fctx2 is not None:
2716 2717 content2 = fctx2.data()
2717 2718
2718 2719 if binary and opts.git and not opts.nobinary:
2719 2720 text = mdiff.b85diff(content1, content2)
2720 2721 if text:
2721 2722 header.append('index %s..%s' %
2722 2723 (gitindex(content1), gitindex(content2)))
2723 2724 hunks = (None, [text]),
2724 2725 else:
2725 2726 if opts.git and opts.index > 0:
2726 2727 flag = flag1
2727 2728 if flag is None:
2728 2729 flag = flag2
2729 2730 header.append('index %s..%s %s' %
2730 2731 (gitindex(content1)[0:opts.index],
2731 2732 gitindex(content2)[0:opts.index],
2732 2733 gitmode[flag]))
2733 2734
2734 2735 uheaders, hunks = mdiff.unidiff(content1, date1,
2735 2736 content2, date2,
2736 2737 path1, path2,
2737 2738 binary=binary, opts=opts)
2738 2739 header.extend(uheaders)
2739 2740 yield fctx1, fctx2, header, hunks
2740 2741
2741 2742 def diffstatsum(stats):
2742 2743 maxfile, maxtotal, addtotal, removetotal, binary = 0, 0, 0, 0, False
2743 2744 for f, a, r, b in stats:
2744 2745 maxfile = max(maxfile, encoding.colwidth(f))
2745 2746 maxtotal = max(maxtotal, a + r)
2746 2747 addtotal += a
2747 2748 removetotal += r
2748 2749 binary = binary or b
2749 2750
2750 2751 return maxfile, maxtotal, addtotal, removetotal, binary
2751 2752
2752 2753 def diffstatdata(lines):
2753 2754 diffre = re.compile(br'^diff .*-r [a-z0-9]+\s(.*)$')
2754 2755
2755 2756 results = []
2756 2757 filename, adds, removes, isbinary = None, 0, 0, False
2757 2758
2758 2759 def addresult():
2759 2760 if filename:
2760 2761 results.append((filename, adds, removes, isbinary))
2761 2762
2762 2763 # inheader is used to track if a line is in the
2763 2764 # header portion of the diff. This helps properly account
2764 2765 # for lines that start with '--' or '++'
2765 2766 inheader = False
2766 2767
2767 2768 for line in lines:
2768 2769 if line.startswith('diff'):
2769 2770 addresult()
2770 2771 # starting a new file diff
2771 2772 # set numbers to 0 and reset inheader
2772 2773 inheader = True
2773 2774 adds, removes, isbinary = 0, 0, False
2774 2775 if line.startswith('diff --git a/'):
2775 2776 filename = gitre.search(line).group(2)
2776 2777 elif line.startswith('diff -r'):
2777 2778 # format: "diff -r ... -r ... filename"
2778 2779 filename = diffre.search(line).group(1)
2779 2780 elif line.startswith('@@'):
2780 2781 inheader = False
2781 2782 elif line.startswith('+') and not inheader:
2782 2783 adds += 1
2783 2784 elif line.startswith('-') and not inheader:
2784 2785 removes += 1
2785 2786 elif (line.startswith('GIT binary patch') or
2786 2787 line.startswith('Binary file')):
2787 2788 isbinary = True
2788 2789 elif line.startswith('rename from'):
2789 2790 filename = line[12:]
2790 2791 elif line.startswith('rename to'):
2791 2792 filename += ' => %s' % line[10:]
2792 2793 addresult()
2793 2794 return results
2794 2795
2795 2796 def diffstat(lines, width=80):
2796 2797 output = []
2797 2798 stats = diffstatdata(lines)
2798 2799 maxname, maxtotal, totaladds, totalremoves, hasbinary = diffstatsum(stats)
2799 2800
2800 2801 countwidth = len(str(maxtotal))
2801 2802 if hasbinary and countwidth < 3:
2802 2803 countwidth = 3
2803 2804 graphwidth = width - countwidth - maxname - 6
2804 2805 if graphwidth < 10:
2805 2806 graphwidth = 10
2806 2807
2807 2808 def scale(i):
2808 2809 if maxtotal <= graphwidth:
2809 2810 return i
2810 2811 # If diffstat runs out of room it doesn't print anything,
2811 2812 # which isn't very useful, so always print at least one + or -
2812 2813 # if there were at least some changes.
2813 2814 return max(i * graphwidth // maxtotal, int(bool(i)))
2814 2815
2815 2816 for filename, adds, removes, isbinary in stats:
2816 2817 if isbinary:
2817 2818 count = 'Bin'
2818 2819 else:
2819 2820 count = '%d' % (adds + removes)
2820 2821 pluses = '+' * scale(adds)
2821 2822 minuses = '-' * scale(removes)
2822 2823 output.append(' %s%s | %*s %s%s\n' %
2823 2824 (filename, ' ' * (maxname - encoding.colwidth(filename)),
2824 2825 countwidth, count, pluses, minuses))
2825 2826
2826 2827 if stats:
2827 2828 output.append(_(' %d files changed, %d insertions(+), '
2828 2829 '%d deletions(-)\n')
2829 2830 % (len(stats), totaladds, totalremoves))
2830 2831
2831 2832 return ''.join(output)
2832 2833
2833 2834 def diffstatui(*args, **kw):
2834 2835 '''like diffstat(), but yields 2-tuples of (output, label) for
2835 2836 ui.write()
2836 2837 '''
2837 2838
2838 2839 for line in diffstat(*args, **kw).splitlines():
2839 2840 if line and line[-1] in '+-':
2840 2841 name, graph = line.rsplit(' ', 1)
2841 2842 yield (name + ' ', '')
2842 2843 m = re.search(br'\++', graph)
2843 2844 if m:
2844 2845 yield (m.group(0), 'diffstat.inserted')
2845 2846 m = re.search(br'-+', graph)
2846 2847 if m:
2847 2848 yield (m.group(0), 'diffstat.deleted')
2848 2849 else:
2849 2850 yield (line, '')
2850 2851 yield ('\n', '')
General Comments 0
You need to be logged in to leave comments. Login now