##// END OF EJS Templates
changegroup: remove unused variable caught by pyflakes
Sean Farley -
r20936:bfb40168 stable
parent child Browse files
Show More
@@ -1,430 +1,430 b''
1 # changegroup.py - Mercurial changegroup manipulation functions
1 # changegroup.py - Mercurial changegroup manipulation functions
2 #
2 #
3 # Copyright 2006 Matt Mackall <mpm@selenic.com>
3 # Copyright 2006 Matt Mackall <mpm@selenic.com>
4 #
4 #
5 # This software may be used and distributed according to the terms of the
5 # This software may be used and distributed according to the terms of the
6 # GNU General Public License version 2 or any later version.
6 # GNU General Public License version 2 or any later version.
7
7
8 from i18n import _
8 from i18n import _
9 from node import nullrev, hex
9 from node import nullrev, hex
10 import mdiff, util, dagutil
10 import mdiff, util, dagutil
11 import struct, os, bz2, zlib, tempfile
11 import struct, os, bz2, zlib, tempfile
12
12
13 _BUNDLE10_DELTA_HEADER = "20s20s20s20s"
13 _BUNDLE10_DELTA_HEADER = "20s20s20s20s"
14
14
15 def readexactly(stream, n):
15 def readexactly(stream, n):
16 '''read n bytes from stream.read and abort if less was available'''
16 '''read n bytes from stream.read and abort if less was available'''
17 s = stream.read(n)
17 s = stream.read(n)
18 if len(s) < n:
18 if len(s) < n:
19 raise util.Abort(_("stream ended unexpectedly"
19 raise util.Abort(_("stream ended unexpectedly"
20 " (got %d bytes, expected %d)")
20 " (got %d bytes, expected %d)")
21 % (len(s), n))
21 % (len(s), n))
22 return s
22 return s
23
23
24 def getchunk(stream):
24 def getchunk(stream):
25 """return the next chunk from stream as a string"""
25 """return the next chunk from stream as a string"""
26 d = readexactly(stream, 4)
26 d = readexactly(stream, 4)
27 l = struct.unpack(">l", d)[0]
27 l = struct.unpack(">l", d)[0]
28 if l <= 4:
28 if l <= 4:
29 if l:
29 if l:
30 raise util.Abort(_("invalid chunk length %d") % l)
30 raise util.Abort(_("invalid chunk length %d") % l)
31 return ""
31 return ""
32 return readexactly(stream, l - 4)
32 return readexactly(stream, l - 4)
33
33
34 def chunkheader(length):
34 def chunkheader(length):
35 """return a changegroup chunk header (string)"""
35 """return a changegroup chunk header (string)"""
36 return struct.pack(">l", length + 4)
36 return struct.pack(">l", length + 4)
37
37
38 def closechunk():
38 def closechunk():
39 """return a changegroup chunk header (string) for a zero-length chunk"""
39 """return a changegroup chunk header (string) for a zero-length chunk"""
40 return struct.pack(">l", 0)
40 return struct.pack(">l", 0)
41
41
42 class nocompress(object):
42 class nocompress(object):
43 def compress(self, x):
43 def compress(self, x):
44 return x
44 return x
45 def flush(self):
45 def flush(self):
46 return ""
46 return ""
47
47
48 bundletypes = {
48 bundletypes = {
49 "": ("", nocompress), # only when using unbundle on ssh and old http servers
49 "": ("", nocompress), # only when using unbundle on ssh and old http servers
50 # since the unification ssh accepts a header but there
50 # since the unification ssh accepts a header but there
51 # is no capability signaling it.
51 # is no capability signaling it.
52 "HG10UN": ("HG10UN", nocompress),
52 "HG10UN": ("HG10UN", nocompress),
53 "HG10BZ": ("HG10", lambda: bz2.BZ2Compressor()),
53 "HG10BZ": ("HG10", lambda: bz2.BZ2Compressor()),
54 "HG10GZ": ("HG10GZ", lambda: zlib.compressobj()),
54 "HG10GZ": ("HG10GZ", lambda: zlib.compressobj()),
55 }
55 }
56
56
57 # hgweb uses this list to communicate its preferred type
57 # hgweb uses this list to communicate its preferred type
58 bundlepriority = ['HG10GZ', 'HG10BZ', 'HG10UN']
58 bundlepriority = ['HG10GZ', 'HG10BZ', 'HG10UN']
59
59
60 def writebundle(cg, filename, bundletype):
60 def writebundle(cg, filename, bundletype):
61 """Write a bundle file and return its filename.
61 """Write a bundle file and return its filename.
62
62
63 Existing files will not be overwritten.
63 Existing files will not be overwritten.
64 If no filename is specified, a temporary file is created.
64 If no filename is specified, a temporary file is created.
65 bz2 compression can be turned off.
65 bz2 compression can be turned off.
66 The bundle file will be deleted in case of errors.
66 The bundle file will be deleted in case of errors.
67 """
67 """
68
68
69 fh = None
69 fh = None
70 cleanup = None
70 cleanup = None
71 try:
71 try:
72 if filename:
72 if filename:
73 fh = open(filename, "wb")
73 fh = open(filename, "wb")
74 else:
74 else:
75 fd, filename = tempfile.mkstemp(prefix="hg-bundle-", suffix=".hg")
75 fd, filename = tempfile.mkstemp(prefix="hg-bundle-", suffix=".hg")
76 fh = os.fdopen(fd, "wb")
76 fh = os.fdopen(fd, "wb")
77 cleanup = filename
77 cleanup = filename
78
78
79 header, compressor = bundletypes[bundletype]
79 header, compressor = bundletypes[bundletype]
80 fh.write(header)
80 fh.write(header)
81 z = compressor()
81 z = compressor()
82
82
83 # parse the changegroup data, otherwise we will block
83 # parse the changegroup data, otherwise we will block
84 # in case of sshrepo because we don't know the end of the stream
84 # in case of sshrepo because we don't know the end of the stream
85
85
86 # an empty chunkgroup is the end of the changegroup
86 # an empty chunkgroup is the end of the changegroup
87 # a changegroup has at least 2 chunkgroups (changelog and manifest).
87 # a changegroup has at least 2 chunkgroups (changelog and manifest).
88 # after that, an empty chunkgroup is the end of the changegroup
88 # after that, an empty chunkgroup is the end of the changegroup
89 empty = False
89 empty = False
90 count = 0
90 count = 0
91 while not empty or count <= 2:
91 while not empty or count <= 2:
92 empty = True
92 empty = True
93 count += 1
93 count += 1
94 while True:
94 while True:
95 chunk = getchunk(cg)
95 chunk = getchunk(cg)
96 if not chunk:
96 if not chunk:
97 break
97 break
98 empty = False
98 empty = False
99 fh.write(z.compress(chunkheader(len(chunk))))
99 fh.write(z.compress(chunkheader(len(chunk))))
100 pos = 0
100 pos = 0
101 while pos < len(chunk):
101 while pos < len(chunk):
102 next = pos + 2**20
102 next = pos + 2**20
103 fh.write(z.compress(chunk[pos:next]))
103 fh.write(z.compress(chunk[pos:next]))
104 pos = next
104 pos = next
105 fh.write(z.compress(closechunk()))
105 fh.write(z.compress(closechunk()))
106 fh.write(z.flush())
106 fh.write(z.flush())
107 cleanup = None
107 cleanup = None
108 return filename
108 return filename
109 finally:
109 finally:
110 if fh is not None:
110 if fh is not None:
111 fh.close()
111 fh.close()
112 if cleanup is not None:
112 if cleanup is not None:
113 os.unlink(cleanup)
113 os.unlink(cleanup)
114
114
115 def decompressor(fh, alg):
115 def decompressor(fh, alg):
116 if alg == 'UN':
116 if alg == 'UN':
117 return fh
117 return fh
118 elif alg == 'GZ':
118 elif alg == 'GZ':
119 def generator(f):
119 def generator(f):
120 zd = zlib.decompressobj()
120 zd = zlib.decompressobj()
121 for chunk in util.filechunkiter(f):
121 for chunk in util.filechunkiter(f):
122 yield zd.decompress(chunk)
122 yield zd.decompress(chunk)
123 elif alg == 'BZ':
123 elif alg == 'BZ':
124 def generator(f):
124 def generator(f):
125 zd = bz2.BZ2Decompressor()
125 zd = bz2.BZ2Decompressor()
126 zd.decompress("BZ")
126 zd.decompress("BZ")
127 for chunk in util.filechunkiter(f, 4096):
127 for chunk in util.filechunkiter(f, 4096):
128 yield zd.decompress(chunk)
128 yield zd.decompress(chunk)
129 else:
129 else:
130 raise util.Abort("unknown bundle compression '%s'" % alg)
130 raise util.Abort("unknown bundle compression '%s'" % alg)
131 return util.chunkbuffer(generator(fh))
131 return util.chunkbuffer(generator(fh))
132
132
133 class unbundle10(object):
133 class unbundle10(object):
134 deltaheader = _BUNDLE10_DELTA_HEADER
134 deltaheader = _BUNDLE10_DELTA_HEADER
135 deltaheadersize = struct.calcsize(deltaheader)
135 deltaheadersize = struct.calcsize(deltaheader)
136 def __init__(self, fh, alg):
136 def __init__(self, fh, alg):
137 self._stream = decompressor(fh, alg)
137 self._stream = decompressor(fh, alg)
138 self._type = alg
138 self._type = alg
139 self.callback = None
139 self.callback = None
140 def compressed(self):
140 def compressed(self):
141 return self._type != 'UN'
141 return self._type != 'UN'
142 def read(self, l):
142 def read(self, l):
143 return self._stream.read(l)
143 return self._stream.read(l)
144 def seek(self, pos):
144 def seek(self, pos):
145 return self._stream.seek(pos)
145 return self._stream.seek(pos)
146 def tell(self):
146 def tell(self):
147 return self._stream.tell()
147 return self._stream.tell()
148 def close(self):
148 def close(self):
149 return self._stream.close()
149 return self._stream.close()
150
150
151 def chunklength(self):
151 def chunklength(self):
152 d = readexactly(self._stream, 4)
152 d = readexactly(self._stream, 4)
153 l = struct.unpack(">l", d)[0]
153 l = struct.unpack(">l", d)[0]
154 if l <= 4:
154 if l <= 4:
155 if l:
155 if l:
156 raise util.Abort(_("invalid chunk length %d") % l)
156 raise util.Abort(_("invalid chunk length %d") % l)
157 return 0
157 return 0
158 if self.callback:
158 if self.callback:
159 self.callback()
159 self.callback()
160 return l - 4
160 return l - 4
161
161
162 def changelogheader(self):
162 def changelogheader(self):
163 """v10 does not have a changelog header chunk"""
163 """v10 does not have a changelog header chunk"""
164 return {}
164 return {}
165
165
166 def manifestheader(self):
166 def manifestheader(self):
167 """v10 does not have a manifest header chunk"""
167 """v10 does not have a manifest header chunk"""
168 return {}
168 return {}
169
169
170 def filelogheader(self):
170 def filelogheader(self):
171 """return the header of the filelogs chunk, v10 only has the filename"""
171 """return the header of the filelogs chunk, v10 only has the filename"""
172 l = self.chunklength()
172 l = self.chunklength()
173 if not l:
173 if not l:
174 return {}
174 return {}
175 fname = readexactly(self._stream, l)
175 fname = readexactly(self._stream, l)
176 return dict(filename=fname)
176 return dict(filename=fname)
177
177
178 def _deltaheader(self, headertuple, prevnode):
178 def _deltaheader(self, headertuple, prevnode):
179 node, p1, p2, cs = headertuple
179 node, p1, p2, cs = headertuple
180 if prevnode is None:
180 if prevnode is None:
181 deltabase = p1
181 deltabase = p1
182 else:
182 else:
183 deltabase = prevnode
183 deltabase = prevnode
184 return node, p1, p2, deltabase, cs
184 return node, p1, p2, deltabase, cs
185
185
186 def deltachunk(self, prevnode):
186 def deltachunk(self, prevnode):
187 l = self.chunklength()
187 l = self.chunklength()
188 if not l:
188 if not l:
189 return {}
189 return {}
190 headerdata = readexactly(self._stream, self.deltaheadersize)
190 headerdata = readexactly(self._stream, self.deltaheadersize)
191 header = struct.unpack(self.deltaheader, headerdata)
191 header = struct.unpack(self.deltaheader, headerdata)
192 delta = readexactly(self._stream, l - self.deltaheadersize)
192 delta = readexactly(self._stream, l - self.deltaheadersize)
193 node, p1, p2, deltabase, cs = self._deltaheader(header, prevnode)
193 node, p1, p2, deltabase, cs = self._deltaheader(header, prevnode)
194 return dict(node=node, p1=p1, p2=p2, cs=cs,
194 return dict(node=node, p1=p1, p2=p2, cs=cs,
195 deltabase=deltabase, delta=delta)
195 deltabase=deltabase, delta=delta)
196
196
197 class headerlessfixup(object):
197 class headerlessfixup(object):
198 def __init__(self, fh, h):
198 def __init__(self, fh, h):
199 self._h = h
199 self._h = h
200 self._fh = fh
200 self._fh = fh
201 def read(self, n):
201 def read(self, n):
202 if self._h:
202 if self._h:
203 d, self._h = self._h[:n], self._h[n:]
203 d, self._h = self._h[:n], self._h[n:]
204 if len(d) < n:
204 if len(d) < n:
205 d += readexactly(self._fh, n - len(d))
205 d += readexactly(self._fh, n - len(d))
206 return d
206 return d
207 return readexactly(self._fh, n)
207 return readexactly(self._fh, n)
208
208
209 def readbundle(fh, fname):
209 def readbundle(fh, fname):
210 header = readexactly(fh, 6)
210 header = readexactly(fh, 6)
211
211
212 if not fname:
212 if not fname:
213 fname = "stream"
213 fname = "stream"
214 if not header.startswith('HG') and header.startswith('\0'):
214 if not header.startswith('HG') and header.startswith('\0'):
215 fh = headerlessfixup(fh, header)
215 fh = headerlessfixup(fh, header)
216 header = "HG10UN"
216 header = "HG10UN"
217
217
218 magic, version, alg = header[0:2], header[2:4], header[4:6]
218 magic, version, alg = header[0:2], header[2:4], header[4:6]
219
219
220 if magic != 'HG':
220 if magic != 'HG':
221 raise util.Abort(_('%s: not a Mercurial bundle') % fname)
221 raise util.Abort(_('%s: not a Mercurial bundle') % fname)
222 if version != '10':
222 if version != '10':
223 raise util.Abort(_('%s: unknown bundle version %s') % (fname, version))
223 raise util.Abort(_('%s: unknown bundle version %s') % (fname, version))
224 return unbundle10(fh, alg)
224 return unbundle10(fh, alg)
225
225
226 class bundle10(object):
226 class bundle10(object):
227 deltaheader = _BUNDLE10_DELTA_HEADER
227 deltaheader = _BUNDLE10_DELTA_HEADER
228 def __init__(self, repo, bundlecaps=None):
228 def __init__(self, repo, bundlecaps=None):
229 """Given a source repo, construct a bundler.
229 """Given a source repo, construct a bundler.
230
230
231 bundlecaps is optional and can be used to specify the set of
231 bundlecaps is optional and can be used to specify the set of
232 capabilities which can be used to build the bundle.
232 capabilities which can be used to build the bundle.
233 """
233 """
234 # Set of capabilities we can use to build the bundle.
234 # Set of capabilities we can use to build the bundle.
235 if bundlecaps is None:
235 if bundlecaps is None:
236 bundlecaps = set()
236 bundlecaps = set()
237 self._bundlecaps = bundlecaps
237 self._bundlecaps = bundlecaps
238 self._changelog = repo.changelog
238 self._changelog = repo.changelog
239 self._manifest = repo.manifest
239 self._manifest = repo.manifest
240 reorder = repo.ui.config('bundle', 'reorder', 'auto')
240 reorder = repo.ui.config('bundle', 'reorder', 'auto')
241 if reorder == 'auto':
241 if reorder == 'auto':
242 reorder = None
242 reorder = None
243 else:
243 else:
244 reorder = util.parsebool(reorder)
244 reorder = util.parsebool(reorder)
245 self._repo = repo
245 self._repo = repo
246 self._reorder = reorder
246 self._reorder = reorder
247 self._progress = repo.ui.progress
247 self._progress = repo.ui.progress
248 def close(self):
248 def close(self):
249 return closechunk()
249 return closechunk()
250
250
251 def fileheader(self, fname):
251 def fileheader(self, fname):
252 return chunkheader(len(fname)) + fname
252 return chunkheader(len(fname)) + fname
253
253
254 def group(self, nodelist, revlog, lookup, units=None, reorder=None):
254 def group(self, nodelist, revlog, lookup, units=None, reorder=None):
255 """Calculate a delta group, yielding a sequence of changegroup chunks
255 """Calculate a delta group, yielding a sequence of changegroup chunks
256 (strings).
256 (strings).
257
257
258 Given a list of changeset revs, return a set of deltas and
258 Given a list of changeset revs, return a set of deltas and
259 metadata corresponding to nodes. The first delta is
259 metadata corresponding to nodes. The first delta is
260 first parent(nodelist[0]) -> nodelist[0], the receiver is
260 first parent(nodelist[0]) -> nodelist[0], the receiver is
261 guaranteed to have this parent as it has all history before
261 guaranteed to have this parent as it has all history before
262 these changesets. In the case firstparent is nullrev the
262 these changesets. In the case firstparent is nullrev the
263 changegroup starts with a full revision.
263 changegroup starts with a full revision.
264
264
265 If units is not None, progress detail will be generated, units specifies
265 If units is not None, progress detail will be generated, units specifies
266 the type of revlog that is touched (changelog, manifest, etc.).
266 the type of revlog that is touched (changelog, manifest, etc.).
267 """
267 """
268 # if we don't have any revisions touched by these changesets, bail
268 # if we don't have any revisions touched by these changesets, bail
269 if len(nodelist) == 0:
269 if len(nodelist) == 0:
270 yield self.close()
270 yield self.close()
271 return
271 return
272
272
273 # for generaldelta revlogs, we linearize the revs; this will both be
273 # for generaldelta revlogs, we linearize the revs; this will both be
274 # much quicker and generate a much smaller bundle
274 # much quicker and generate a much smaller bundle
275 if (revlog._generaldelta and reorder is not False) or reorder:
275 if (revlog._generaldelta and reorder is not False) or reorder:
276 dag = dagutil.revlogdag(revlog)
276 dag = dagutil.revlogdag(revlog)
277 revs = set(revlog.rev(n) for n in nodelist)
277 revs = set(revlog.rev(n) for n in nodelist)
278 revs = dag.linearize(revs)
278 revs = dag.linearize(revs)
279 else:
279 else:
280 revs = sorted([revlog.rev(n) for n in nodelist])
280 revs = sorted([revlog.rev(n) for n in nodelist])
281
281
282 # add the parent of the first rev
282 # add the parent of the first rev
283 p = revlog.parentrevs(revs[0])[0]
283 p = revlog.parentrevs(revs[0])[0]
284 revs.insert(0, p)
284 revs.insert(0, p)
285
285
286 # build deltas
286 # build deltas
287 total = len(revs) - 1
287 total = len(revs) - 1
288 msgbundling = _('bundling')
288 msgbundling = _('bundling')
289 for r in xrange(len(revs) - 1):
289 for r in xrange(len(revs) - 1):
290 if units is not None:
290 if units is not None:
291 self._progress(msgbundling, r + 1, unit=units, total=total)
291 self._progress(msgbundling, r + 1, unit=units, total=total)
292 prev, curr = revs[r], revs[r + 1]
292 prev, curr = revs[r], revs[r + 1]
293 linknode = lookup(revlog.node(curr))
293 linknode = lookup(revlog.node(curr))
294 for c in self.revchunk(revlog, curr, prev, linknode):
294 for c in self.revchunk(revlog, curr, prev, linknode):
295 yield c
295 yield c
296
296
297 yield self.close()
297 yield self.close()
298
298
299 # filter any nodes that claim to be part of the known set
299 # filter any nodes that claim to be part of the known set
300 def prune(self, revlog, missing, commonrevs, source):
300 def prune(self, revlog, missing, commonrevs, source):
301 rr, rl = revlog.rev, revlog.linkrev
301 rr, rl = revlog.rev, revlog.linkrev
302 return [n for n in missing if rl(rr(n)) not in commonrevs]
302 return [n for n in missing if rl(rr(n)) not in commonrevs]
303
303
304 def generate(self, commonrevs, clnodes, fastpathlinkrev, source):
304 def generate(self, commonrevs, clnodes, fastpathlinkrev, source):
305 '''yield a sequence of changegroup chunks (strings)'''
305 '''yield a sequence of changegroup chunks (strings)'''
306 repo = self._repo
306 repo = self._repo
307 cl = self._changelog
307 cl = self._changelog
308 mf = self._manifest
308 mf = self._manifest
309 reorder = self._reorder
309 reorder = self._reorder
310 progress = self._progress
310 progress = self._progress
311
311
312 # for progress output
312 # for progress output
313 msgbundling = _('bundling')
313 msgbundling = _('bundling')
314
314
315 mfs = {} # needed manifests
315 mfs = {} # needed manifests
316 fnodes = {} # needed file nodes
316 fnodes = {} # needed file nodes
317 changedfiles = set()
317 changedfiles = set()
318
318
319 # Callback for the changelog, used to collect changed files and manifest
319 # Callback for the changelog, used to collect changed files and manifest
320 # nodes.
320 # nodes.
321 # Returns the linkrev node (identity in the changelog case).
321 # Returns the linkrev node (identity in the changelog case).
322 def lookupcl(x):
322 def lookupcl(x):
323 c = cl.read(x)
323 c = cl.read(x)
324 changedfiles.update(c[3])
324 changedfiles.update(c[3])
325 # record the first changeset introducing this manifest version
325 # record the first changeset introducing this manifest version
326 mfs.setdefault(c[0], x)
326 mfs.setdefault(c[0], x)
327 return x
327 return x
328
328
329 # Callback for the manifest, used to collect linkrevs for filelog
329 # Callback for the manifest, used to collect linkrevs for filelog
330 # revisions.
330 # revisions.
331 # Returns the linkrev node (collected in lookupcl).
331 # Returns the linkrev node (collected in lookupcl).
332 def lookupmf(x):
332 def lookupmf(x):
333 clnode = mfs[x]
333 clnode = mfs[x]
334 if not fastpathlinkrev:
334 if not fastpathlinkrev:
335 mdata = mf.readfast(x)
335 mdata = mf.readfast(x)
336 for f, n in mdata.iteritems():
336 for f, n in mdata.iteritems():
337 if f in changedfiles:
337 if f in changedfiles:
338 # record the first changeset introducing this filelog
338 # record the first changeset introducing this filelog
339 # version
339 # version
340 fnodes[f].setdefault(n, clnode)
340 fnodes[f].setdefault(n, clnode)
341 return clnode
341 return clnode
342
342
343 for chunk in self.group(clnodes, cl, lookupcl, units=_('changesets'),
343 for chunk in self.group(clnodes, cl, lookupcl, units=_('changesets'),
344 reorder=reorder):
344 reorder=reorder):
345 yield chunk
345 yield chunk
346 progress(msgbundling, None)
346 progress(msgbundling, None)
347
347
348 for f in changedfiles:
348 for f in changedfiles:
349 fnodes[f] = {}
349 fnodes[f] = {}
350 mfnodes = self.prune(mf, mfs, commonrevs, source)
350 mfnodes = self.prune(mf, mfs, commonrevs, source)
351 for chunk in self.group(mfnodes, mf, lookupmf, units=_('manifests'),
351 for chunk in self.group(mfnodes, mf, lookupmf, units=_('manifests'),
352 reorder=reorder):
352 reorder=reorder):
353 yield chunk
353 yield chunk
354 progress(msgbundling, None)
354 progress(msgbundling, None)
355
355
356 mfs.clear()
356 mfs.clear()
357 needed = set(cl.rev(x) for x in clnodes)
357 needed = set(cl.rev(x) for x in clnodes)
358
358
359 def linknodes(filerevlog, fname):
359 def linknodes(filerevlog, fname):
360 if fastpathlinkrev:
360 if fastpathlinkrev:
361 ln, llr = filerevlog.node, filerevlog.linkrev
361 llr = filerevlog.linkrev
362 def genfilenodes():
362 def genfilenodes():
363 for r in filerevlog:
363 for r in filerevlog:
364 linkrev = llr(r)
364 linkrev = llr(r)
365 if linkrev in needed:
365 if linkrev in needed:
366 yield filerevlog.node(r), cl.node(linkrev)
366 yield filerevlog.node(r), cl.node(linkrev)
367 fnodes[fname] = dict(genfilenodes())
367 fnodes[fname] = dict(genfilenodes())
368 return fnodes.get(fname, {})
368 return fnodes.get(fname, {})
369
369
370 for chunk in self.generatefiles(changedfiles, linknodes, commonrevs,
370 for chunk in self.generatefiles(changedfiles, linknodes, commonrevs,
371 source):
371 source):
372 yield chunk
372 yield chunk
373
373
374 yield self.close()
374 yield self.close()
375 progress(msgbundling, None)
375 progress(msgbundling, None)
376
376
377 if clnodes:
377 if clnodes:
378 repo.hook('outgoing', node=hex(clnodes[0]), source=source)
378 repo.hook('outgoing', node=hex(clnodes[0]), source=source)
379
379
380 def generatefiles(self, changedfiles, linknodes, commonrevs, source):
380 def generatefiles(self, changedfiles, linknodes, commonrevs, source):
381 repo = self._repo
381 repo = self._repo
382 progress = self._progress
382 progress = self._progress
383 reorder = self._reorder
383 reorder = self._reorder
384 msgbundling = _('bundling')
384 msgbundling = _('bundling')
385
385
386 total = len(changedfiles)
386 total = len(changedfiles)
387 # for progress output
387 # for progress output
388 msgfiles = _('files')
388 msgfiles = _('files')
389 for i, fname in enumerate(sorted(changedfiles)):
389 for i, fname in enumerate(sorted(changedfiles)):
390 filerevlog = repo.file(fname)
390 filerevlog = repo.file(fname)
391 if not filerevlog:
391 if not filerevlog:
392 raise util.Abort(_("empty or missing revlog for %s") % fname)
392 raise util.Abort(_("empty or missing revlog for %s") % fname)
393
393
394 linkrevnodes = linknodes(filerevlog, fname)
394 linkrevnodes = linknodes(filerevlog, fname)
395 # Lookup for filenodes, we collected the linkrev nodes above in the
395 # Lookup for filenodes, we collected the linkrev nodes above in the
396 # fastpath case and with lookupmf in the slowpath case.
396 # fastpath case and with lookupmf in the slowpath case.
397 def lookupfilelog(x):
397 def lookupfilelog(x):
398 return linkrevnodes[x]
398 return linkrevnodes[x]
399
399
400 filenodes = self.prune(filerevlog, linkrevnodes, commonrevs, source)
400 filenodes = self.prune(filerevlog, linkrevnodes, commonrevs, source)
401 if filenodes:
401 if filenodes:
402 progress(msgbundling, i + 1, item=fname, unit=msgfiles,
402 progress(msgbundling, i + 1, item=fname, unit=msgfiles,
403 total=total)
403 total=total)
404 yield self.fileheader(fname)
404 yield self.fileheader(fname)
405 for chunk in self.group(filenodes, filerevlog, lookupfilelog,
405 for chunk in self.group(filenodes, filerevlog, lookupfilelog,
406 reorder=reorder):
406 reorder=reorder):
407 yield chunk
407 yield chunk
408
408
409 def revchunk(self, revlog, rev, prev, linknode):
409 def revchunk(self, revlog, rev, prev, linknode):
410 node = revlog.node(rev)
410 node = revlog.node(rev)
411 p1, p2 = revlog.parentrevs(rev)
411 p1, p2 = revlog.parentrevs(rev)
412 base = prev
412 base = prev
413
413
414 prefix = ''
414 prefix = ''
415 if base == nullrev:
415 if base == nullrev:
416 delta = revlog.revision(node)
416 delta = revlog.revision(node)
417 prefix = mdiff.trivialdiffheader(len(delta))
417 prefix = mdiff.trivialdiffheader(len(delta))
418 else:
418 else:
419 delta = revlog.revdiff(base, rev)
419 delta = revlog.revdiff(base, rev)
420 p1n, p2n = revlog.parents(node)
420 p1n, p2n = revlog.parents(node)
421 basenode = revlog.node(base)
421 basenode = revlog.node(base)
422 meta = self.builddeltaheader(node, p1n, p2n, basenode, linknode)
422 meta = self.builddeltaheader(node, p1n, p2n, basenode, linknode)
423 meta += prefix
423 meta += prefix
424 l = len(meta) + len(delta)
424 l = len(meta) + len(delta)
425 yield chunkheader(l)
425 yield chunkheader(l)
426 yield meta
426 yield meta
427 yield delta
427 yield delta
428 def builddeltaheader(self, node, p1n, p2n, basenode, linknode):
428 def builddeltaheader(self, node, p1n, p2n, basenode, linknode):
429 # do nothing with basenode, it is implicitly the previous one in HG10
429 # do nothing with basenode, it is implicitly the previous one in HG10
430 return struct.pack(self.deltaheader, node, p1n, p2n, linknode)
430 return struct.pack(self.deltaheader, node, p1n, p2n, linknode)
General Comments 0
You need to be logged in to leave comments. Login now