Show More
@@ -270,11 +270,12 b" def generatebundlev1(repo, compression='" | |||
|
270 | 270 | assert compression == 'UN' |
|
271 | 271 | |
|
272 | 272 | seen = 0 |
|
273 | repo.ui.progress(_('bundle'), 0, total=bytecount) | |
|
273 | repo.ui.progress(_('bundle'), 0, total=bytecount, unit=_('bytes')) | |
|
274 | 274 | |
|
275 | 275 | for chunk in it: |
|
276 | 276 | seen += len(chunk) |
|
277 |
repo.ui.progress(_('bundle'), seen, total=bytecount |
|
|
277 | repo.ui.progress(_('bundle'), seen, total=bytecount, | |
|
278 | unit=_('bytes')) | |
|
278 | 279 | yield chunk |
|
279 | 280 | |
|
280 | 281 | repo.ui.progress(_('bundle'), None) |
@@ -294,7 +295,7 b' def consumev1(repo, fp, filecount, bytec' | |||
|
294 | 295 | repo.ui.status(_('%d files to transfer, %s of data\n') % |
|
295 | 296 | (filecount, util.bytecount(bytecount))) |
|
296 | 297 | handled_bytes = 0 |
|
297 | repo.ui.progress(_('clone'), 0, total=bytecount) | |
|
298 | repo.ui.progress(_('clone'), 0, total=bytecount, unit=_('bytes')) | |
|
298 | 299 | start = time.time() |
|
299 | 300 | |
|
300 | 301 | with repo.transaction('clone'): |
@@ -317,7 +318,7 b' def consumev1(repo, fp, filecount, bytec' | |||
|
317 | 318 | for chunk in util.filechunkiter(fp, limit=size): |
|
318 | 319 | handled_bytes += len(chunk) |
|
319 | 320 | repo.ui.progress(_('clone'), handled_bytes, |
|
320 | total=bytecount) | |
|
321 | total=bytecount, unit=_('bytes')) | |
|
321 | 322 | ofp.write(chunk) |
|
322 | 323 | |
|
323 | 324 | # Writing straight to files circumvented the inmemory caches |
General Comments 0
You need to be logged in to leave comments.
Login now