##// END OF EJS Templates
py3: add 13 new passing tests to whitelist...
py3: add 13 new passing tests to whitelist Differential Revision: https://phab.mercurial-scm.org/D3615

File last commit:

r37783:ab04972a default
r38102:42476b8b default
Show More
blobstore.py
580 lines | 21.2 KiB | text/x-python | PythonLexer
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 # blobstore.py - local and remote (speaking Git-LFS protocol) blob storages
#
# Copyright 2017 Facebook, Inc.
#
# This software may be used and distributed according to the terms of the
# GNU General Public License version 2 or any later version.
from __future__ import absolute_import
Matt Harbison
lfs: add the ability to disable the usercache...
r37535 import errno
Matt Harbison
lfs: verify lfs object content when transferring to and from the remote store...
r35492 import hashlib
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 import json
import os
import re
Matt Harbison
lfs: narrow the exceptions that trigger a transfer retry...
r35491 import socket
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097
Matt Harbison
lfs: quiesce check-module-import warnings...
r35098 from mercurial.i18n import _
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 from mercurial import (
error,
Matt Harbison
lfs: override walk() in lfsvfs...
r35363 pathutil,
Augie Fackler
lfs: add some bytestring wrappers in blobstore.py...
r36619 pycompat,
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 url as urlmod,
util,
vfs as vfsmod,
Wojciech Lis
lfs: using workers in lfs prefetch...
r35449 worker,
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 )
Matt Harbison
lfs: introduce a user level cache for lfs files...
r35281 from ..largefiles import lfutil
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 # 64 bytes for SHA256
Pulkit Goyal
py3: make sure regexes are bytes...
r36473 _lfsre = re.compile(br'\A[a-f0-9]{64}\Z')
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097
class lfsvfs(vfsmod.vfs):
def join(self, path):
"""split the path at first two characters, like: XX/XXXXX..."""
if not _lfsre.match(path):
raise error.ProgrammingError('unexpected lfs path: %s' % path)
return super(lfsvfs, self).join(path[0:2], path[2:])
Matt Harbison
lfs: override walk() in lfsvfs...
r35363 def walk(self, path=None, onerror=None):
Matt Harbison
lfs: correct the directory list value returned by lfsvfs.walk()...
r35397 """Yield (dirpath, [], oids) tuple for blobs under path
Matt Harbison
lfs: override walk() in lfsvfs...
r35363
Oids only exist in the root of this vfs, so dirpath is always ''.
"""
root = os.path.normpath(self.base)
# when dirpath == root, dirpath[prefixlen:] becomes empty
# because len(dirpath) < prefixlen.
prefixlen = len(pathutil.normasprefix(root))
oids = []
for dirpath, dirs, files in os.walk(self.reljoin(self.base, path or ''),
onerror=onerror):
dirpath = dirpath[prefixlen:]
# Silently skip unexpected files and directories
if len(dirpath) == 2:
oids.extend([dirpath + f for f in files
if _lfsre.match(dirpath + f)])
Matt Harbison
lfs: correct the directory list value returned by lfsvfs.walk()...
r35397 yield ('', [], oids)
Matt Harbison
lfs: override walk() in lfsvfs...
r35363
Matt Harbison
lfs: add the ability to disable the usercache...
r37535 class nullvfs(lfsvfs):
def __init__(self):
pass
def exists(self, oid):
return False
def read(self, oid):
# store.read() calls into here if the blob doesn't exist in its
# self.vfs. Raise the same error as a normal vfs when asked to read a
# file that doesn't exist. The only difference is the full file path
# isn't available in the error.
raise IOError(errno.ENOENT, '%s: No such file or directory' % oid)
def walk(self, path=None, onerror=None):
return ('', [], [])
def write(self, oid, data):
pass
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 class filewithprogress(object):
"""a file-like object that supports __len__ and read.
Useful to provide progress information for how many bytes are read.
"""
def __init__(self, fp, callback):
self._fp = fp
self._callback = callback # func(readsize)
fp.seek(0, os.SEEK_END)
self._len = fp.tell()
fp.seek(0)
def __len__(self):
return self._len
def read(self, size):
if self._fp is None:
return b''
data = self._fp.read(size)
if data:
if self._callback:
self._callback(len(data))
else:
self._fp.close()
self._fp = None
return data
class local(object):
"""Local blobstore for large file contents.
This blobstore is used both as a cache and as a staging area for large blobs
to be uploaded to the remote blobstore.
"""
def __init__(self, repo):
fullpath = repo.svfs.join('lfs/objects')
self.vfs = lfsvfs(fullpath)
Matt Harbison
lfs: special case the null:// usercache instead of treating it as a url...
r37580
if repo.ui.configbool('experimental', 'lfs.disableusercache'):
Matt Harbison
lfs: add the ability to disable the usercache...
r37535 self.cachevfs = nullvfs()
else:
Matt Harbison
lfs: special case the null:// usercache instead of treating it as a url...
r37580 usercache = lfutil._usercachedir(repo.ui, 'lfs')
self.cachevfs = lfsvfs(usercache)
Matt Harbison
lfs: add note messages indicating what store holds the lfs blob...
r35489 self.ui = repo.ui
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097
Matt Harbison
lfs: add a local store method for opening a blob...
r35543 def open(self, oid):
"""Open a read-only file descriptor to the named blob, in either the
usercache or the local store."""
Matt Harbison
lfs: add a comment to describe subtle local blobstore open() behavior
r35555 # The usercache is the most likely place to hold the file. Commit will
# write to both it and the local store, as will anything that downloads
# the blobs. However, things like clone without an update won't
# populate the local store. For an init + push of a local clone,
# the usercache is the only place it _could_ be. If not present, the
# missing file msg here will indicate the local repo, not the usercache.
Matt Harbison
lfs: add a local store method for opening a blob...
r35543 if self.cachevfs.exists(oid):
return self.cachevfs(oid, 'rb')
return self.vfs(oid, 'rb')
Matt Harbison
lfs: introduce a localstore method for downloading from remote stores...
r35565 def download(self, oid, src):
"""Read the blob from the remote source in chunks, verify the content,
and write to this local blobstore."""
sha256 = hashlib.sha256()
with self.vfs(oid, 'wb', atomictemp=True) as fp:
for chunk in util.filechunkiter(src, size=1048576):
fp.write(chunk)
sha256.update(chunk)
realoid = sha256.hexdigest()
if realoid != oid:
Matt Harbison
lfs: gracefully handle aborts on the server when corrupt blobs are detected...
r37710 raise LfsCorruptionError(_('corrupt remote lfs object: %s')
% oid)
Matt Harbison
lfs: introduce a localstore method for downloading from remote stores...
r35565
Matt Harbison
lfs: add the ability to disable the usercache...
r37535 self._linktousercache(oid)
Matt Harbison
lfs: introduce a localstore method for downloading from remote stores...
r35565
Matt Harbison
lfs: remove the verification option when writing to the local store...
r35567 def write(self, oid, data):
"""Write blob to local blobstore.
Matt Harbison
lfs: verify lfs object content when transferring to and from the remote store...
r35492
Matt Harbison
lfs: remove the verification option when writing to the local store...
r35567 This should only be called from the filelog during a commit or similar.
As such, there is no need to verify the data. Imports from a remote
store must use ``download()`` instead."""
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 with self.vfs(oid, 'wb', atomictemp=True) as fp:
fp.write(data)
Matt Harbison
lfs: add the ability to disable the usercache...
r37535 self._linktousercache(oid)
def _linktousercache(self, oid):
Matt Harbison
lfs: introduce a user level cache for lfs files...
r35281 # XXX: should we verify the content of the cache, and hardlink back to
# the local store on success, but truncate, write and link on failure?
Matt Harbison
lfs: add the ability to disable the usercache...
r37535 if (not self.cachevfs.exists(oid)
and not isinstance(self.cachevfs, nullvfs)):
Matt Harbison
lfs: remove the verification option when writing to the local store...
r35567 self.ui.note(_('lfs: adding %s to the usercache\n') % oid)
lfutil.link(self.vfs.join(oid), self.cachevfs.join(oid))
Matt Harbison
lfs: introduce a user level cache for lfs files...
r35281
Matt Harbison
lfs: verify lfs object content when transferring to and from the remote store...
r35492 def read(self, oid, verify=True):
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 """Read blob from local blobstore."""
Matt Harbison
lfs: introduce a user level cache for lfs files...
r35281 if not self.vfs.exists(oid):
Matt Harbison
lfs: verify lfs object content when transferring to and from the remote store...
r35492 blob = self._read(self.cachevfs, oid, verify)
Matt Harbison
lfs: only hardlink between the usercache and local store if the blob verifies...
r35493
# Even if revlog will verify the content, it needs to be verified
# now before making the hardlink to avoid propagating corrupt blobs.
# Don't abort if corruption is detected, because `hg verify` will
# give more useful info about the corruption- simply don't add the
# hardlink.
if verify or hashlib.sha256(blob).hexdigest() == oid:
self.ui.note(_('lfs: found %s in the usercache\n') % oid)
lfutil.link(self.cachevfs.join(oid), self.vfs.join(oid))
Matt Harbison
lfs: add note messages indicating what store holds the lfs blob...
r35489 else:
self.ui.note(_('lfs: found %s in the local lfs store\n') % oid)
Matt Harbison
lfs: verify lfs object content when transferring to and from the remote store...
r35492 blob = self._read(self.vfs, oid, verify)
return blob
def _read(self, vfs, oid, verify):
"""Read blob (after verifying) from the given store"""
blob = vfs.read(oid)
if verify:
_verify(oid, blob)
return blob
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097
Matt Harbison
lfs: add a blob verification method to the local store...
r37163 def verify(self, oid):
"""Indicate whether or not the hash of the underlying file matches its
name."""
sha256 = hashlib.sha256()
with self.open(oid) as fp:
for chunk in util.filechunkiter(fp, size=1048576):
sha256.update(chunk)
return oid == sha256.hexdigest()
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 def has(self, oid):
"""Returns True if the local blobstore contains the requested blob,
False otherwise."""
Matt Harbison
lfs: introduce a user level cache for lfs files...
r35281 return self.cachevfs.exists(oid) or self.vfs.exists(oid)
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097
class _gitlfsremote(object):
def __init__(self, repo, url):
ui = repo.ui
self.ui = ui
baseurl, authinfo = url.authinfo()
self.baseurl = baseurl.rstrip('/')
Matt Harbison
lfs: add an experimental config to override User-Agent for the blob transfer...
r35456 useragent = repo.ui.config('experimental', 'lfs.user-agent')
if not useragent:
Matt Harbison
lfs: default the User-Agent header for blob transfers to 'git-lfs'...
r35751 useragent = 'git-lfs/2.3.4 (Mercurial %s)' % util.version()
Matt Harbison
lfs: add git to the User-Agent header for blob transfers...
r35455 self.urlopener = urlmod.opener(ui, authinfo, useragent)
Matt Harbison
lfs: register config options...
r35099 self.retry = ui.configint('lfs', 'retry')
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097
def writebatch(self, pointers, fromstore):
"""Batch upload from local to remote blobstore."""
Matt Harbison
lfs: deduplicate oids in the transfer...
r35945 self._batch(_deduplicate(pointers), fromstore, 'upload')
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097
def readbatch(self, pointers, tostore):
"""Batch download from remote to local blostore."""
Matt Harbison
lfs: deduplicate oids in the transfer...
r35945 self._batch(_deduplicate(pointers), tostore, 'download')
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097
def _batchrequest(self, pointers, action):
"""Get metadata about objects pointed by pointers for given action
Return decoded JSON object like {'objects': [{'oid': '', 'size': 1}]}
See https://github.com/git-lfs/git-lfs/blob/master/docs/api/batch.md
"""
objects = [{'oid': p.oid(), 'size': p.size()} for p in pointers]
requestdata = json.dumps({
'objects': objects,
'operation': action,
})
batchreq = util.urlreq.request('%s/objects/batch' % self.baseurl,
data=requestdata)
batchreq.add_header('Accept', 'application/vnd.git-lfs+json')
batchreq.add_header('Content-Type', 'application/vnd.git-lfs+json')
try:
Matt Harbison
lfs: debug print HTTP headers and JSON payload received from the server...
r36944 rsp = self.urlopener.open(batchreq)
rawjson = rsp.read()
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 except util.urlerr.httperror as ex:
raise LfsRemoteError(_('LFS HTTP error: %s (action=%s)')
% (ex, action))
try:
response = json.loads(rawjson)
except ValueError:
raise LfsRemoteError(_('LFS server returns invalid JSON: %s')
% rawjson)
Matt Harbison
lfs: debug print HTTP headers and JSON payload received from the server...
r36944
if self.ui.debugflag:
self.ui.debug('Status: %d\n' % rsp.status)
# lfs-test-server and hg serve return headers in different order
self.ui.debug('%s\n'
% '\n'.join(sorted(str(rsp.info()).splitlines())))
if 'objects' in response:
response['objects'] = sorted(response['objects'],
key=lambda p: p['oid'])
self.ui.debug('%s\n'
Matt Harbison
test-lfs: drop trailing ', ' item separators from debug JSON output...
r37164 % json.dumps(response, indent=2,
separators=('', ': '), sort_keys=True))
Matt Harbison
lfs: debug print HTTP headers and JSON payload received from the server...
r36944
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 return response
Jun Wu
lfs: remove internal url in test...
r35684 def _checkforservererror(self, pointers, responses, action):
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 """Scans errors from objects
Matt Harbison
lfs: correct documentation typo
r35712 Raises LfsRemoteError if any objects have an error"""
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 for response in responses:
Jun Wu
lfs: remove internal url in test...
r35684 # The server should return 404 when objects cannot be found. Some
# server implementation (ex. lfs-test-server) does not set "error"
# but just removes "download" from "actions". Treat that case
# as the same as 404 error.
Matt Harbison
lfs: improve the client message when the server signals an object error...
r37259 if 'error' not in response:
if (action == 'download'
and action not in response.get('actions', [])):
code = 404
Matt Harbison
lfs: raise an error if the server sends an unsolicited oid...
r35713 else:
Matt Harbison
lfs: improve the client message when the server signals an object error...
r37259 continue
else:
# An error dict without a code doesn't make much sense, so
# treat as a server error.
code = response.get('error').get('code', 500)
ptrmap = {p.oid(): p for p in pointers}
p = ptrmap.get(response['oid'], None)
if p:
filename = getattr(p, 'filename', 'unknown')
errors = {
404: 'The object does not exist',
410: 'The object was removed by the owner',
422: 'Validation error',
500: 'Internal server error',
}
msg = errors.get(code, 'status code %d' % code)
raise LfsRemoteError(_('LFS server error for "%s": %s')
% (filename, msg))
else:
raise LfsRemoteError(
_('LFS server error. Unsolicited response for oid %s')
% response['oid'])
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097
def _extractobjects(self, response, pointers, action):
"""extract objects from response of the batch API
response: parsed JSON object returned by batch API
return response['objects'] filtered by action
raise if any object has an error
"""
# Scan errors from objects - fail early
objects = response.get('objects', [])
Jun Wu
lfs: remove internal url in test...
r35684 self._checkforservererror(pointers, objects, action)
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097
# Filter objects with given action. Practically, this skips uploading
# objects which exist in the server.
filteredobjects = [o for o in objects if action in o.get('actions', [])]
return filteredobjects
Wojciech Lis
lfs: using workers in lfs prefetch...
r35449 def _basictransfer(self, obj, action, localstore):
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 """Download or upload a single object using basic transfer protocol
obj: dict, an object description returned by batch API
action: string, one of ['upload', 'download']
localstore: blobstore.local
See https://github.com/git-lfs/git-lfs/blob/master/docs/api/\
basic-transfers.md
"""
Augie Fackler
lfs: add some bytestring wrappers in blobstore.py...
r36619 oid = pycompat.bytestr(obj['oid'])
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097
Augie Fackler
lfs: add some bytestring wrappers in blobstore.py...
r36619 href = pycompat.bytestr(obj['actions'][action].get('href'))
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 headers = obj['actions'][action].get('header', {}).items()
request = util.urlreq.request(href)
if action == 'upload':
# If uploading blobs, read data from local blobstore.
Matt Harbison
lfs: drop a duplicate blob verification method
r37234 if not localstore.verify(oid):
raise error.Abort(_('detected corrupt lfs object: %s') % oid,
hint=_('run hg verify'))
Matt Harbison
lfs: use the local store method for opening a blob...
r35544 request.data = filewithprogress(localstore.open(oid), None)
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 request.get_method = lambda: 'PUT'
Matt Harbison
lfs: add the 'Content-Type' header called out in the file transfer spec...
r37260 request.add_header('Content-Type', 'application/octet-stream')
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097
for k, v in headers:
request.add_header(k, v)
response = b''
try:
req = self.urlopener.open(request)
Matt Harbison
lfs: debug print HTTP headers and JSON payload received from the server...
r36944
if self.ui.debugflag:
self.ui.debug('Status: %d\n' % req.status)
# lfs-test-server and hg serve return headers in different order
self.ui.debug('%s\n'
% '\n'.join(sorted(str(req.info()).splitlines())))
Matt Harbison
lfs: use the localstore download method to transfer from remote stores...
r35566 if action == 'download':
# If downloading blobs, store downloaded data to local blobstore
localstore.download(oid, req)
else:
while True:
data = req.read(1048576)
if not data:
break
response += data
if response:
self.ui.debug('lfs %s response: %s' % (action, response))
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 except util.urlerr.httperror as ex:
Matt Harbison
lfs: dump the full response on httperror in debug mode...
r35752 if self.ui.debugflag:
Matt Harbison
lfs: separate a debug message from the subsequent abort message
r35771 self.ui.debug('%s: %s\n' % (oid, ex.read()))
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 raise LfsRemoteError(_('HTTP error: %s (oid=%s, action=%s)')
% (ex, oid, action))
def _batch(self, pointers, localstore, action):
if action not in ['upload', 'download']:
raise error.ProgrammingError('invalid Git-LFS action: %s' % action)
response = self._batchrequest(pointers, action)
objects = self._extractobjects(response, pointers, action)
total = sum(x.get('size', 0) for x in objects)
Wojciech Lis
lfs: using workers in lfs prefetch...
r35449 sizes = {}
for obj in objects:
sizes[obj.get('oid')] = obj.get('size', 0)
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 topic = {'upload': _('lfs uploading'),
'download': _('lfs downloading')}[action]
Matt Harbison
lfs: use ui.note() and ui.debug() instead of ui.write() and their flags...
r35494 if len(objects) > 1:
self.ui.note(_('lfs: need to transfer %d objects (%s)\n')
% (len(objects), util.bytecount(total)))
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 self.ui.progress(topic, 0, total=total)
Wojciech Lis
lfs: using workers in lfs prefetch...
r35449 def transfer(chunk):
for obj in chunk:
objsize = obj.get('size', 0)
if self.ui.verbose:
if action == 'download':
msg = _('lfs: downloading %s (%s)\n')
elif action == 'upload':
msg = _('lfs: uploading %s (%s)\n')
Matt Harbison
lfs: use ui.note() and ui.debug() instead of ui.write() and their flags...
r35494 self.ui.note(msg % (obj.get('oid'),
util.bytecount(objsize)))
Wojciech Lis
lfs: using workers in lfs prefetch...
r35449 retry = self.retry
while True:
try:
self._basictransfer(obj, action, localstore)
yield 1, obj.get('oid')
break
Matt Harbison
lfs: narrow the exceptions that trigger a transfer retry...
r35491 except socket.error as ex:
Wojciech Lis
lfs: using workers in lfs prefetch...
r35449 if retry > 0:
Matt Harbison
lfs: use ui.note() and ui.debug() instead of ui.write() and their flags...
r35494 self.ui.note(
_('lfs: failed: %r (remaining retry %d)\n')
% (ex, retry))
Wojciech Lis
lfs: using workers in lfs prefetch...
r35449 retry -= 1
continue
raise
Matt Harbison
lfs: default to not using workers for upload/download...
r35750 # Until https multiplexing gets sorted out
if self.ui.configbool('experimental', 'lfs.worker-enable'):
oids = worker.worker(self.ui, 0.1, transfer, (),
sorted(objects, key=lambda o: o.get('oid')))
else:
oids = transfer(sorted(objects, key=lambda o: o.get('oid')))
Wojciech Lis
lfs: using workers in lfs prefetch...
r35449 processed = 0
Matt Harbison
lfs: emit a status message to indicate how many blobs were uploaded...
r35899 blobs = 0
Wojciech Lis
lfs: using workers in lfs prefetch...
r35449 for _one, oid in oids:
processed += sizes[oid]
Matt Harbison
lfs: emit a status message to indicate how many blobs were uploaded...
r35899 blobs += 1
Wojciech Lis
lfs: using workers in lfs prefetch...
r35449 self.ui.progress(topic, processed, total=total)
Matt Harbison
lfs: use ui.note() and ui.debug() instead of ui.write() and their flags...
r35494 self.ui.note(_('lfs: processed: %s\n') % oid)
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 self.ui.progress(topic, pos=None, total=total)
Matt Harbison
lfs: emit a status message to indicate how many blobs were uploaded...
r35899 if blobs > 0:
if action == 'upload':
self.ui.status(_('lfs: uploaded %d files (%s)\n')
% (blobs, util.bytecount(processed)))
Matt Harbison
lfs: enable the final download count status message...
r37783 elif action == 'download':
self.ui.status(_('lfs: downloaded %d files (%s)\n')
% (blobs, util.bytecount(processed)))
Matt Harbison
lfs: emit a status message to indicate how many blobs were uploaded...
r35899
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 def __del__(self):
# copied from mercurial/httppeer.py
urlopener = getattr(self, 'urlopener', None)
if urlopener:
for h in urlopener.handlers:
h.close()
getattr(h, "close_all", lambda : None)()
class _dummyremote(object):
"""Dummy store storing blobs to temp directory."""
def __init__(self, repo, url):
fullpath = repo.vfs.join('lfs', url.path)
self.vfs = lfsvfs(fullpath)
def writebatch(self, pointers, fromstore):
Matt Harbison
lfs: deduplicate oids in the transfer...
r35945 for p in _deduplicate(pointers):
Matt Harbison
lfs: verify lfs object content when transferring to and from the remote store...
r35492 content = fromstore.read(p.oid(), verify=True)
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 with self.vfs(p.oid(), 'wb', atomictemp=True) as fp:
fp.write(content)
def readbatch(self, pointers, tostore):
Matt Harbison
lfs: deduplicate oids in the transfer...
r35945 for p in _deduplicate(pointers):
Matt Harbison
lfs: use the localstore download method to transfer from remote stores...
r35566 with self.vfs(p.oid(), 'rb') as fp:
tostore.download(p.oid(), fp)
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097
class _nullremote(object):
"""Null store storing blobs to /dev/null."""
def __init__(self, repo, url):
pass
def writebatch(self, pointers, fromstore):
pass
def readbatch(self, pointers, tostore):
pass
class _promptremote(object):
"""Prompt user to set lfs.url when accessed."""
def __init__(self, repo, url):
pass
def writebatch(self, pointers, fromstore, ui=None):
self._prompt()
def readbatch(self, pointers, tostore, ui=None):
self._prompt()
def _prompt(self):
raise error.Abort(_('lfs.url needs to be configured'))
_storemap = {
'https': _gitlfsremote,
'http': _gitlfsremote,
'file': _dummyremote,
'null': _nullremote,
None: _promptremote,
}
Matt Harbison
lfs: deduplicate oids in the transfer...
r35945 def _deduplicate(pointers):
"""Remove any duplicate oids that exist in the list"""
reduced = util.sortdict()
for p in pointers:
reduced[p.oid()] = p
return reduced.values()
Matt Harbison
lfs: verify lfs object content when transferring to and from the remote store...
r35492 def _verify(oid, content):
realoid = hashlib.sha256(content).hexdigest()
if realoid != oid:
Matt Harbison
lfs: gracefully handle aborts on the server when corrupt blobs are detected...
r37710 raise LfsCorruptionError(_('detected corrupt lfs object: %s') % oid,
hint=_('run hg verify'))
Matt Harbison
lfs: verify lfs object content when transferring to and from the remote store...
r35492
Matt Harbison
lfs: infer the blob store URL from an explicit push dest or default-push...
r37582 def remote(repo, remote=None):
Matt Harbison
lfs: infer the blob store URL from paths.default...
r37536 """remotestore factory. return a store in _storemap depending on config
If ``lfs.url`` is specified, use that remote endpoint. Otherwise, try to
infer the endpoint, based on the remote repository using the same path
adjustments as git. As an extension, 'http' is supported as well so that
``hg serve`` works out of the box.
https://github.com/git-lfs/git-lfs/blob/master/docs/api/server-discovery.md
"""
Matt Harbison
lfs: handle paths that don't end with '/' when inferring the blob store...
r37583 lfsurl = repo.ui.config('lfs', 'url')
url = util.url(lfsurl or '')
if lfsurl is None:
Matt Harbison
lfs: infer the blob store URL from an explicit push dest or default-push...
r37582 if remote:
Matt Harbison
lfs: handle paths that don't end with '/' when inferring the blob store...
r37583 path = remote
Matt Harbison
lfs: infer the blob store URL from an explicit push dest or default-push...
r37582 elif util.safehasattr(repo, '_subtoppath'):
Matt Harbison
lfs: infer the blob store URL from an explicit pull source...
r37581 # The pull command sets this during the optional update phase, which
# tells exactly where the pull originated, whether 'paths.default'
# or explicit.
Matt Harbison
lfs: handle paths that don't end with '/' when inferring the blob store...
r37583 path = repo._subtoppath
Matt Harbison
lfs: infer the blob store URL from an explicit pull source...
r37581 else:
# TODO: investigate 'paths.remote:lfsurl' style path customization,
# and fall back to inferring from 'paths.remote' if unspecified.
Matt Harbison
lfs: handle paths that don't end with '/' when inferring the blob store...
r37583 path = repo.ui.config('paths', 'default') or ''
defaulturl = util.url(path)
Matt Harbison
lfs: infer the blob store URL from paths.default...
r37536
# TODO: support local paths as well.
# TODO: consider the ssh -> https transformation that git applies
if defaulturl.scheme in (b'http', b'https'):
Matt Harbison
lfs: handle paths that don't end with '/' when inferring the blob store...
r37583 if defaulturl.path and defaulturl.path[:-1] != b'/':
defaulturl.path += b'/'
Matt Harbison
lfs: fix the inferred remote store path when using a --prefix...
r37709 defaulturl.path = (defaulturl.path or b'') + b'.git/info/lfs'
Matt Harbison
lfs: infer the blob store URL from paths.default...
r37536
url = util.url(bytes(defaulturl))
repo.ui.note(_('lfs: assuming remote store: %s\n') % url)
Matt Harbison
lfs: import the Facebook git-lfs client extension...
r35097 scheme = url.scheme
if scheme not in _storemap:
raise error.Abort(_('lfs: unknown url scheme: %s') % scheme)
return _storemap[scheme](repo, url)
class LfsRemoteError(error.RevlogError):
pass
Matt Harbison
lfs: gracefully handle aborts on the server when corrupt blobs are detected...
r37710
class LfsCorruptionError(error.Abort):
"""Raised when a corrupt blob is detected, aborting an operation
It exists to allow specialized handling on the server side."""