Show More
@@ -0,0 +1,4 b'' | |||||
|
1 | Greg Ward, author of the original bfiles extension | |||
|
2 | Na'Tosha Bard of Unity Technologies | |||
|
3 | Fog Creek Software | |||
|
4 | Special thanks to the University of Toronto and the UCOSP program |
@@ -0,0 +1,40 b'' | |||||
|
1 | # Copyright 2009-2010 Gregory P. Ward | |||
|
2 | # Copyright 2009-2010 Intelerad Medical Systems Incorporated | |||
|
3 | # Copyright 2010-2011 Fog Creek Software | |||
|
4 | # Copyright 2010-2011 Unity Technologies | |||
|
5 | # | |||
|
6 | # This software may be used and distributed according to the terms of the | |||
|
7 | # GNU General Public License version 2 or any later version. | |||
|
8 | ||||
|
9 | '''track large binary files | |||
|
10 | ||||
|
11 | Large binary files tend to be not very compressible, not very "diffable", and | |||
|
12 | not at all mergeable. Such files are not handled well by Mercurial\'s storage | |||
|
13 | format (revlog), which is based on compressed binary deltas. largefiles solves | |||
|
14 | this problem by adding a centralized client-server layer on top of Mercurial: | |||
|
15 | largefiles live in a *central store* out on the network somewhere, and you only | |||
|
16 | fetch the ones that you need when you need them. | |||
|
17 | ||||
|
18 | largefiles works by maintaining a *standin* in .hglf/ for each largefile. The | |||
|
19 | standins are small (41 bytes: an SHA-1 hash plus newline) and are tracked by | |||
|
20 | Mercurial. Largefile revisions are identified by the SHA-1 hash of their | |||
|
21 | contents, which is written to the standin. largefiles uses that revision ID to | |||
|
22 | get/put largefile revisions from/to the central store. | |||
|
23 | ||||
|
24 | A complete tutorial for using lfiles is included in ``usage.txt`` in the lfiles | |||
|
25 | source distribution. See | |||
|
26 | https://developers.kilnhg.com/Repo/Kiln/largefiles/largefiles/File/usage.txt | |||
|
27 | ''' | |||
|
28 | ||||
|
29 | from mercurial import commands | |||
|
30 | ||||
|
31 | import lfcommands | |||
|
32 | import reposetup | |||
|
33 | import uisetup | |||
|
34 | ||||
|
35 | reposetup = reposetup.reposetup | |||
|
36 | uisetup = uisetup.uisetup | |||
|
37 | ||||
|
38 | commands.norepo += " lfconvert" | |||
|
39 | ||||
|
40 | cmdtable = lfcommands.cmdtable |
@@ -0,0 +1,201 b'' | |||||
|
1 | # Copyright 2009-2010 Gregory P. Ward | |||
|
2 | # Copyright 2009-2010 Intelerad Medical Systems Incorporated | |||
|
3 | # Copyright 2010-2011 Fog Creek Software | |||
|
4 | # Copyright 2010-2011 Unity Technologies | |||
|
5 | # | |||
|
6 | # This software may be used and distributed according to the terms of the | |||
|
7 | # GNU General Public License version 2 or any later version. | |||
|
8 | ||||
|
9 | '''Base class for store implementations and store-related utility code.''' | |||
|
10 | ||||
|
11 | import os | |||
|
12 | import tempfile | |||
|
13 | import binascii | |||
|
14 | import re | |||
|
15 | ||||
|
16 | from mercurial import util, node, hg | |||
|
17 | from mercurial.i18n import _ | |||
|
18 | ||||
|
19 | import lfutil | |||
|
20 | ||||
|
21 | class StoreError(Exception): | |||
|
22 | '''Raised when there is a problem getting files from or putting | |||
|
23 | files to a central store.''' | |||
|
24 | def __init__(self, filename, hash, url, detail): | |||
|
25 | self.filename = filename | |||
|
26 | self.hash = hash | |||
|
27 | self.url = url | |||
|
28 | self.detail = detail | |||
|
29 | ||||
|
30 | def longmessage(self): | |||
|
31 | if self.url: | |||
|
32 | return ('%s: %s\n' | |||
|
33 | '(failed URL: %s)\n' | |||
|
34 | % (self.filename, self.detail, self.url)) | |||
|
35 | else: | |||
|
36 | return ('%s: %s\n' | |||
|
37 | '(no default or default-push path set in hgrc)\n' | |||
|
38 | % (self.filename, self.detail)) | |||
|
39 | ||||
|
40 | def __str__(self): | |||
|
41 | return "%s: %s" % (self.url, self.detail) | |||
|
42 | ||||
|
43 | class basestore(object): | |||
|
44 | def __init__(self, ui, repo, url): | |||
|
45 | self.ui = ui | |||
|
46 | self.repo = repo | |||
|
47 | self.url = url | |||
|
48 | ||||
|
49 | def put(self, source, hash): | |||
|
50 | '''Put source file into the store under <filename>/<hash>.''' | |||
|
51 | raise NotImplementedError('abstract method') | |||
|
52 | ||||
|
53 | def exists(self, hash): | |||
|
54 | '''Check to see if the store contains the given hash.''' | |||
|
55 | raise NotImplementedError('abstract method') | |||
|
56 | ||||
|
57 | def get(self, files): | |||
|
58 | '''Get the specified largefiles from the store and write to local | |||
|
59 | files under repo.root. files is a list of (filename, hash) | |||
|
60 | tuples. Return (success, missing), lists of files successfuly | |||
|
61 | downloaded and those not found in the store. success is a list | |||
|
62 | of (filename, hash) tuples; missing is a list of filenames that | |||
|
63 | we could not get. (The detailed error message will already have | |||
|
64 | been presented to the user, so missing is just supplied as a | |||
|
65 | summary.)''' | |||
|
66 | success = [] | |||
|
67 | missing = [] | |||
|
68 | ui = self.ui | |||
|
69 | ||||
|
70 | at = 0 | |||
|
71 | for filename, hash in files: | |||
|
72 | ui.progress(_('getting largefiles'), at, unit='lfile', | |||
|
73 | total=len(files)) | |||
|
74 | at += 1 | |||
|
75 | ui.note(_('getting %s:%s\n') % (filename, hash)) | |||
|
76 | ||||
|
77 | cachefilename = lfutil.cachepath(self.repo, hash) | |||
|
78 | cachedir = os.path.dirname(cachefilename) | |||
|
79 | ||||
|
80 | # No need to pass mode='wb' to fdopen(), since mkstemp() already | |||
|
81 | # opened the file in binary mode. | |||
|
82 | (tmpfd, tmpfilename) = tempfile.mkstemp( | |||
|
83 | dir=cachedir, prefix=os.path.basename(filename)) | |||
|
84 | tmpfile = os.fdopen(tmpfd, 'w') | |||
|
85 | ||||
|
86 | try: | |||
|
87 | hhash = binascii.hexlify(self._getfile(tmpfile, filename, hash)) | |||
|
88 | except StoreError, err: | |||
|
89 | ui.warn(err.longmessage()) | |||
|
90 | hhash = "" | |||
|
91 | ||||
|
92 | if hhash != hash: | |||
|
93 | if hhash != "": | |||
|
94 | ui.warn(_('%s: data corruption (expected %s, got %s)\n') | |||
|
95 | % (filename, hash, hhash)) | |||
|
96 | tmpfile.close() # no-op if it's already closed | |||
|
97 | os.remove(tmpfilename) | |||
|
98 | missing.append(filename) | |||
|
99 | continue | |||
|
100 | ||||
|
101 | if os.path.exists(cachefilename): # Windows | |||
|
102 | os.remove(cachefilename) | |||
|
103 | os.rename(tmpfilename, cachefilename) | |||
|
104 | lfutil.linktosystemcache(self.repo, hash) | |||
|
105 | success.append((filename, hhash)) | |||
|
106 | ||||
|
107 | ui.progress(_('getting largefiles'), None) | |||
|
108 | return (success, missing) | |||
|
109 | ||||
|
110 | def verify(self, revs, contents=False): | |||
|
111 | '''Verify the existence (and, optionally, contents) of every big | |||
|
112 | file revision referenced by every changeset in revs. | |||
|
113 | Return 0 if all is well, non-zero on any errors.''' | |||
|
114 | write = self.ui.write | |||
|
115 | failed = False | |||
|
116 | ||||
|
117 | write(_('searching %d changesets for largefiles\n') % len(revs)) | |||
|
118 | verified = set() # set of (filename, filenode) tuples | |||
|
119 | ||||
|
120 | for rev in revs: | |||
|
121 | cctx = self.repo[rev] | |||
|
122 | cset = "%d:%s" % (cctx.rev(), node.short(cctx.node())) | |||
|
123 | ||||
|
124 | failed = lfutil.any_(self._verifyfile( | |||
|
125 | cctx, cset, contents, standin, verified) for standin in cctx) | |||
|
126 | ||||
|
127 | num_revs = len(verified) | |||
|
128 | num_lfiles = len(set([fname for (fname, fnode) in verified])) | |||
|
129 | if contents: | |||
|
130 | write(_('verified contents of %d revisions of %d largefiles\n') | |||
|
131 | % (num_revs, num_lfiles)) | |||
|
132 | else: | |||
|
133 | write(_('verified existence of %d revisions of %d largefiles\n') | |||
|
134 | % (num_revs, num_lfiles)) | |||
|
135 | ||||
|
136 | return int(failed) | |||
|
137 | ||||
|
138 | def _getfile(self, tmpfile, filename, hash): | |||
|
139 | '''Fetch one revision of one file from the store and write it | |||
|
140 | to tmpfile. Compute the hash of the file on-the-fly as it | |||
|
141 | downloads and return the binary hash. Close tmpfile. Raise | |||
|
142 | StoreError if unable to download the file (e.g. it does not | |||
|
143 | exist in the store).''' | |||
|
144 | raise NotImplementedError('abstract method') | |||
|
145 | ||||
|
146 | def _verifyfile(self, cctx, cset, contents, standin, verified): | |||
|
147 | '''Perform the actual verification of a file in the store. | |||
|
148 | ''' | |||
|
149 | raise NotImplementedError('abstract method') | |||
|
150 | ||||
|
151 | import localstore, wirestore | |||
|
152 | ||||
|
153 | _storeprovider = { | |||
|
154 | 'file': [localstore.localstore], | |||
|
155 | 'http': [wirestore.wirestore], | |||
|
156 | 'https': [wirestore.wirestore], | |||
|
157 | 'ssh': [wirestore.wirestore], | |||
|
158 | } | |||
|
159 | ||||
|
160 | _scheme_re = re.compile(r'^([a-zA-Z0-9+-.]+)://') | |||
|
161 | ||||
|
162 | # During clone this function is passed the src's ui object | |||
|
163 | # but it needs the dest's ui object so it can read out of | |||
|
164 | # the config file. Use repo.ui instead. | |||
|
165 | def _openstore(repo, remote=None, put=False): | |||
|
166 | ui = repo.ui | |||
|
167 | ||||
|
168 | if not remote: | |||
|
169 | path = getattr(repo, 'lfpullsource', None) or \ | |||
|
170 | ui.expandpath('default-push', 'default') | |||
|
171 | # If 'default-push' and 'default' can't be expanded | |||
|
172 | # they are just returned. In that case use the empty string which | |||
|
173 | # use the filescheme. | |||
|
174 | if path == 'default-push' or path == 'default': | |||
|
175 | path = '' | |||
|
176 | remote = repo | |||
|
177 | else: | |||
|
178 | remote = hg.peer(repo, {}, path) | |||
|
179 | ||||
|
180 | # The path could be a scheme so use Mercurial's normal functionality | |||
|
181 | # to resolve the scheme to a repository and use its path | |||
|
182 | path = hasattr(remote, 'url') and remote.url() or remote.path | |||
|
183 | ||||
|
184 | match = _scheme_re.match(path) | |||
|
185 | if not match: # regular filesystem path | |||
|
186 | scheme = 'file' | |||
|
187 | else: | |||
|
188 | scheme = match.group(1) | |||
|
189 | ||||
|
190 | try: | |||
|
191 | storeproviders = _storeprovider[scheme] | |||
|
192 | except KeyError: | |||
|
193 | raise util.Abort(_('unsupported URL scheme %r') % scheme) | |||
|
194 | ||||
|
195 | for class_obj in storeproviders: | |||
|
196 | try: | |||
|
197 | return class_obj(ui, repo, remote) | |||
|
198 | except lfutil.storeprotonotcapable: | |||
|
199 | pass | |||
|
200 | ||||
|
201 | raise util.Abort(_('%s does not appear to be a lfile store'), path) |
@@ -0,0 +1,49 b'' | |||||
|
1 | = largefiles - manage large binary files = | |||
|
2 | This extension is based off of Greg Ward's bfiles extension which can be found | |||
|
3 | at http://mercurial.selenic.com/wiki/BfilesExtension. | |||
|
4 | ||||
|
5 | == The largefile store == | |||
|
6 | ||||
|
7 | largefile stores are, in the typical use case, centralized servers that have | |||
|
8 | every past revision of a given binary file. Each largefile is identified by | |||
|
9 | its sha1 hash, and all interactions with the store take one of the following | |||
|
10 | forms. | |||
|
11 | ||||
|
12 | -Download a bfile with this hash | |||
|
13 | -Upload a bfile with this hash | |||
|
14 | -Check if the store has a bfile with this hash | |||
|
15 | ||||
|
16 | largefiles stores can take one of two forms: | |||
|
17 | ||||
|
18 | -Directories on a network file share | |||
|
19 | -Mercurial wireproto servers, either via ssh or http (hgweb) | |||
|
20 | ||||
|
21 | == The Local Repository == | |||
|
22 | ||||
|
23 | The local repository has a largefile cache in .hg/largefiles which holds a | |||
|
24 | subset of the largefiles needed. On a clone only the largefiles at tip are | |||
|
25 | downloaded. When largefiles are downloaded from the central store, a copy is | |||
|
26 | saved in this store. | |||
|
27 | ||||
|
28 | == The Global Cache == | |||
|
29 | ||||
|
30 | largefiles in a local repository cache are hardlinked to files in the global | |||
|
31 | cache. Before a file is downloaded we check if it is in the global cache. | |||
|
32 | ||||
|
33 | == Implementation Details == | |||
|
34 | ||||
|
35 | Each largefile has a standin which is in .hglf. The standin is tracked by | |||
|
36 | Mercurial. The standin contains the SHA1 hash of the largefile. When a | |||
|
37 | largefile is added/removed/copied/renamed/etc the same operation is applied to | |||
|
38 | the standin. Thus the history of the standin is the history of the largefile. | |||
|
39 | ||||
|
40 | For performance reasons, the contents of a standin are only updated before a | |||
|
41 | commit. Standins are added/removed/copied/renamed from add/remove/copy/rename | |||
|
42 | Mercurial commands but their contents will not be updated. The contents of a | |||
|
43 | standin will always be the hash of the largefile as of the last commit. To | |||
|
44 | support some commands (revert) some standins are temporarily updated but will | |||
|
45 | be changed back after the command is finished. | |||
|
46 | ||||
|
47 | A Mercurial dirstate object tracks the state of the largefiles. The dirstate | |||
|
48 | uses the last modified time and current size to detect if a file has changed | |||
|
49 | (without reading the entire contents of the file). |
@@ -0,0 +1,483 b'' | |||||
|
1 | # Copyright 2009-2010 Gregory P. Ward | |||
|
2 | # Copyright 2009-2010 Intelerad Medical Systems Incorporated | |||
|
3 | # Copyright 2010-2011 Fog Creek Software | |||
|
4 | # Copyright 2010-2011 Unity Technologies | |||
|
5 | # | |||
|
6 | # This software may be used and distributed according to the terms of the | |||
|
7 | # GNU General Public License version 2 or any later version. | |||
|
8 | ||||
|
9 | '''High-level command functions: lfadd() et. al, plus the cmdtable.''' | |||
|
10 | ||||
|
11 | import os | |||
|
12 | import shutil | |||
|
13 | ||||
|
14 | from mercurial import util, match as match_, hg, node, context, error | |||
|
15 | from mercurial.i18n import _ | |||
|
16 | ||||
|
17 | import lfutil | |||
|
18 | import basestore | |||
|
19 | ||||
|
20 | # -- Commands ---------------------------------------------------------- | |||
|
21 | ||||
|
22 | def lfconvert(ui, src, dest, *pats, **opts): | |||
|
23 | '''Convert a normal repository to a largefiles repository | |||
|
24 | ||||
|
25 | Convert source repository creating an identical repository, except that all | |||
|
26 | files that match the patterns given, or are over the given size will be | |||
|
27 | added as largefiles. The size used to determine whether or not to track a | |||
|
28 | file as a largefile is the size of the first version of the file. After | |||
|
29 | running this command you will need to make sure that largefiles is enabled | |||
|
30 | anywhere you intend to push the new repository.''' | |||
|
31 | ||||
|
32 | if opts['tonormal']: | |||
|
33 | tolfile = False | |||
|
34 | else: | |||
|
35 | tolfile = True | |||
|
36 | size = opts['size'] | |||
|
37 | if not size: | |||
|
38 | size = ui.config(lfutil.longname, 'size', default=None) | |||
|
39 | try: | |||
|
40 | size = int(size) | |||
|
41 | except ValueError: | |||
|
42 | raise util.Abort(_('largefiles.size must be integer, was %s\n') % \ | |||
|
43 | size) | |||
|
44 | except TypeError: | |||
|
45 | raise util.Abort(_('size must be specified')) | |||
|
46 | ||||
|
47 | try: | |||
|
48 | rsrc = hg.repository(ui, src) | |||
|
49 | if not rsrc.local(): | |||
|
50 | raise util.Abort(_('%s is not a local Mercurial repo') % src) | |||
|
51 | except error.RepoError, err: | |||
|
52 | ui.traceback() | |||
|
53 | raise util.Abort(err.args[0]) | |||
|
54 | if os.path.exists(dest): | |||
|
55 | if not os.path.isdir(dest): | |||
|
56 | raise util.Abort(_('destination %s already exists') % dest) | |||
|
57 | elif os.listdir(dest): | |||
|
58 | raise util.Abort(_('destination %s is not empty') % dest) | |||
|
59 | try: | |||
|
60 | ui.status(_('initializing destination %s\n') % dest) | |||
|
61 | rdst = hg.repository(ui, dest, create=True) | |||
|
62 | if not rdst.local(): | |||
|
63 | raise util.Abort(_('%s is not a local Mercurial repo') % dest) | |||
|
64 | except error.RepoError: | |||
|
65 | ui.traceback() | |||
|
66 | raise util.Abort(_('%s is not a repo') % dest) | |||
|
67 | ||||
|
68 | try: | |||
|
69 | # Lock destination to prevent modification while it is converted to. | |||
|
70 | # Don't need to lock src because we are just reading from its history | |||
|
71 | # which can't change. | |||
|
72 | dst_lock = rdst.lock() | |||
|
73 | ||||
|
74 | # Get a list of all changesets in the source. The easy way to do this | |||
|
75 | # is to simply walk the changelog, using changelog.nodesbewteen(). | |||
|
76 | # Take a look at mercurial/revlog.py:639 for more details. | |||
|
77 | # Use a generator instead of a list to decrease memory usage | |||
|
78 | ctxs = (rsrc[ctx] for ctx in rsrc.changelog.nodesbetween(None, | |||
|
79 | rsrc.heads())[0]) | |||
|
80 | revmap = {node.nullid: node.nullid} | |||
|
81 | if tolfile: | |||
|
82 | lfiles = set() | |||
|
83 | normalfiles = set() | |||
|
84 | if not pats: | |||
|
85 | pats = ui.config(lfutil.longname, 'patterns', default=()) | |||
|
86 | if pats: | |||
|
87 | pats = pats.split(' ') | |||
|
88 | if pats: | |||
|
89 | matcher = match_.match(rsrc.root, '', list(pats)) | |||
|
90 | else: | |||
|
91 | matcher = None | |||
|
92 | ||||
|
93 | lfiletohash = {} | |||
|
94 | for ctx in ctxs: | |||
|
95 | ui.progress(_('converting revisions'), ctx.rev(), | |||
|
96 | unit=_('revision'), total=rsrc['tip'].rev()) | |||
|
97 | _lfconvert_addchangeset(rsrc, rdst, ctx, revmap, | |||
|
98 | lfiles, normalfiles, matcher, size, lfiletohash) | |||
|
99 | ui.progress(_('converting revisions'), None) | |||
|
100 | ||||
|
101 | if os.path.exists(rdst.wjoin(lfutil.shortname)): | |||
|
102 | shutil.rmtree(rdst.wjoin(lfutil.shortname)) | |||
|
103 | ||||
|
104 | for f in lfiletohash.keys(): | |||
|
105 | if os.path.isfile(rdst.wjoin(f)): | |||
|
106 | os.unlink(rdst.wjoin(f)) | |||
|
107 | try: | |||
|
108 | os.removedirs(os.path.dirname(rdst.wjoin(f))) | |||
|
109 | except: | |||
|
110 | pass | |||
|
111 | ||||
|
112 | else: | |||
|
113 | for ctx in ctxs: | |||
|
114 | ui.progress(_('converting revisions'), ctx.rev(), | |||
|
115 | unit=_('revision'), total=rsrc['tip'].rev()) | |||
|
116 | _addchangeset(ui, rsrc, rdst, ctx, revmap) | |||
|
117 | ||||
|
118 | ui.progress(_('converting revisions'), None) | |||
|
119 | except: | |||
|
120 | # we failed, remove the new directory | |||
|
121 | shutil.rmtree(rdst.root) | |||
|
122 | raise | |||
|
123 | finally: | |||
|
124 | dst_lock.release() | |||
|
125 | ||||
|
126 | def _addchangeset(ui, rsrc, rdst, ctx, revmap): | |||
|
127 | # Convert src parents to dst parents | |||
|
128 | parents = [] | |||
|
129 | for p in ctx.parents(): | |||
|
130 | parents.append(revmap[p.node()]) | |||
|
131 | while len(parents) < 2: | |||
|
132 | parents.append(node.nullid) | |||
|
133 | ||||
|
134 | # Generate list of changed files | |||
|
135 | files = set(ctx.files()) | |||
|
136 | if node.nullid not in parents: | |||
|
137 | mc = ctx.manifest() | |||
|
138 | mp1 = ctx.parents()[0].manifest() | |||
|
139 | mp2 = ctx.parents()[1].manifest() | |||
|
140 | files |= (set(mp1) | set(mp2)) - set(mc) | |||
|
141 | for f in mc: | |||
|
142 | if mc[f] != mp1.get(f, None) or mc[f] != mp2.get(f, None): | |||
|
143 | files.add(f) | |||
|
144 | ||||
|
145 | def getfilectx(repo, memctx, f): | |||
|
146 | if lfutil.standin(f) in files: | |||
|
147 | # if the file isn't in the manifest then it was removed | |||
|
148 | # or renamed, raise IOError to indicate this | |||
|
149 | try: | |||
|
150 | fctx = ctx.filectx(lfutil.standin(f)) | |||
|
151 | except error.LookupError: | |||
|
152 | raise IOError() | |||
|
153 | renamed = fctx.renamed() | |||
|
154 | if renamed: | |||
|
155 | renamed = lfutil.splitstandin(renamed[0]) | |||
|
156 | ||||
|
157 | hash = fctx.data().strip() | |||
|
158 | path = lfutil.findfile(rsrc, hash) | |||
|
159 | ### TODO: What if the file is not cached? | |||
|
160 | data = '' | |||
|
161 | fd = None | |||
|
162 | try: | |||
|
163 | fd = open(path, 'rb') | |||
|
164 | data = fd.read() | |||
|
165 | finally: | |||
|
166 | if fd: fd.close() | |||
|
167 | return context.memfilectx(f, data, 'l' in fctx.flags(), | |||
|
168 | 'x' in fctx.flags(), renamed) | |||
|
169 | else: | |||
|
170 | try: | |||
|
171 | fctx = ctx.filectx(f) | |||
|
172 | except error.LookupError: | |||
|
173 | raise IOError() | |||
|
174 | renamed = fctx.renamed() | |||
|
175 | if renamed: | |||
|
176 | renamed = renamed[0] | |||
|
177 | data = fctx.data() | |||
|
178 | if f == '.hgtags': | |||
|
179 | newdata = [] | |||
|
180 | for line in data.splitlines(): | |||
|
181 | id, name = line.split(' ', 1) | |||
|
182 | newdata.append('%s %s\n' % (node.hex(revmap[node.bin(id)]), | |||
|
183 | name)) | |||
|
184 | data = ''.join(newdata) | |||
|
185 | return context.memfilectx(f, data, 'l' in fctx.flags(), | |||
|
186 | 'x' in fctx.flags(), renamed) | |||
|
187 | ||||
|
188 | dstfiles = [] | |||
|
189 | for file in files: | |||
|
190 | if lfutil.isstandin(file): | |||
|
191 | dstfiles.append(lfutil.splitstandin(file)) | |||
|
192 | else: | |||
|
193 | dstfiles.append(file) | |||
|
194 | # Commit | |||
|
195 | mctx = context.memctx(rdst, parents, ctx.description(), dstfiles, | |||
|
196 | getfilectx, ctx.user(), ctx.date(), ctx.extra()) | |||
|
197 | ret = rdst.commitctx(mctx) | |||
|
198 | rdst.dirstate.setparents(ret) | |||
|
199 | revmap[ctx.node()] = rdst.changelog.tip() | |||
|
200 | ||||
|
201 | def _lfconvert_addchangeset(rsrc, rdst, ctx, revmap, lfiles, normalfiles, | |||
|
202 | matcher, size, lfiletohash): | |||
|
203 | # Convert src parents to dst parents | |||
|
204 | parents = [] | |||
|
205 | for p in ctx.parents(): | |||
|
206 | parents.append(revmap[p.node()]) | |||
|
207 | while len(parents) < 2: | |||
|
208 | parents.append(node.nullid) | |||
|
209 | ||||
|
210 | # Generate list of changed files | |||
|
211 | files = set(ctx.files()) | |||
|
212 | if node.nullid not in parents: | |||
|
213 | mc = ctx.manifest() | |||
|
214 | mp1 = ctx.parents()[0].manifest() | |||
|
215 | mp2 = ctx.parents()[1].manifest() | |||
|
216 | files |= (set(mp1) | set(mp2)) - set(mc) | |||
|
217 | for f in mc: | |||
|
218 | if mc[f] != mp1.get(f, None) or mc[f] != mp2.get(f, None): | |||
|
219 | files.add(f) | |||
|
220 | ||||
|
221 | dstfiles = [] | |||
|
222 | for f in files: | |||
|
223 | if f not in lfiles and f not in normalfiles: | |||
|
224 | islfile = _islfile(f, ctx, matcher, size) | |||
|
225 | # If this file was renamed or copied then copy | |||
|
226 | # the lfileness of its predecessor | |||
|
227 | if f in ctx.manifest(): | |||
|
228 | fctx = ctx.filectx(f) | |||
|
229 | renamed = fctx.renamed() | |||
|
230 | renamedlfile = renamed and renamed[0] in lfiles | |||
|
231 | islfile |= renamedlfile | |||
|
232 | if 'l' in fctx.flags(): | |||
|
233 | if renamedlfile: | |||
|
234 | raise util.Abort( | |||
|
235 | _('Renamed/copied largefile %s becomes symlink') % f) | |||
|
236 | islfile = False | |||
|
237 | if islfile: | |||
|
238 | lfiles.add(f) | |||
|
239 | else: | |||
|
240 | normalfiles.add(f) | |||
|
241 | ||||
|
242 | if f in lfiles: | |||
|
243 | dstfiles.append(lfutil.standin(f)) | |||
|
244 | # lfile in manifest if it has not been removed/renamed | |||
|
245 | if f in ctx.manifest(): | |||
|
246 | if 'l' in ctx.filectx(f).flags(): | |||
|
247 | if renamed and renamed[0] in lfiles: | |||
|
248 | raise util.Abort(_('largefile %s becomes symlink') % f) | |||
|
249 | ||||
|
250 | # lfile was modified, update standins | |||
|
251 | fullpath = rdst.wjoin(f) | |||
|
252 | lfutil.createdir(os.path.dirname(fullpath)) | |||
|
253 | m = util.sha1('') | |||
|
254 | m.update(ctx[f].data()) | |||
|
255 | hash = m.hexdigest() | |||
|
256 | if f not in lfiletohash or lfiletohash[f] != hash: | |||
|
257 | try: | |||
|
258 | fd = open(fullpath, 'wb') | |||
|
259 | fd.write(ctx[f].data()) | |||
|
260 | finally: | |||
|
261 | if fd: | |||
|
262 | fd.close() | |||
|
263 | executable = 'x' in ctx[f].flags() | |||
|
264 | os.chmod(fullpath, lfutil.getmode(executable)) | |||
|
265 | lfutil.writestandin(rdst, lfutil.standin(f), hash, | |||
|
266 | executable) | |||
|
267 | lfiletohash[f] = hash | |||
|
268 | else: | |||
|
269 | # normal file | |||
|
270 | dstfiles.append(f) | |||
|
271 | ||||
|
272 | def getfilectx(repo, memctx, f): | |||
|
273 | if lfutil.isstandin(f): | |||
|
274 | # if the file isn't in the manifest then it was removed | |||
|
275 | # or renamed, raise IOError to indicate this | |||
|
276 | srcfname = lfutil.splitstandin(f) | |||
|
277 | try: | |||
|
278 | fctx = ctx.filectx(srcfname) | |||
|
279 | except error.LookupError: | |||
|
280 | raise IOError() | |||
|
281 | renamed = fctx.renamed() | |||
|
282 | if renamed: | |||
|
283 | # standin is always a lfile because lfileness | |||
|
284 | # doesn't change after rename or copy | |||
|
285 | renamed = lfutil.standin(renamed[0]) | |||
|
286 | ||||
|
287 | return context.memfilectx(f, lfiletohash[srcfname], 'l' in | |||
|
288 | fctx.flags(), 'x' in fctx.flags(), renamed) | |||
|
289 | else: | |||
|
290 | try: | |||
|
291 | fctx = ctx.filectx(f) | |||
|
292 | except error.LookupError: | |||
|
293 | raise IOError() | |||
|
294 | renamed = fctx.renamed() | |||
|
295 | if renamed: | |||
|
296 | renamed = renamed[0] | |||
|
297 | ||||
|
298 | data = fctx.data() | |||
|
299 | if f == '.hgtags': | |||
|
300 | newdata = [] | |||
|
301 | for line in data.splitlines(): | |||
|
302 | id, name = line.split(' ', 1) | |||
|
303 | newdata.append('%s %s\n' % (node.hex(revmap[node.bin(id)]), | |||
|
304 | name)) | |||
|
305 | data = ''.join(newdata) | |||
|
306 | return context.memfilectx(f, data, 'l' in fctx.flags(), | |||
|
307 | 'x' in fctx.flags(), renamed) | |||
|
308 | ||||
|
309 | # Commit | |||
|
310 | mctx = context.memctx(rdst, parents, ctx.description(), dstfiles, | |||
|
311 | getfilectx, ctx.user(), ctx.date(), ctx.extra()) | |||
|
312 | ret = rdst.commitctx(mctx) | |||
|
313 | rdst.dirstate.setparents(ret) | |||
|
314 | revmap[ctx.node()] = rdst.changelog.tip() | |||
|
315 | ||||
|
316 | def _islfile(file, ctx, matcher, size): | |||
|
317 | ''' | |||
|
318 | A file is a lfile if it matches a pattern or is over | |||
|
319 | the given size. | |||
|
320 | ''' | |||
|
321 | # Never store hgtags or hgignore as lfiles | |||
|
322 | if file == '.hgtags' or file == '.hgignore' or file == '.hgsigs': | |||
|
323 | return False | |||
|
324 | if matcher and matcher(file): | |||
|
325 | return True | |||
|
326 | try: | |||
|
327 | return ctx.filectx(file).size() >= size * 1024 * 1024 | |||
|
328 | except error.LookupError: | |||
|
329 | return False | |||
|
330 | ||||
|
331 | def uploadlfiles(ui, rsrc, rdst, files): | |||
|
332 | '''upload largefiles to the central store''' | |||
|
333 | ||||
|
334 | # Don't upload locally. All largefiles are in the system wide cache | |||
|
335 | # so the other repo can just get them from there. | |||
|
336 | if not files or rdst.local(): | |||
|
337 | return | |||
|
338 | ||||
|
339 | store = basestore._openstore(rsrc, rdst, put=True) | |||
|
340 | ||||
|
341 | at = 0 | |||
|
342 | files = filter(lambda h: not store.exists(h), files) | |||
|
343 | for hash in files: | |||
|
344 | ui.progress(_('uploading largefiles'), at, unit='largefile', total=len(files)) | |||
|
345 | source = lfutil.findfile(rsrc, hash) | |||
|
346 | if not source: | |||
|
347 | raise util.Abort(_('Missing largefile %s needs to be uploaded') % hash) | |||
|
348 | # XXX check for errors here | |||
|
349 | store.put(source, hash) | |||
|
350 | at += 1 | |||
|
351 | ui.progress('uploading largefiles', None) | |||
|
352 | ||||
|
353 | def verifylfiles(ui, repo, all=False, contents=False): | |||
|
354 | '''Verify that every big file revision in the current changeset | |||
|
355 | exists in the central store. With --contents, also verify that | |||
|
356 | the contents of each big file revision are correct (SHA-1 hash | |||
|
357 | matches the revision ID). With --all, check every changeset in | |||
|
358 | this repository.''' | |||
|
359 | if all: | |||
|
360 | # Pass a list to the function rather than an iterator because we know a | |||
|
361 | # list will work. | |||
|
362 | revs = range(len(repo)) | |||
|
363 | else: | |||
|
364 | revs = ['.'] | |||
|
365 | ||||
|
366 | store = basestore._openstore(repo) | |||
|
367 | return store.verify(revs, contents=contents) | |||
|
368 | ||||
|
369 | def cachelfiles(ui, repo, node): | |||
|
370 | '''cachelfiles ensures that all largefiles needed by the specified revision | |||
|
371 | are present in the repository's largefile cache. | |||
|
372 | ||||
|
373 | returns a tuple (cached, missing). cached is the list of files downloaded | |||
|
374 | by this operation; missing is the list of files that were needed but could | |||
|
375 | not be found.''' | |||
|
376 | lfiles = lfutil.listlfiles(repo, node) | |||
|
377 | toget = [] | |||
|
378 | ||||
|
379 | for lfile in lfiles: | |||
|
380 | expectedhash = repo[node][lfutil.standin(lfile)].data().strip() | |||
|
381 | # if it exists and its hash matches, it might have been locally | |||
|
382 | # modified before updating and the user chose 'local'. in this case, | |||
|
383 | # it will not be in any store, so don't look for it. | |||
|
384 | if (not os.path.exists(repo.wjoin(lfile)) \ | |||
|
385 | or expectedhash != lfutil.hashfile(repo.wjoin(lfile))) and \ | |||
|
386 | not lfutil.findfile(repo, expectedhash): | |||
|
387 | toget.append((lfile, expectedhash)) | |||
|
388 | ||||
|
389 | if toget: | |||
|
390 | store = basestore._openstore(repo) | |||
|
391 | ret = store.get(toget) | |||
|
392 | return ret | |||
|
393 | ||||
|
394 | return ([], []) | |||
|
395 | ||||
|
396 | def updatelfiles(ui, repo, filelist=None, printmessage=True): | |||
|
397 | wlock = repo.wlock() | |||
|
398 | try: | |||
|
399 | lfdirstate = lfutil.openlfdirstate(ui, repo) | |||
|
400 | lfiles = set(lfutil.listlfiles(repo)) | set(lfdirstate) | |||
|
401 | ||||
|
402 | if filelist is not None: | |||
|
403 | lfiles = [f for f in lfiles if f in filelist] | |||
|
404 | ||||
|
405 | printed = False | |||
|
406 | if printmessage and lfiles: | |||
|
407 | ui.status(_('getting changed largefiles\n')) | |||
|
408 | printed = True | |||
|
409 | cachelfiles(ui, repo, '.') | |||
|
410 | ||||
|
411 | updated, removed = 0, 0 | |||
|
412 | for i in map(lambda f: _updatelfile(repo, lfdirstate, f), lfiles): | |||
|
413 | # increment the appropriate counter according to _updatelfile's | |||
|
414 | # return value | |||
|
415 | updated += i > 0 and i or 0 | |||
|
416 | removed -= i < 0 and i or 0 | |||
|
417 | if printmessage and (removed or updated) and not printed: | |||
|
418 | ui.status(_('getting changed largefiles\n')) | |||
|
419 | printed = True | |||
|
420 | ||||
|
421 | lfdirstate.write() | |||
|
422 | if printed and printmessage: | |||
|
423 | ui.status(_('%d largefiles updated, %d removed\n') % (updated, | |||
|
424 | removed)) | |||
|
425 | finally: | |||
|
426 | wlock.release() | |||
|
427 | ||||
|
428 | def _updatelfile(repo, lfdirstate, lfile): | |||
|
429 | '''updates a single largefile and copies the state of its standin from | |||
|
430 | the repository's dirstate to its state in the lfdirstate. | |||
|
431 | ||||
|
432 | returns 1 if the file was modified, -1 if the file was removed, 0 if the | |||
|
433 | file was unchanged, and None if the needed largefile was missing from the | |||
|
434 | cache.''' | |||
|
435 | ret = 0 | |||
|
436 | abslfile = repo.wjoin(lfile) | |||
|
437 | absstandin = repo.wjoin(lfutil.standin(lfile)) | |||
|
438 | if os.path.exists(absstandin): | |||
|
439 | if os.path.exists(absstandin+'.orig'): | |||
|
440 | shutil.copyfile(abslfile, abslfile+'.orig') | |||
|
441 | expecthash = lfutil.readstandin(repo, lfile) | |||
|
442 | if expecthash != '' and \ | |||
|
443 | (not os.path.exists(abslfile) or \ | |||
|
444 | expecthash != lfutil.hashfile(abslfile)): | |||
|
445 | if not lfutil.copyfromcache(repo, expecthash, lfile): | |||
|
446 | return None # don't try to set the mode or update the dirstate | |||
|
447 | ret = 1 | |||
|
448 | mode = os.stat(absstandin).st_mode | |||
|
449 | if mode != os.stat(abslfile).st_mode: | |||
|
450 | os.chmod(abslfile, mode) | |||
|
451 | ret = 1 | |||
|
452 | else: | |||
|
453 | if os.path.exists(abslfile): | |||
|
454 | os.unlink(abslfile) | |||
|
455 | ret = -1 | |||
|
456 | state = repo.dirstate[lfutil.standin(lfile)] | |||
|
457 | if state == 'n': | |||
|
458 | lfdirstate.normal(lfile) | |||
|
459 | elif state == 'r': | |||
|
460 | lfdirstate.remove(lfile) | |||
|
461 | elif state == 'a': | |||
|
462 | lfdirstate.add(lfile) | |||
|
463 | elif state == '?': | |||
|
464 | try: | |||
|
465 | # Mercurial >= 1.9 | |||
|
466 | lfdirstate.drop(lfile) | |||
|
467 | except AttributeError: | |||
|
468 | # Mercurial <= 1.8 | |||
|
469 | lfdirstate.forget(lfile) | |||
|
470 | return ret | |||
|
471 | ||||
|
472 | # -- hg commands declarations ------------------------------------------------ | |||
|
473 | ||||
|
474 | ||||
|
475 | cmdtable = { | |||
|
476 | 'lfconvert': (lfconvert, | |||
|
477 | [('s', 'size', 0, 'All files over this size (in megabytes) ' | |||
|
478 | 'will be considered largefiles. This can also be specified in ' | |||
|
479 | 'your hgrc as [largefiles].size.'), | |||
|
480 | ('','tonormal',False, | |||
|
481 | 'Convert from a largefiles repo to a normal repo')], | |||
|
482 | _('hg lfconvert SOURCE DEST [FILE ...]')), | |||
|
483 | } |
This diff has been collapsed as it changes many lines, (502 lines changed) Show them Hide them | |||||
@@ -0,0 +1,502 b'' | |||||
|
1 | # Copyright 2009-2010 Gregory P. Ward | |||
|
2 | # Copyright 2009-2010 Intelerad Medical Systems Incorporated | |||
|
3 | # Copyright 2010-2011 Fog Creek Software | |||
|
4 | # Copyright 2010-2011 Unity Technologies | |||
|
5 | # | |||
|
6 | # This software may be used and distributed according to the terms of the | |||
|
7 | # GNU General Public License version 2 or any later version. | |||
|
8 | ||||
|
9 | '''largefiles utility code: must not import other modules in this package.''' | |||
|
10 | ||||
|
11 | import os | |||
|
12 | import errno | |||
|
13 | import inspect | |||
|
14 | import shutil | |||
|
15 | import stat | |||
|
16 | import hashlib | |||
|
17 | ||||
|
18 | from mercurial import cmdutil, dirstate, httpconnection, match as match_, \ | |||
|
19 | url as url_, util | |||
|
20 | from mercurial.i18n import _ | |||
|
21 | ||||
|
22 | try: | |||
|
23 | from mercurial import scmutil | |||
|
24 | except ImportError: | |||
|
25 | pass | |||
|
26 | ||||
|
27 | shortname = '.hglf' | |||
|
28 | longname = 'largefiles' | |||
|
29 | ||||
|
30 | ||||
|
31 | # -- Portability wrappers ---------------------------------------------- | |||
|
32 | ||||
|
33 | if 'subrepos' in inspect.getargspec(dirstate.dirstate.status)[0]: | |||
|
34 | # for Mercurial >= 1.5 | |||
|
35 | def dirstate_walk(dirstate, matcher, unknown=False, ignored=False): | |||
|
36 | return dirstate.walk(matcher, [], unknown, ignored) | |||
|
37 | else: | |||
|
38 | # for Mercurial <= 1.4 | |||
|
39 | def dirstate_walk(dirstate, matcher, unknown=False, ignored=False): | |||
|
40 | return dirstate.walk(matcher, unknown, ignored) | |||
|
41 | ||||
|
42 | def repo_add(repo, list): | |||
|
43 | try: | |||
|
44 | # Mercurial <= 1.5 | |||
|
45 | add = repo.add | |||
|
46 | except AttributeError: | |||
|
47 | # Mercurial >= 1.6 | |||
|
48 | add = repo[None].add | |||
|
49 | return add(list) | |||
|
50 | ||||
|
51 | def repo_remove(repo, list, unlink=False): | |||
|
52 | try: | |||
|
53 | # Mercurial <= 1.5 | |||
|
54 | remove = repo.remove | |||
|
55 | except AttributeError: | |||
|
56 | # Mercurial >= 1.6 | |||
|
57 | try: | |||
|
58 | # Mercurial <= 1.8 | |||
|
59 | remove = repo[None].remove | |||
|
60 | except AttributeError: | |||
|
61 | # Mercurial >= 1.9 | |||
|
62 | def remove(list, unlink): | |||
|
63 | wlock = repo.wlock() | |||
|
64 | try: | |||
|
65 | if unlink: | |||
|
66 | for f in list: | |||
|
67 | try: | |||
|
68 | util.unlinkpath(repo.wjoin(f)) | |||
|
69 | except OSError, inst: | |||
|
70 | if inst.errno != errno.ENOENT: | |||
|
71 | raise | |||
|
72 | repo[None].forget(list) | |||
|
73 | finally: | |||
|
74 | wlock.release() | |||
|
75 | ||||
|
76 | return remove(list, unlink=unlink) | |||
|
77 | ||||
|
78 | def repo_forget(repo, list): | |||
|
79 | try: | |||
|
80 | # Mercurial <= 1.5 | |||
|
81 | forget = repo.forget | |||
|
82 | except AttributeError: | |||
|
83 | # Mercurial >= 1.6 | |||
|
84 | forget = repo[None].forget | |||
|
85 | return forget(list) | |||
|
86 | ||||
|
87 | def findoutgoing(repo, remote, force): | |||
|
88 | # First attempt is for Mercurial <= 1.5 second is for >= 1.6 | |||
|
89 | try: | |||
|
90 | return repo.findoutgoing(remote) | |||
|
91 | except AttributeError: | |||
|
92 | from mercurial import discovery | |||
|
93 | try: | |||
|
94 | # Mercurial <= 1.8 | |||
|
95 | return discovery.findoutgoing(repo, remote, force=force) | |||
|
96 | except AttributeError: | |||
|
97 | # Mercurial >= 1.9 | |||
|
98 | common, _anyinc, _heads = discovery.findcommonincoming(repo, | |||
|
99 | remote, force=force) | |||
|
100 | return repo.changelog.findmissing(common) | |||
|
101 | ||||
|
102 | # -- Private worker functions ------------------------------------------ | |||
|
103 | ||||
|
104 | if os.name == 'nt': | |||
|
105 | from mercurial import win32 | |||
|
106 | try: | |||
|
107 | linkfn = win32.oslink | |||
|
108 | except: | |||
|
109 | linkfn = win32.os_link | |||
|
110 | else: | |||
|
111 | linkfn = os.link | |||
|
112 | ||||
|
113 | def link(src, dest): | |||
|
114 | try: | |||
|
115 | linkfn(src, dest) | |||
|
116 | except OSError: | |||
|
117 | # If hardlinks fail fall back on copy | |||
|
118 | shutil.copyfile(src, dest) | |||
|
119 | os.chmod(dest, os.stat(src).st_mode) | |||
|
120 | ||||
|
121 | def systemcachepath(ui, hash): | |||
|
122 | path = ui.config(longname, 'systemcache', None) | |||
|
123 | if path: | |||
|
124 | path = os.path.join(path, hash) | |||
|
125 | else: | |||
|
126 | if os.name == 'nt': | |||
|
127 | path = os.path.join(os.getenv('LOCALAPPDATA') or \ | |||
|
128 | os.getenv('APPDATA'), longname, hash) | |||
|
129 | elif os.name == 'posix': | |||
|
130 | path = os.path.join(os.getenv('HOME'), '.' + longname, hash) | |||
|
131 | else: | |||
|
132 | raise util.Abort(_('Unknown operating system: %s\n') % os.name) | |||
|
133 | return path | |||
|
134 | ||||
|
135 | def insystemcache(ui, hash): | |||
|
136 | return os.path.exists(systemcachepath(ui, hash)) | |||
|
137 | ||||
|
138 | def findfile(repo, hash): | |||
|
139 | if incache(repo, hash): | |||
|
140 | repo.ui.note(_('Found %s in cache\n') % hash) | |||
|
141 | return cachepath(repo, hash) | |||
|
142 | if insystemcache(repo.ui, hash): | |||
|
143 | repo.ui.note(_('Found %s in system cache\n') % hash) | |||
|
144 | return systemcachepath(repo.ui, hash) | |||
|
145 | return None | |||
|
146 | ||||
|
147 | class largefiles_dirstate(dirstate.dirstate): | |||
|
148 | def __getitem__(self, key): | |||
|
149 | return super(largefiles_dirstate, self).__getitem__(unixpath(key)) | |||
|
150 | def normal(self, f): | |||
|
151 | return super(largefiles_dirstate, self).normal(unixpath(f)) | |||
|
152 | def remove(self, f): | |||
|
153 | return super(largefiles_dirstate, self).remove(unixpath(f)) | |||
|
154 | def add(self, f): | |||
|
155 | return super(largefiles_dirstate, self).add(unixpath(f)) | |||
|
156 | def drop(self, f): | |||
|
157 | return super(largefiles_dirstate, self).drop(unixpath(f)) | |||
|
158 | def forget(self, f): | |||
|
159 | return super(largefiles_dirstate, self).forget(unixpath(f)) | |||
|
160 | ||||
|
161 | def openlfdirstate(ui, repo): | |||
|
162 | ''' | |||
|
163 | Return a dirstate object that tracks big files: i.e. its root is the | |||
|
164 | repo root, but it is saved in .hg/largefiles/dirstate. | |||
|
165 | ''' | |||
|
166 | admin = repo.join(longname) | |||
|
167 | try: | |||
|
168 | # Mercurial >= 1.9 | |||
|
169 | opener = scmutil.opener(admin) | |||
|
170 | except ImportError: | |||
|
171 | # Mercurial <= 1.8 | |||
|
172 | opener = util.opener(admin) | |||
|
173 | if hasattr(repo.dirstate, '_validate'): | |||
|
174 | lfdirstate = largefiles_dirstate(opener, ui, repo.root, | |||
|
175 | repo.dirstate._validate) | |||
|
176 | else: | |||
|
177 | lfdirstate = largefiles_dirstate(opener, ui, repo.root) | |||
|
178 | ||||
|
179 | # If the largefiles dirstate does not exist, populate and create it. This | |||
|
180 | # ensures that we create it on the first meaningful largefiles operation in | |||
|
181 | # a new clone. It also gives us an easy way to forcibly rebuild largefiles | |||
|
182 | # state: | |||
|
183 | # rm .hg/largefiles/dirstate && hg status | |||
|
184 | # Or even, if things are really messed up: | |||
|
185 | # rm -rf .hg/largefiles && hg status | |||
|
186 | if not os.path.exists(os.path.join(admin, 'dirstate')): | |||
|
187 | util.makedirs(admin) | |||
|
188 | matcher = getstandinmatcher(repo) | |||
|
189 | for standin in dirstate_walk(repo.dirstate, matcher): | |||
|
190 | lfile = splitstandin(standin) | |||
|
191 | hash = readstandin(repo, lfile) | |||
|
192 | lfdirstate.normallookup(lfile) | |||
|
193 | try: | |||
|
194 | if hash == hashfile(lfile): | |||
|
195 | lfdirstate.normal(lfile) | |||
|
196 | except IOError, err: | |||
|
197 | if err.errno != errno.ENOENT: | |||
|
198 | raise | |||
|
199 | ||||
|
200 | lfdirstate.write() | |||
|
201 | ||||
|
202 | return lfdirstate | |||
|
203 | ||||
|
204 | def lfdirstate_status(lfdirstate, repo, rev): | |||
|
205 | wlock = repo.wlock() | |||
|
206 | try: | |||
|
207 | match = match_.always(repo.root, repo.getcwd()) | |||
|
208 | s = lfdirstate.status(match, [], False, False, False) | |||
|
209 | unsure, modified, added, removed, missing, unknown, ignored, clean = s | |||
|
210 | for lfile in unsure: | |||
|
211 | if repo[rev][standin(lfile)].data().strip() != \ | |||
|
212 | hashfile(repo.wjoin(lfile)): | |||
|
213 | modified.append(lfile) | |||
|
214 | else: | |||
|
215 | clean.append(lfile) | |||
|
216 | lfdirstate.normal(lfile) | |||
|
217 | lfdirstate.write() | |||
|
218 | finally: | |||
|
219 | wlock.release() | |||
|
220 | return (modified, added, removed, missing, unknown, ignored, clean) | |||
|
221 | ||||
|
222 | def listlfiles(repo, rev=None, matcher=None): | |||
|
223 | '''list largefiles in the working copy or specified changeset''' | |||
|
224 | ||||
|
225 | if matcher is None: | |||
|
226 | matcher = getstandinmatcher(repo) | |||
|
227 | ||||
|
228 | # ignore unknown files in working directory | |||
|
229 | return [splitstandin(f) for f in repo[rev].walk(matcher) \ | |||
|
230 | if rev is not None or repo.dirstate[f] != '?'] | |||
|
231 | ||||
|
232 | def incache(repo, hash): | |||
|
233 | return os.path.exists(cachepath(repo, hash)) | |||
|
234 | ||||
|
235 | def createdir(dir): | |||
|
236 | if not os.path.exists(dir): | |||
|
237 | os.makedirs(dir) | |||
|
238 | ||||
|
239 | def cachepath(repo, hash): | |||
|
240 | return repo.join(os.path.join(longname, hash)) | |||
|
241 | ||||
|
242 | def copyfromcache(repo, hash, filename): | |||
|
243 | '''copyfromcache copies the specified largefile from the repo or system | |||
|
244 | cache to the specified location in the repository. It will not throw an | |||
|
245 | exception on failure, as it is meant to be called only after ensuring that | |||
|
246 | the needed largefile exists in the cache.''' | |||
|
247 | path = findfile(repo, hash) | |||
|
248 | if path is None: | |||
|
249 | return False | |||
|
250 | util.makedirs(os.path.dirname(repo.wjoin(filename))) | |||
|
251 | shutil.copy(path, repo.wjoin(filename)) | |||
|
252 | return True | |||
|
253 | ||||
|
254 | def copytocache(repo, rev, file, uploaded=False): | |||
|
255 | hash = readstandin(repo, file) | |||
|
256 | if incache(repo, hash): | |||
|
257 | return | |||
|
258 | copytocacheabsolute(repo, repo.wjoin(file), hash) | |||
|
259 | ||||
|
260 | def copytocacheabsolute(repo, file, hash): | |||
|
261 | createdir(os.path.dirname(cachepath(repo, hash))) | |||
|
262 | if insystemcache(repo.ui, hash): | |||
|
263 | link(systemcachepath(repo.ui, hash), cachepath(repo, hash)) | |||
|
264 | else: | |||
|
265 | shutil.copyfile(file, cachepath(repo, hash)) | |||
|
266 | os.chmod(cachepath(repo, hash), os.stat(file).st_mode) | |||
|
267 | linktosystemcache(repo, hash) | |||
|
268 | ||||
|
269 | def linktosystemcache(repo, hash): | |||
|
270 | createdir(os.path.dirname(systemcachepath(repo.ui, hash))) | |||
|
271 | link(cachepath(repo, hash), systemcachepath(repo.ui, hash)) | |||
|
272 | ||||
|
273 | def getstandinmatcher(repo, pats=[], opts={}): | |||
|
274 | '''Return a match object that applies pats to the standin directory''' | |||
|
275 | standindir = repo.pathto(shortname) | |||
|
276 | if pats: | |||
|
277 | # patterns supplied: search standin directory relative to current dir | |||
|
278 | cwd = repo.getcwd() | |||
|
279 | if os.path.isabs(cwd): | |||
|
280 | # cwd is an absolute path for hg -R <reponame> | |||
|
281 | # work relative to the repository root in this case | |||
|
282 | cwd = '' | |||
|
283 | pats = [os.path.join(standindir, cwd, pat) for pat in pats] | |||
|
284 | elif os.path.isdir(standindir): | |||
|
285 | # no patterns: relative to repo root | |||
|
286 | pats = [standindir] | |||
|
287 | else: | |||
|
288 | # no patterns and no standin dir: return matcher that matches nothing | |||
|
289 | match = match_.match(repo.root, None, [], exact=True) | |||
|
290 | match.matchfn = lambda f: False | |||
|
291 | return match | |||
|
292 | return getmatcher(repo, pats, opts, showbad=False) | |||
|
293 | ||||
|
294 | def getmatcher(repo, pats=[], opts={}, showbad=True): | |||
|
295 | '''Wrapper around scmutil.match() that adds showbad: if false, neuter | |||
|
296 | the match object\'s bad() method so it does not print any warnings | |||
|
297 | about missing files or directories.''' | |||
|
298 | try: | |||
|
299 | # Mercurial >= 1.9 | |||
|
300 | match = scmutil.match(repo[None], pats, opts) | |||
|
301 | except ImportError: | |||
|
302 | # Mercurial <= 1.8 | |||
|
303 | match = cmdutil.match(repo, pats, opts) | |||
|
304 | ||||
|
305 | if not showbad: | |||
|
306 | match.bad = lambda f, msg: None | |||
|
307 | return match | |||
|
308 | ||||
|
309 | def composestandinmatcher(repo, rmatcher): | |||
|
310 | '''Return a matcher that accepts standins corresponding to the files | |||
|
311 | accepted by rmatcher. Pass the list of files in the matcher as the | |||
|
312 | paths specified by the user.''' | |||
|
313 | smatcher = getstandinmatcher(repo, rmatcher.files()) | |||
|
314 | isstandin = smatcher.matchfn | |||
|
315 | def composed_matchfn(f): | |||
|
316 | return isstandin(f) and rmatcher.matchfn(splitstandin(f)) | |||
|
317 | smatcher.matchfn = composed_matchfn | |||
|
318 | ||||
|
319 | return smatcher | |||
|
320 | ||||
|
321 | def standin(filename): | |||
|
322 | '''Return the repo-relative path to the standin for the specified big | |||
|
323 | file.''' | |||
|
324 | # Notes: | |||
|
325 | # 1) Most callers want an absolute path, but _create_standin() needs | |||
|
326 | # it repo-relative so lfadd() can pass it to repo_add(). So leave | |||
|
327 | # it up to the caller to use repo.wjoin() to get an absolute path. | |||
|
328 | # 2) Join with '/' because that's what dirstate always uses, even on | |||
|
329 | # Windows. Change existing separator to '/' first in case we are | |||
|
330 | # passed filenames from an external source (like the command line). | |||
|
331 | return shortname + '/' + filename.replace(os.sep, '/') | |||
|
332 | ||||
|
333 | def isstandin(filename): | |||
|
334 | '''Return true if filename is a big file standin. filename must | |||
|
335 | be in Mercurial\'s internal form (slash-separated).''' | |||
|
336 | return filename.startswith(shortname + '/') | |||
|
337 | ||||
|
338 | def splitstandin(filename): | |||
|
339 | # Split on / because that's what dirstate always uses, even on Windows. | |||
|
340 | # Change local separator to / first just in case we are passed filenames | |||
|
341 | # from an external source (like the command line). | |||
|
342 | bits = filename.replace(os.sep, '/').split('/', 1) | |||
|
343 | if len(bits) == 2 and bits[0] == shortname: | |||
|
344 | return bits[1] | |||
|
345 | else: | |||
|
346 | return None | |||
|
347 | ||||
|
348 | def updatestandin(repo, standin): | |||
|
349 | file = repo.wjoin(splitstandin(standin)) | |||
|
350 | if os.path.exists(file): | |||
|
351 | hash = hashfile(file) | |||
|
352 | executable = getexecutable(file) | |||
|
353 | writestandin(repo, standin, hash, executable) | |||
|
354 | ||||
|
355 | def readstandin(repo, filename, node=None): | |||
|
356 | '''read hex hash from standin for filename at given node, or working | |||
|
357 | directory if no node is given''' | |||
|
358 | return repo[node][standin(filename)].data().strip() | |||
|
359 | ||||
|
360 | def writestandin(repo, standin, hash, executable): | |||
|
361 | '''write hhash to <repo.root>/<standin>''' | |||
|
362 | writehash(hash, repo.wjoin(standin), executable) | |||
|
363 | ||||
|
364 | def copyandhash(instream, outfile): | |||
|
365 | '''Read bytes from instream (iterable) and write them to outfile, | |||
|
366 | computing the SHA-1 hash of the data along the way. Close outfile | |||
|
367 | when done and return the binary hash.''' | |||
|
368 | hasher = util.sha1('') | |||
|
369 | for data in instream: | |||
|
370 | hasher.update(data) | |||
|
371 | outfile.write(data) | |||
|
372 | ||||
|
373 | # Blecch: closing a file that somebody else opened is rude and | |||
|
374 | # wrong. But it's so darn convenient and practical! After all, | |||
|
375 | # outfile was opened just to copy and hash. | |||
|
376 | outfile.close() | |||
|
377 | ||||
|
378 | return hasher.digest() | |||
|
379 | ||||
|
380 | def hashrepofile(repo, file): | |||
|
381 | return hashfile(repo.wjoin(file)) | |||
|
382 | ||||
|
383 | def hashfile(file): | |||
|
384 | if not os.path.exists(file): | |||
|
385 | return '' | |||
|
386 | hasher = util.sha1('') | |||
|
387 | fd = open(file, 'rb') | |||
|
388 | for data in blockstream(fd): | |||
|
389 | hasher.update(data) | |||
|
390 | fd.close() | |||
|
391 | return hasher.hexdigest() | |||
|
392 | ||||
|
393 | class limitreader(object): | |||
|
394 | def __init__(self, f, limit): | |||
|
395 | self.f = f | |||
|
396 | self.limit = limit | |||
|
397 | ||||
|
398 | def read(self, length): | |||
|
399 | if self.limit == 0: | |||
|
400 | return '' | |||
|
401 | length = length > self.limit and self.limit or length | |||
|
402 | self.limit -= length | |||
|
403 | return self.f.read(length) | |||
|
404 | ||||
|
405 | def close(self): | |||
|
406 | pass | |||
|
407 | ||||
|
408 | def blockstream(infile, blocksize=128 * 1024): | |||
|
409 | """Generator that yields blocks of data from infile and closes infile.""" | |||
|
410 | while True: | |||
|
411 | data = infile.read(blocksize) | |||
|
412 | if not data: | |||
|
413 | break | |||
|
414 | yield data | |||
|
415 | # Same blecch as above. | |||
|
416 | infile.close() | |||
|
417 | ||||
|
418 | def readhash(filename): | |||
|
419 | rfile = open(filename, 'rb') | |||
|
420 | hash = rfile.read(40) | |||
|
421 | rfile.close() | |||
|
422 | if len(hash) < 40: | |||
|
423 | raise util.Abort(_('bad hash in \'%s\' (only %d bytes long)') | |||
|
424 | % (filename, len(hash))) | |||
|
425 | return hash | |||
|
426 | ||||
|
427 | def writehash(hash, filename, executable): | |||
|
428 | util.makedirs(os.path.dirname(filename)) | |||
|
429 | if os.path.exists(filename): | |||
|
430 | os.unlink(filename) | |||
|
431 | wfile = open(filename, 'wb') | |||
|
432 | ||||
|
433 | try: | |||
|
434 | wfile.write(hash) | |||
|
435 | wfile.write('\n') | |||
|
436 | finally: | |||
|
437 | wfile.close() | |||
|
438 | if os.path.exists(filename): | |||
|
439 | os.chmod(filename, getmode(executable)) | |||
|
440 | ||||
|
441 | def getexecutable(filename): | |||
|
442 | mode = os.stat(filename).st_mode | |||
|
443 | return (mode & stat.S_IXUSR) and (mode & stat.S_IXGRP) and (mode & \ | |||
|
444 | stat.S_IXOTH) | |||
|
445 | ||||
|
446 | def getmode(executable): | |||
|
447 | if executable: | |||
|
448 | return 0755 | |||
|
449 | else: | |||
|
450 | return 0644 | |||
|
451 | ||||
|
452 | def urljoin(first, second, *arg): | |||
|
453 | def join(left, right): | |||
|
454 | if not left.endswith('/'): | |||
|
455 | left += '/' | |||
|
456 | if right.startswith('/'): | |||
|
457 | right = right[1:] | |||
|
458 | return left + right | |||
|
459 | ||||
|
460 | url = join(first, second) | |||
|
461 | for a in arg: | |||
|
462 | url = join(url, a) | |||
|
463 | return url | |||
|
464 | ||||
|
465 | def hexsha1(data): | |||
|
466 | """hexsha1 returns the hex-encoded sha1 sum of the data in the file-like | |||
|
467 | object data""" | |||
|
468 | h = hashlib.sha1() | |||
|
469 | for chunk in util.filechunkiter(data): | |||
|
470 | h.update(chunk) | |||
|
471 | return h.hexdigest() | |||
|
472 | ||||
|
473 | def httpsendfile(ui, filename): | |||
|
474 | try: | |||
|
475 | # Mercurial >= 1.9 | |||
|
476 | return httpconnection.httpsendfile(ui, filename, 'rb') | |||
|
477 | except ImportError: | |||
|
478 | if 'ui' in inspect.getargspec(url_.httpsendfile.__init__)[0]: | |||
|
479 | # Mercurial == 1.8 | |||
|
480 | return url_.httpsendfile(ui, filename, 'rb') | |||
|
481 | else: | |||
|
482 | # Mercurial <= 1.7 | |||
|
483 | return url_.httpsendfile(filename, 'rb') | |||
|
484 | ||||
|
485 | # Convert a path to a unix style path. This is used to give a | |||
|
486 | # canonical path to the lfdirstate. | |||
|
487 | def unixpath(path): | |||
|
488 | return os.path.normpath(path).replace(os.sep, '/') | |||
|
489 | ||||
|
490 | def islfilesrepo(repo): | |||
|
491 | return 'largefiles' in repo.requirements and any_(shortname+'/' in f[0] for f in | |||
|
492 | repo.store.datafiles()) | |||
|
493 | ||||
|
494 | def any_(gen): | |||
|
495 | for x in gen: | |||
|
496 | if x: | |||
|
497 | return True | |||
|
498 | return False | |||
|
499 | ||||
|
500 | class storeprotonotcapable(BaseException): | |||
|
501 | def __init__(self, storetypes): | |||
|
502 | self.storetypes = storetypes |
@@ -0,0 +1,71 b'' | |||||
|
1 | # Copyright 2009-2010 Gregory P. Ward | |||
|
2 | # Copyright 2009-2010 Intelerad Medical Systems Incorporated | |||
|
3 | # Copyright 2010-2011 Fog Creek Software | |||
|
4 | # Copyright 2010-2011 Unity Technologies | |||
|
5 | # | |||
|
6 | # This software may be used and distributed according to the terms of the | |||
|
7 | # GNU General Public License version 2 or any later version. | |||
|
8 | ||||
|
9 | '''Store class for local filesystem.''' | |||
|
10 | ||||
|
11 | import os | |||
|
12 | ||||
|
13 | from mercurial import util | |||
|
14 | from mercurial.i18n import _ | |||
|
15 | ||||
|
16 | import lfutil | |||
|
17 | import basestore | |||
|
18 | ||||
|
19 | class localstore(basestore.basestore): | |||
|
20 | '''Because there is a system wide cache, the local store always uses that | |||
|
21 | cache. Since the cache is updated elsewhere, we can just read from it here | |||
|
22 | as if it were the store.''' | |||
|
23 | ||||
|
24 | def __init__(self, ui, repo, remote): | |||
|
25 | url = os.path.join(remote.path, '.hg', lfutil.longname) | |||
|
26 | super(localstore, self).__init__(ui, repo, util.expandpath(url)) | |||
|
27 | ||||
|
28 | def put(self, source, filename, hash): | |||
|
29 | '''Any file that is put must already be in the system wide cache so do | |||
|
30 | nothing.''' | |||
|
31 | return | |||
|
32 | ||||
|
33 | def exists(self, hash): | |||
|
34 | return lfutil.insystemcache(self.repo.ui, hash) | |||
|
35 | ||||
|
36 | def _getfile(self, tmpfile, filename, hash): | |||
|
37 | if lfutil.insystemcache(self.ui, hash): | |||
|
38 | return lfutil.systemcachepath(self.ui, hash) | |||
|
39 | raise basestore.StoreError(filename, hash, '', | |||
|
40 | _("Can't get file locally")) | |||
|
41 | ||||
|
42 | def _verifyfile(self, cctx, cset, contents, standin, verified): | |||
|
43 | filename = lfutil.splitstandin(standin) | |||
|
44 | if not filename: | |||
|
45 | return False | |||
|
46 | fctx = cctx[standin] | |||
|
47 | key = (filename, fctx.filenode()) | |||
|
48 | if key in verified: | |||
|
49 | return False | |||
|
50 | ||||
|
51 | expecthash = fctx.data()[0:40] | |||
|
52 | verified.add(key) | |||
|
53 | if not lfutil.insystemcache(self.ui, expecthash): | |||
|
54 | self.ui.warn( | |||
|
55 | _('changeset %s: %s missing\n' | |||
|
56 | ' (looked for hash %s)\n') | |||
|
57 | % (cset, filename, expecthash)) | |||
|
58 | return True # failed | |||
|
59 | ||||
|
60 | if contents: | |||
|
61 | storepath = lfutil.systemcachepath(self.ui, expecthash) | |||
|
62 | actualhash = lfutil.hashfile(storepath) | |||
|
63 | if actualhash != expecthash: | |||
|
64 | self.ui.warn( | |||
|
65 | _('changeset %s: %s: contents differ\n' | |||
|
66 | ' (%s:\n' | |||
|
67 | ' expected hash %s,\n' | |||
|
68 | ' but got %s)\n') | |||
|
69 | % (cset, filename, storepath, expecthash, actualhash)) | |||
|
70 | return True # failed | |||
|
71 | return False |
This diff has been collapsed as it changes many lines, (902 lines changed) Show them Hide them | |||||
@@ -0,0 +1,902 b'' | |||||
|
1 | # Copyright 2009-2010 Gregory P. Ward | |||
|
2 | # Copyright 2009-2010 Intelerad Medical Systems Incorporated | |||
|
3 | # Copyright 2010-2011 Fog Creek Software | |||
|
4 | # Copyright 2010-2011 Unity Technologies | |||
|
5 | # | |||
|
6 | # This software may be used and distributed according to the terms of the | |||
|
7 | # GNU General Public License version 2 or any later version. | |||
|
8 | ||||
|
9 | '''Overridden Mercurial commands and functions for the largefiles extension''' | |||
|
10 | ||||
|
11 | import os | |||
|
12 | import copy | |||
|
13 | ||||
|
14 | from mercurial import hg, commands, util, cmdutil, match as match_, node, \ | |||
|
15 | archival, error, merge | |||
|
16 | from mercurial.i18n import _ | |||
|
17 | from mercurial.node import hex | |||
|
18 | from hgext import rebase | |||
|
19 | ||||
|
20 | try: | |||
|
21 | from mercurial import scmutil | |||
|
22 | except ImportError: | |||
|
23 | pass | |||
|
24 | ||||
|
25 | import lfutil | |||
|
26 | import lfcommands | |||
|
27 | ||||
|
28 | def installnormalfilesmatchfn(manifest): | |||
|
29 | '''overrides scmutil.match so that the matcher it returns will ignore all | |||
|
30 | largefiles''' | |||
|
31 | oldmatch = None # for the closure | |||
|
32 | def override_match(repo, pats=[], opts={}, globbed=False, | |||
|
33 | default='relpath'): | |||
|
34 | match = oldmatch(repo, pats, opts, globbed, default) | |||
|
35 | m = copy.copy(match) | |||
|
36 | notlfile = lambda f: not (lfutil.isstandin(f) or lfutil.standin(f) in | |||
|
37 | manifest) | |||
|
38 | m._files = filter(notlfile, m._files) | |||
|
39 | m._fmap = set(m._files) | |||
|
40 | orig_matchfn = m.matchfn | |||
|
41 | m.matchfn = lambda f: notlfile(f) and orig_matchfn(f) or None | |||
|
42 | return m | |||
|
43 | oldmatch = installmatchfn(override_match) | |||
|
44 | ||||
|
45 | def installmatchfn(f): | |||
|
46 | try: | |||
|
47 | # Mercurial >= 1.9 | |||
|
48 | oldmatch = scmutil.match | |||
|
49 | except ImportError: | |||
|
50 | # Mercurial <= 1.8 | |||
|
51 | oldmatch = cmdutil.match | |||
|
52 | setattr(f, 'oldmatch', oldmatch) | |||
|
53 | try: | |||
|
54 | # Mercurial >= 1.9 | |||
|
55 | scmutil.match = f | |||
|
56 | except ImportError: | |||
|
57 | # Mercurial <= 1.8 | |||
|
58 | cmdutil.match = f | |||
|
59 | return oldmatch | |||
|
60 | ||||
|
61 | def restorematchfn(): | |||
|
62 | '''restores scmutil.match to what it was before installnormalfilesmatchfn | |||
|
63 | was called. no-op if scmutil.match is its original function. | |||
|
64 | ||||
|
65 | Note that n calls to installnormalfilesmatchfn will require n calls to | |||
|
66 | restore matchfn to reverse''' | |||
|
67 | try: | |||
|
68 | # Mercurial >= 1.9 | |||
|
69 | scmutil.match = getattr(scmutil.match, 'oldmatch', scmutil.match) | |||
|
70 | except ImportError: | |||
|
71 | # Mercurial <= 1.8 | |||
|
72 | cmdutil.match = getattr(cmdutil.match, 'oldmatch', cmdutil.match) | |||
|
73 | ||||
|
74 | # -- Wrappers: modify existing commands -------------------------------- | |||
|
75 | ||||
|
76 | # Add works by going through the files that the user wanted to add | |||
|
77 | # and checking if they should be added as lfiles. Then making a new | |||
|
78 | # matcher which matches only the normal files and running the original | |||
|
79 | # version of add. | |||
|
80 | def override_add(orig, ui, repo, *pats, **opts): | |||
|
81 | large = opts.pop('large', None) | |||
|
82 | ||||
|
83 | lfsize = opts.pop('lfsize', None) | |||
|
84 | if not lfsize and lfutil.islfilesrepo(repo): | |||
|
85 | lfsize = ui.config(lfutil.longname, 'size', default='10') | |||
|
86 | if lfsize: | |||
|
87 | try: | |||
|
88 | lfsize = int(lfsize) | |||
|
89 | except ValueError: | |||
|
90 | raise util.Abort(_('largefiles: size must be an integer, was %s\n') % lfsize) | |||
|
91 | ||||
|
92 | lfmatcher = None | |||
|
93 | if os.path.exists(repo.wjoin(lfutil.shortname)): | |||
|
94 | lfpats = ui.config(lfutil.longname, 'patterns', default=()) | |||
|
95 | if lfpats: | |||
|
96 | lfpats = lfpats.split(' ') | |||
|
97 | lfmatcher = match_.match(repo.root, '', list(lfpats)) | |||
|
98 | ||||
|
99 | lfnames = [] | |||
|
100 | try: | |||
|
101 | # Mercurial >= 1.9 | |||
|
102 | m = scmutil.match(repo[None], pats, opts) | |||
|
103 | except ImportError: | |||
|
104 | # Mercurial <= 1.8 | |||
|
105 | m = cmdutil.match(repo, pats, opts) | |||
|
106 | m.bad = lambda x, y: None | |||
|
107 | wctx = repo[None] | |||
|
108 | for f in repo.walk(m): | |||
|
109 | exact = m.exact(f) | |||
|
110 | lfile = lfutil.standin(f) in wctx | |||
|
111 | nfile = f in wctx | |||
|
112 | exists = lfile or nfile | |||
|
113 | ||||
|
114 | # Don't warn the user when they attempt to add a normal tracked file. | |||
|
115 | # The normal add code will do that for us. | |||
|
116 | if exact and exists: | |||
|
117 | if lfile: | |||
|
118 | ui.warn(_('%s already a largefile\n') % f) | |||
|
119 | continue | |||
|
120 | ||||
|
121 | if exact or not exists: | |||
|
122 | if large or (lfsize and os.path.getsize(repo.wjoin(f)) >= \ | |||
|
123 | lfsize * 1024 * 1024) or (lfmatcher and lfmatcher(f)): | |||
|
124 | lfnames.append(f) | |||
|
125 | if ui.verbose or not exact: | |||
|
126 | ui.status(_('adding %s as a largefile\n') % m.rel(f)) | |||
|
127 | ||||
|
128 | bad = [] | |||
|
129 | standins = [] | |||
|
130 | ||||
|
131 | # Need to lock otherwise there could be a race condition inbetween when | |||
|
132 | # standins are created and added to the repo | |||
|
133 | wlock = repo.wlock() | |||
|
134 | try: | |||
|
135 | if not opts.get('dry_run'): | |||
|
136 | lfdirstate = lfutil.openlfdirstate(ui, repo) | |||
|
137 | for f in lfnames: | |||
|
138 | standinname = lfutil.standin(f) | |||
|
139 | lfutil.writestandin(repo, standinname, hash='', | |||
|
140 | executable=lfutil.getexecutable(repo.wjoin(f))) | |||
|
141 | standins.append(standinname) | |||
|
142 | if lfdirstate[f] == 'r': | |||
|
143 | lfdirstate.normallookup(f) | |||
|
144 | else: | |||
|
145 | lfdirstate.add(f) | |||
|
146 | lfdirstate.write() | |||
|
147 | bad += [lfutil.splitstandin(f) for f in lfutil.repo_add(repo, | |||
|
148 | standins) if f in m.files()] | |||
|
149 | finally: | |||
|
150 | wlock.release() | |||
|
151 | ||||
|
152 | installnormalfilesmatchfn(repo[None].manifest()) | |||
|
153 | result = orig(ui, repo, *pats, **opts) | |||
|
154 | restorematchfn() | |||
|
155 | ||||
|
156 | return (result == 1 or bad) and 1 or 0 | |||
|
157 | ||||
|
158 | def override_remove(orig, ui, repo, *pats, **opts): | |||
|
159 | manifest = repo[None].manifest() | |||
|
160 | installnormalfilesmatchfn(manifest) | |||
|
161 | orig(ui, repo, *pats, **opts) | |||
|
162 | restorematchfn() | |||
|
163 | ||||
|
164 | after, force = opts.get('after'), opts.get('force') | |||
|
165 | if not pats and not after: | |||
|
166 | raise util.Abort(_('no files specified')) | |||
|
167 | try: | |||
|
168 | # Mercurial >= 1.9 | |||
|
169 | m = scmutil.match(repo[None], pats, opts) | |||
|
170 | except ImportError: | |||
|
171 | # Mercurial <= 1.8 | |||
|
172 | m = cmdutil.match(repo, pats, opts) | |||
|
173 | try: | |||
|
174 | repo.lfstatus = True | |||
|
175 | s = repo.status(match=m, clean=True) | |||
|
176 | finally: | |||
|
177 | repo.lfstatus = False | |||
|
178 | modified, added, deleted, clean = [[f for f in list if lfutil.standin(f) \ | |||
|
179 | in manifest] for list in [s[0], s[1], s[3], s[6]]] | |||
|
180 | ||||
|
181 | def warn(files, reason): | |||
|
182 | for f in files: | |||
|
183 | ui.warn(_('not removing %s: file %s (use -f to force removal)\n') | |||
|
184 | % (m.rel(f), reason)) | |||
|
185 | ||||
|
186 | if force: | |||
|
187 | remove, forget = modified + deleted + clean, added | |||
|
188 | elif after: | |||
|
189 | remove, forget = deleted, [] | |||
|
190 | warn(modified + added + clean, _('still exists')) | |||
|
191 | else: | |||
|
192 | remove, forget = deleted + clean, [] | |||
|
193 | warn(modified, _('is modified')) | |||
|
194 | warn(added, _('has been marked for add')) | |||
|
195 | ||||
|
196 | for f in sorted(remove + forget): | |||
|
197 | if ui.verbose or not m.exact(f): | |||
|
198 | ui.status(_('removing %s\n') % m.rel(f)) | |||
|
199 | ||||
|
200 | # Need to lock because standin files are deleted then removed from the | |||
|
201 | # repository and we could race inbetween. | |||
|
202 | wlock = repo.wlock() | |||
|
203 | try: | |||
|
204 | lfdirstate = lfutil.openlfdirstate(ui, repo) | |||
|
205 | for f in remove: | |||
|
206 | if not after: | |||
|
207 | os.unlink(repo.wjoin(f)) | |||
|
208 | currentdir = os.path.split(f)[0] | |||
|
209 | while currentdir and not os.listdir(repo.wjoin(currentdir)): | |||
|
210 | os.rmdir(repo.wjoin(currentdir)) | |||
|
211 | currentdir = os.path.split(currentdir)[0] | |||
|
212 | lfdirstate.remove(f) | |||
|
213 | lfdirstate.write() | |||
|
214 | ||||
|
215 | forget = [lfutil.standin(f) for f in forget] | |||
|
216 | remove = [lfutil.standin(f) for f in remove] | |||
|
217 | lfutil.repo_forget(repo, forget) | |||
|
218 | lfutil.repo_remove(repo, remove, unlink=True) | |||
|
219 | finally: | |||
|
220 | wlock.release() | |||
|
221 | ||||
|
222 | def override_status(orig, ui, repo, *pats, **opts): | |||
|
223 | try: | |||
|
224 | repo.lfstatus = True | |||
|
225 | return orig(ui, repo, *pats, **opts) | |||
|
226 | finally: | |||
|
227 | repo.lfstatus = False | |||
|
228 | ||||
|
229 | def override_log(orig, ui, repo, *pats, **opts): | |||
|
230 | try: | |||
|
231 | repo.lfstatus = True | |||
|
232 | orig(ui, repo, *pats, **opts) | |||
|
233 | finally: | |||
|
234 | repo.lfstatus = False | |||
|
235 | ||||
|
236 | def override_verify(orig, ui, repo, *pats, **opts): | |||
|
237 | large = opts.pop('large', False) | |||
|
238 | all = opts.pop('lfa', False) | |||
|
239 | contents = opts.pop('lfc', False) | |||
|
240 | ||||
|
241 | result = orig(ui, repo, *pats, **opts) | |||
|
242 | if large: | |||
|
243 | result = result or lfcommands.verifylfiles(ui, repo, all, contents) | |||
|
244 | return result | |||
|
245 | ||||
|
246 | # Override needs to refresh standins so that update's normal merge | |||
|
247 | # will go through properly. Then the other update hook (overriding repo.update) | |||
|
248 | # will get the new files. Filemerge is also overriden so that the merge | |||
|
249 | # will merge standins correctly. | |||
|
250 | def override_update(orig, ui, repo, *pats, **opts): | |||
|
251 | lfdirstate = lfutil.openlfdirstate(ui, repo) | |||
|
252 | s = lfdirstate.status(match_.always(repo.root, repo.getcwd()), [], False, | |||
|
253 | False, False) | |||
|
254 | (unsure, modified, added, removed, missing, unknown, ignored, clean) = s | |||
|
255 | ||||
|
256 | # Need to lock between the standins getting updated and their lfiles | |||
|
257 | # getting updated | |||
|
258 | wlock = repo.wlock() | |||
|
259 | try: | |||
|
260 | if opts['check']: | |||
|
261 | mod = len(modified) > 0 | |||
|
262 | for lfile in unsure: | |||
|
263 | standin = lfutil.standin(lfile) | |||
|
264 | if repo['.'][standin].data().strip() != \ | |||
|
265 | lfutil.hashfile(repo.wjoin(lfile)): | |||
|
266 | mod = True | |||
|
267 | else: | |||
|
268 | lfdirstate.normal(lfile) | |||
|
269 | lfdirstate.write() | |||
|
270 | if mod: | |||
|
271 | raise util.Abort(_('uncommitted local changes')) | |||
|
272 | # XXX handle removed differently | |||
|
273 | if not opts['clean']: | |||
|
274 | for lfile in unsure + modified + added: | |||
|
275 | lfutil.updatestandin(repo, lfutil.standin(lfile)) | |||
|
276 | finally: | |||
|
277 | wlock.release() | |||
|
278 | return orig(ui, repo, *pats, **opts) | |||
|
279 | ||||
|
280 | # Override filemerge to prompt the user about how they wish to merge lfiles. | |||
|
281 | # This will handle identical edits, and copy/rename + edit without prompting | |||
|
282 | # the user. | |||
|
283 | def override_filemerge(origfn, repo, mynode, orig, fcd, fco, fca): | |||
|
284 | # Use better variable names here. Because this is a wrapper we cannot | |||
|
285 | # change the variable names in the function declaration. | |||
|
286 | fcdest, fcother, fcancestor = fcd, fco, fca | |||
|
287 | if not lfutil.isstandin(orig): | |||
|
288 | return origfn(repo, mynode, orig, fcdest, fcother, fcancestor) | |||
|
289 | else: | |||
|
290 | if not fcother.cmp(fcdest): # files identical? | |||
|
291 | return None | |||
|
292 | ||||
|
293 | # backwards, use working dir parent as ancestor | |||
|
294 | if fcancestor == fcother: | |||
|
295 | fcancestor = fcdest.parents()[0] | |||
|
296 | ||||
|
297 | if orig != fcother.path(): | |||
|
298 | repo.ui.status(_('merging %s and %s to %s\n') | |||
|
299 | % (lfutil.splitstandin(orig), | |||
|
300 | lfutil.splitstandin(fcother.path()), | |||
|
301 | lfutil.splitstandin(fcdest.path()))) | |||
|
302 | else: | |||
|
303 | repo.ui.status(_('merging %s\n') | |||
|
304 | % lfutil.splitstandin(fcdest.path())) | |||
|
305 | ||||
|
306 | if fcancestor.path() != fcother.path() and fcother.data() == \ | |||
|
307 | fcancestor.data(): | |||
|
308 | return 0 | |||
|
309 | if fcancestor.path() != fcdest.path() and fcdest.data() == \ | |||
|
310 | fcancestor.data(): | |||
|
311 | repo.wwrite(fcdest.path(), fcother.data(), fcother.flags()) | |||
|
312 | return 0 | |||
|
313 | ||||
|
314 | if repo.ui.promptchoice(_('largefile %s has a merge conflict\n' | |||
|
315 | 'keep (l)ocal or take (o)ther?') % | |||
|
316 | lfutil.splitstandin(orig), | |||
|
317 | (_('&Local'), _('&Other')), 0) == 0: | |||
|
318 | return 0 | |||
|
319 | else: | |||
|
320 | repo.wwrite(fcdest.path(), fcother.data(), fcother.flags()) | |||
|
321 | return 0 | |||
|
322 | ||||
|
323 | # Copy first changes the matchers to match standins instead of lfiles. | |||
|
324 | # Then it overrides util.copyfile in that function it checks if the destination | |||
|
325 | # lfile already exists. It also keeps a list of copied files so that the lfiles | |||
|
326 | # can be copied and the dirstate updated. | |||
|
327 | def override_copy(orig, ui, repo, pats, opts, rename=False): | |||
|
328 | # doesn't remove lfile on rename | |||
|
329 | if len(pats) < 2: | |||
|
330 | # this isn't legal, let the original function deal with it | |||
|
331 | return orig(ui, repo, pats, opts, rename) | |||
|
332 | ||||
|
333 | def makestandin(relpath): | |||
|
334 | try: | |||
|
335 | # Mercurial >= 1.9 | |||
|
336 | path = scmutil.canonpath(repo.root, repo.getcwd(), relpath) | |||
|
337 | except ImportError: | |||
|
338 | # Mercurial <= 1.8 | |||
|
339 | path = util.canonpath(repo.root, repo.getcwd(), relpath) | |||
|
340 | return os.path.join(os.path.relpath('.', repo.getcwd()), | |||
|
341 | lfutil.standin(path)) | |||
|
342 | ||||
|
343 | try: | |||
|
344 | # Mercurial >= 1.9 | |||
|
345 | fullpats = scmutil.expandpats(pats) | |||
|
346 | except ImportError: | |||
|
347 | # Mercurial <= 1.8 | |||
|
348 | fullpats = cmdutil.expandpats(pats) | |||
|
349 | dest = fullpats[-1] | |||
|
350 | ||||
|
351 | if os.path.isdir(dest): | |||
|
352 | if not os.path.isdir(makestandin(dest)): | |||
|
353 | os.makedirs(makestandin(dest)) | |||
|
354 | # This could copy both lfiles and normal files in one command, but we don't | |||
|
355 | # want to do that first replace their matcher to only match normal files | |||
|
356 | # and run it then replace it to just match lfiles and run it again | |||
|
357 | nonormalfiles = False | |||
|
358 | nolfiles = False | |||
|
359 | try: | |||
|
360 | installnormalfilesmatchfn(repo[None].manifest()) | |||
|
361 | result = orig(ui, repo, pats, opts, rename) | |||
|
362 | except util.Abort, e: | |||
|
363 | if str(e) != 'no files to copy': | |||
|
364 | raise e | |||
|
365 | else: | |||
|
366 | nonormalfiles = True | |||
|
367 | result = 0 | |||
|
368 | finally: | |||
|
369 | restorematchfn() | |||
|
370 | ||||
|
371 | # The first rename can cause our current working directory to be removed. | |||
|
372 | # In that case there is nothing left to copy/rename so just quit. | |||
|
373 | try: | |||
|
374 | repo.getcwd() | |||
|
375 | except OSError: | |||
|
376 | return result | |||
|
377 | ||||
|
378 | try: | |||
|
379 | # When we call orig below it creates the standins but we don't add them | |||
|
380 | # to the dir state until later so lock during that time. | |||
|
381 | wlock = repo.wlock() | |||
|
382 | ||||
|
383 | manifest = repo[None].manifest() | |||
|
384 | oldmatch = None # for the closure | |||
|
385 | def override_match(repo, pats=[], opts={}, globbed=False, | |||
|
386 | default='relpath'): | |||
|
387 | newpats = [] | |||
|
388 | # The patterns were previously mangled to add the standin | |||
|
389 | # directory; we need to remove that now | |||
|
390 | for pat in pats: | |||
|
391 | if match_.patkind(pat) is None and lfutil.shortname in pat: | |||
|
392 | newpats.append(pat.replace(lfutil.shortname, '')) | |||
|
393 | else: | |||
|
394 | newpats.append(pat) | |||
|
395 | match = oldmatch(repo, newpats, opts, globbed, default) | |||
|
396 | m = copy.copy(match) | |||
|
397 | lfile = lambda f: lfutil.standin(f) in manifest | |||
|
398 | m._files = [lfutil.standin(f) for f in m._files if lfile(f)] | |||
|
399 | m._fmap = set(m._files) | |||
|
400 | orig_matchfn = m.matchfn | |||
|
401 | m.matchfn = lambda f: lfutil.isstandin(f) and \ | |||
|
402 | lfile(lfutil.splitstandin(f)) and \ | |||
|
403 | orig_matchfn(lfutil.splitstandin(f)) or None | |||
|
404 | return m | |||
|
405 | oldmatch = installmatchfn(override_match) | |||
|
406 | listpats = [] | |||
|
407 | for pat in pats: | |||
|
408 | if match_.patkind(pat) is not None: | |||
|
409 | listpats.append(pat) | |||
|
410 | else: | |||
|
411 | listpats.append(makestandin(pat)) | |||
|
412 | ||||
|
413 | try: | |||
|
414 | origcopyfile = util.copyfile | |||
|
415 | copiedfiles = [] | |||
|
416 | def override_copyfile(src, dest): | |||
|
417 | if lfutil.shortname in src and lfutil.shortname in dest: | |||
|
418 | destlfile = dest.replace(lfutil.shortname, '') | |||
|
419 | if not opts['force'] and os.path.exists(destlfile): | |||
|
420 | raise IOError('', | |||
|
421 | _('destination largefile already exists')) | |||
|
422 | copiedfiles.append((src, dest)) | |||
|
423 | origcopyfile(src, dest) | |||
|
424 | ||||
|
425 | util.copyfile = override_copyfile | |||
|
426 | result += orig(ui, repo, listpats, opts, rename) | |||
|
427 | finally: | |||
|
428 | util.copyfile = origcopyfile | |||
|
429 | ||||
|
430 | lfdirstate = lfutil.openlfdirstate(ui, repo) | |||
|
431 | for (src, dest) in copiedfiles: | |||
|
432 | if lfutil.shortname in src and lfutil.shortname in dest: | |||
|
433 | srclfile = src.replace(lfutil.shortname, '') | |||
|
434 | destlfile = dest.replace(lfutil.shortname, '') | |||
|
435 | destlfiledir = os.path.dirname(destlfile) or '.' | |||
|
436 | if not os.path.isdir(destlfiledir): | |||
|
437 | os.makedirs(destlfiledir) | |||
|
438 | if rename: | |||
|
439 | os.rename(srclfile, destlfile) | |||
|
440 | lfdirstate.remove(os.path.relpath(srclfile, | |||
|
441 | repo.root)) | |||
|
442 | else: | |||
|
443 | util.copyfile(srclfile, destlfile) | |||
|
444 | lfdirstate.add(os.path.relpath(destlfile, | |||
|
445 | repo.root)) | |||
|
446 | lfdirstate.write() | |||
|
447 | except util.Abort, e: | |||
|
448 | if str(e) != 'no files to copy': | |||
|
449 | raise e | |||
|
450 | else: | |||
|
451 | nolfiles = True | |||
|
452 | finally: | |||
|
453 | restorematchfn() | |||
|
454 | wlock.release() | |||
|
455 | ||||
|
456 | if nolfiles and nonormalfiles: | |||
|
457 | raise util.Abort(_('no files to copy')) | |||
|
458 | ||||
|
459 | return result | |||
|
460 | ||||
|
461 | # When the user calls revert, we have to be careful to not revert any changes | |||
|
462 | # to other lfiles accidentally. This means we have to keep track of the lfiles | |||
|
463 | # that are being reverted so we only pull down the necessary lfiles. | |||
|
464 | # | |||
|
465 | # Standins are only updated (to match the hash of lfiles) before commits. | |||
|
466 | # Update the standins then run the original revert (changing the matcher to hit | |||
|
467 | # standins instead of lfiles). Based on the resulting standins update the | |||
|
468 | # lfiles. Then return the standins to their proper state | |||
|
469 | def override_revert(orig, ui, repo, *pats, **opts): | |||
|
470 | # Because we put the standins in a bad state (by updating them) and then | |||
|
471 | # return them to a correct state we need to lock to prevent others from | |||
|
472 | # changing them in their incorrect state. | |||
|
473 | wlock = repo.wlock() | |||
|
474 | try: | |||
|
475 | lfdirstate = lfutil.openlfdirstate(ui, repo) | |||
|
476 | (modified, added, removed, missing, unknown, ignored, clean) = \ | |||
|
477 | lfutil.lfdirstate_status(lfdirstate, repo, repo['.'].rev()) | |||
|
478 | for lfile in modified: | |||
|
479 | lfutil.updatestandin(repo, lfutil.standin(lfile)) | |||
|
480 | ||||
|
481 | try: | |||
|
482 | ctx = repo[opts.get('rev')] | |||
|
483 | oldmatch = None # for the closure | |||
|
484 | def override_match(ctxorrepo, pats=[], opts={}, globbed=False, | |||
|
485 | default='relpath'): | |||
|
486 | if hasattr(ctxorrepo, 'match'): | |||
|
487 | ctx0 = ctxorrepo | |||
|
488 | else: | |||
|
489 | ctx0 = ctxorrepo[None] | |||
|
490 | match = oldmatch(ctxorrepo, pats, opts, globbed, default) | |||
|
491 | m = copy.copy(match) | |||
|
492 | def tostandin(f): | |||
|
493 | if lfutil.standin(f) in ctx0 or lfutil.standin(f) in ctx: | |||
|
494 | return lfutil.standin(f) | |||
|
495 | elif lfutil.standin(f) in repo[None]: | |||
|
496 | return None | |||
|
497 | return f | |||
|
498 | m._files = [tostandin(f) for f in m._files] | |||
|
499 | m._files = [f for f in m._files if f is not None] | |||
|
500 | m._fmap = set(m._files) | |||
|
501 | orig_matchfn = m.matchfn | |||
|
502 | def matchfn(f): | |||
|
503 | if lfutil.isstandin(f): | |||
|
504 | # We need to keep track of what lfiles are being | |||
|
505 | # matched so we know which ones to update later | |||
|
506 | # (otherwise we revert changes to other lfiles | |||
|
507 | # accidentally). This is repo specific, so duckpunch | |||
|
508 | # the repo object to keep the list of lfiles for us | |||
|
509 | # later. | |||
|
510 | if orig_matchfn(lfutil.splitstandin(f)) and \ | |||
|
511 | (f in repo[None] or f in ctx): | |||
|
512 | lfileslist = getattr(repo, '_lfilestoupdate', []) | |||
|
513 | lfileslist.append(lfutil.splitstandin(f)) | |||
|
514 | repo._lfilestoupdate = lfileslist | |||
|
515 | return True | |||
|
516 | else: | |||
|
517 | return False | |||
|
518 | return orig_matchfn(f) | |||
|
519 | m.matchfn = matchfn | |||
|
520 | return m | |||
|
521 | oldmatch = installmatchfn(override_match) | |||
|
522 | try: | |||
|
523 | # Mercurial >= 1.9 | |||
|
524 | scmutil.match | |||
|
525 | matches = override_match(repo[None], pats, opts) | |||
|
526 | except ImportError: | |||
|
527 | # Mercurial <= 1.8 | |||
|
528 | matches = override_match(repo, pats, opts) | |||
|
529 | orig(ui, repo, *pats, **opts) | |||
|
530 | finally: | |||
|
531 | restorematchfn() | |||
|
532 | lfileslist = getattr(repo, '_lfilestoupdate', []) | |||
|
533 | lfcommands.updatelfiles(ui, repo, filelist=lfileslist, printmessage=False) | |||
|
534 | # Empty out the lfiles list so we start fresh next time | |||
|
535 | repo._lfilestoupdate = [] | |||
|
536 | for lfile in modified: | |||
|
537 | if lfile in lfileslist: | |||
|
538 | if os.path.exists(repo.wjoin(lfutil.standin(lfile))) and lfile\ | |||
|
539 | in repo['.']: | |||
|
540 | lfutil.writestandin(repo, lfutil.standin(lfile), | |||
|
541 | repo['.'][lfile].data().strip(), | |||
|
542 | 'x' in repo['.'][lfile].flags()) | |||
|
543 | lfdirstate = lfutil.openlfdirstate(ui, repo) | |||
|
544 | for lfile in added: | |||
|
545 | standin = lfutil.standin(lfile) | |||
|
546 | if standin not in ctx and (standin in matches or opts.get('all')): | |||
|
547 | if lfile in lfdirstate: | |||
|
548 | try: | |||
|
549 | # Mercurial >= 1.9 | |||
|
550 | lfdirstate.drop(lfile) | |||
|
551 | except AttributeError: | |||
|
552 | # Mercurial <= 1.8 | |||
|
553 | lfdirstate.forget(lfile) | |||
|
554 | util.unlinkpath(repo.wjoin(standin)) | |||
|
555 | lfdirstate.write() | |||
|
556 | finally: | |||
|
557 | wlock.release() | |||
|
558 | ||||
|
559 | def hg_update(orig, repo, node): | |||
|
560 | result = orig(repo, node) | |||
|
561 | # XXX check if it worked first | |||
|
562 | lfcommands.updatelfiles(repo.ui, repo) | |||
|
563 | return result | |||
|
564 | ||||
|
565 | def hg_clean(orig, repo, node, show_stats=True): | |||
|
566 | result = orig(repo, node, show_stats) | |||
|
567 | lfcommands.updatelfiles(repo.ui, repo) | |||
|
568 | return result | |||
|
569 | ||||
|
570 | def hg_merge(orig, repo, node, force=None, remind=True): | |||
|
571 | result = orig(repo, node, force, remind) | |||
|
572 | lfcommands.updatelfiles(repo.ui, repo) | |||
|
573 | return result | |||
|
574 | ||||
|
575 | # When we rebase a repository with remotely changed lfiles, we need | |||
|
576 | # to take some extra care so that the lfiles are correctly updated | |||
|
577 | # in the working copy | |||
|
578 | def override_pull(orig, ui, repo, source=None, **opts): | |||
|
579 | if opts.get('rebase', False): | |||
|
580 | repo._isrebasing = True | |||
|
581 | try: | |||
|
582 | if opts.get('update'): | |||
|
583 | del opts['update'] | |||
|
584 | ui.debug('--update and --rebase are not compatible, ignoring ' | |||
|
585 | 'the update flag\n') | |||
|
586 | del opts['rebase'] | |||
|
587 | try: | |||
|
588 | # Mercurial >= 1.9 | |||
|
589 | cmdutil.bailifchanged(repo) | |||
|
590 | except AttributeError: | |||
|
591 | # Mercurial <= 1.8 | |||
|
592 | cmdutil.bail_if_changed(repo) | |||
|
593 | revsprepull = len(repo) | |||
|
594 | origpostincoming = commands.postincoming | |||
|
595 | def _dummy(*args, **kwargs): | |||
|
596 | pass | |||
|
597 | commands.postincoming = _dummy | |||
|
598 | repo.lfpullsource = source | |||
|
599 | if not source: | |||
|
600 | source = 'default' | |||
|
601 | try: | |||
|
602 | result = commands.pull(ui, repo, source, **opts) | |||
|
603 | finally: | |||
|
604 | commands.postincoming = origpostincoming | |||
|
605 | revspostpull = len(repo) | |||
|
606 | if revspostpull > revsprepull: | |||
|
607 | result = result or rebase.rebase(ui, repo) | |||
|
608 | finally: | |||
|
609 | repo._isrebasing = False | |||
|
610 | else: | |||
|
611 | repo.lfpullsource = source | |||
|
612 | if not source: | |||
|
613 | source = 'default' | |||
|
614 | result = orig(ui, repo, source, **opts) | |||
|
615 | return result | |||
|
616 | ||||
|
617 | def override_rebase(orig, ui, repo, **opts): | |||
|
618 | repo._isrebasing = True | |||
|
619 | try: | |||
|
620 | orig(ui, repo, **opts) | |||
|
621 | finally: | |||
|
622 | repo._isrebasing = False | |||
|
623 | ||||
|
624 | def override_archive(orig, repo, dest, node, kind, decode=True, matchfn=None, | |||
|
625 | prefix=None, mtime=None, subrepos=None): | |||
|
626 | # No need to lock because we are only reading history and lfile caches | |||
|
627 | # neither of which are modified | |||
|
628 | ||||
|
629 | lfcommands.cachelfiles(repo.ui, repo, node) | |||
|
630 | ||||
|
631 | if kind not in archival.archivers: | |||
|
632 | raise util.Abort(_("unknown archive type '%s'") % kind) | |||
|
633 | ||||
|
634 | ctx = repo[node] | |||
|
635 | ||||
|
636 | # In Mercurial <= 1.5 the prefix is passed to the archiver so try that | |||
|
637 | # if that doesn't work we are probably in Mercurial >= 1.6 where the | |||
|
638 | # prefix is not handled by the archiver | |||
|
639 | try: | |||
|
640 | archiver = archival.archivers[kind](dest, prefix, mtime or \ | |||
|
641 | ctx.date()[0]) | |||
|
642 | ||||
|
643 | def write(name, mode, islink, getdata): | |||
|
644 | if matchfn and not matchfn(name): | |||
|
645 | return | |||
|
646 | data = getdata() | |||
|
647 | if decode: | |||
|
648 | data = repo.wwritedata(name, data) | |||
|
649 | archiver.addfile(name, mode, islink, data) | |||
|
650 | except TypeError: | |||
|
651 | if kind == 'files': | |||
|
652 | if prefix: | |||
|
653 | raise util.Abort( | |||
|
654 | _('cannot give prefix when archiving to files')) | |||
|
655 | else: | |||
|
656 | prefix = archival.tidyprefix(dest, kind, prefix) | |||
|
657 | ||||
|
658 | def write(name, mode, islink, getdata): | |||
|
659 | if matchfn and not matchfn(name): | |||
|
660 | return | |||
|
661 | data = getdata() | |||
|
662 | if decode: | |||
|
663 | data = repo.wwritedata(name, data) | |||
|
664 | archiver.addfile(prefix + name, mode, islink, data) | |||
|
665 | ||||
|
666 | archiver = archival.archivers[kind](dest, mtime or ctx.date()[0]) | |||
|
667 | ||||
|
668 | if repo.ui.configbool("ui", "archivemeta", True): | |||
|
669 | def metadata(): | |||
|
670 | base = 'repo: %s\nnode: %s\nbranch: %s\n' % ( | |||
|
671 | hex(repo.changelog.node(0)), hex(node), ctx.branch()) | |||
|
672 | ||||
|
673 | tags = ''.join('tag: %s\n' % t for t in ctx.tags() | |||
|
674 | if repo.tagtype(t) == 'global') | |||
|
675 | if not tags: | |||
|
676 | repo.ui.pushbuffer() | |||
|
677 | opts = {'template': '{latesttag}\n{latesttagdistance}', | |||
|
678 | 'style': '', 'patch': None, 'git': None} | |||
|
679 | cmdutil.show_changeset(repo.ui, repo, opts).show(ctx) | |||
|
680 | ltags, dist = repo.ui.popbuffer().split('\n') | |||
|
681 | tags = ''.join('latesttag: %s\n' % t for t in ltags.split(':')) | |||
|
682 | tags += 'latesttagdistance: %s\n' % dist | |||
|
683 | ||||
|
684 | return base + tags | |||
|
685 | ||||
|
686 | write('.hg_archival.txt', 0644, False, metadata) | |||
|
687 | ||||
|
688 | for f in ctx: | |||
|
689 | ff = ctx.flags(f) | |||
|
690 | getdata = ctx[f].data | |||
|
691 | if lfutil.isstandin(f): | |||
|
692 | path = lfutil.findfile(repo, getdata().strip()) | |||
|
693 | f = lfutil.splitstandin(f) | |||
|
694 | ||||
|
695 | def getdatafn(): | |||
|
696 | try: | |||
|
697 | fd = open(path, 'rb') | |||
|
698 | return fd.read() | |||
|
699 | finally: | |||
|
700 | fd.close() | |||
|
701 | ||||
|
702 | getdata = getdatafn | |||
|
703 | write(f, 'x' in ff and 0755 or 0644, 'l' in ff, getdata) | |||
|
704 | ||||
|
705 | if subrepos: | |||
|
706 | for subpath in ctx.substate: | |||
|
707 | sub = ctx.sub(subpath) | |||
|
708 | try: | |||
|
709 | sub.archive(repo.ui, archiver, prefix) | |||
|
710 | except TypeError: | |||
|
711 | sub.archive(archiver, prefix) | |||
|
712 | ||||
|
713 | archiver.done() | |||
|
714 | ||||
|
715 | # If a lfile is modified the change is not reflected in its standin until a | |||
|
716 | # commit. cmdutil.bailifchanged raises an exception if the repo has | |||
|
717 | # uncommitted changes. Wrap it to also check if lfiles were changed. This is | |||
|
718 | # used by bisect and backout. | |||
|
719 | def override_bailifchanged(orig, repo): | |||
|
720 | orig(repo) | |||
|
721 | repo.lfstatus = True | |||
|
722 | modified, added, removed, deleted = repo.status()[:4] | |||
|
723 | repo.lfstatus = False | |||
|
724 | if modified or added or removed or deleted: | |||
|
725 | raise util.Abort(_('outstanding uncommitted changes')) | |||
|
726 | ||||
|
727 | # Fetch doesn't use cmdutil.bail_if_changed so override it to add the check | |||
|
728 | def override_fetch(orig, ui, repo, *pats, **opts): | |||
|
729 | repo.lfstatus = True | |||
|
730 | modified, added, removed, deleted = repo.status()[:4] | |||
|
731 | repo.lfstatus = False | |||
|
732 | if modified or added or removed or deleted: | |||
|
733 | raise util.Abort(_('outstanding uncommitted changes')) | |||
|
734 | return orig(ui, repo, *pats, **opts) | |||
|
735 | ||||
|
736 | def override_forget(orig, ui, repo, *pats, **opts): | |||
|
737 | installnormalfilesmatchfn(repo[None].manifest()) | |||
|
738 | orig(ui, repo, *pats, **opts) | |||
|
739 | restorematchfn() | |||
|
740 | try: | |||
|
741 | # Mercurial >= 1.9 | |||
|
742 | m = scmutil.match(repo[None], pats, opts) | |||
|
743 | except ImportError: | |||
|
744 | # Mercurial <= 1.8 | |||
|
745 | m = cmdutil.match(repo, pats, opts) | |||
|
746 | ||||
|
747 | try: | |||
|
748 | repo.lfstatus = True | |||
|
749 | s = repo.status(match=m, clean=True) | |||
|
750 | finally: | |||
|
751 | repo.lfstatus = False | |||
|
752 | forget = sorted(s[0] + s[1] + s[3] + s[6]) | |||
|
753 | forget = [f for f in forget if lfutil.standin(f) in repo[None].manifest()] | |||
|
754 | ||||
|
755 | for f in forget: | |||
|
756 | if lfutil.standin(f) not in repo.dirstate and not \ | |||
|
757 | os.path.isdir(m.rel(lfutil.standin(f))): | |||
|
758 | ui.warn(_('not removing %s: file is already untracked\n') | |||
|
759 | % m.rel(f)) | |||
|
760 | ||||
|
761 | for f in forget: | |||
|
762 | if ui.verbose or not m.exact(f): | |||
|
763 | ui.status(_('removing %s\n') % m.rel(f)) | |||
|
764 | ||||
|
765 | # Need to lock because standin files are deleted then removed from the | |||
|
766 | # repository and we could race inbetween. | |||
|
767 | wlock = repo.wlock() | |||
|
768 | try: | |||
|
769 | lfdirstate = lfutil.openlfdirstate(ui, repo) | |||
|
770 | for f in forget: | |||
|
771 | if lfdirstate[f] == 'a': | |||
|
772 | lfdirstate.drop(f) | |||
|
773 | else: | |||
|
774 | lfdirstate.remove(f) | |||
|
775 | lfdirstate.write() | |||
|
776 | lfutil.repo_remove(repo, [lfutil.standin(f) for f in forget], | |||
|
777 | unlink=True) | |||
|
778 | finally: | |||
|
779 | wlock.release() | |||
|
780 | ||||
|
781 | def getoutgoinglfiles(ui, repo, dest=None, **opts): | |||
|
782 | dest = ui.expandpath(dest or 'default-push', dest or 'default') | |||
|
783 | dest, branches = hg.parseurl(dest, opts.get('branch')) | |||
|
784 | revs, checkout = hg.addbranchrevs(repo, repo, branches, opts.get('rev')) | |||
|
785 | if revs: | |||
|
786 | revs = [repo.lookup(rev) for rev in revs] | |||
|
787 | ||||
|
788 | # Mercurial <= 1.5 had remoteui in cmdutil, then it moved to hg | |||
|
789 | try: | |||
|
790 | remoteui = cmdutil.remoteui | |||
|
791 | except AttributeError: | |||
|
792 | remoteui = hg.remoteui | |||
|
793 | ||||
|
794 | try: | |||
|
795 | remote = hg.repository(remoteui(repo, opts), dest) | |||
|
796 | except error.RepoError: | |||
|
797 | return None | |||
|
798 | o = lfutil.findoutgoing(repo, remote, False) | |||
|
799 | if not o: | |||
|
800 | return None | |||
|
801 | o = repo.changelog.nodesbetween(o, revs)[0] | |||
|
802 | if opts.get('newest_first'): | |||
|
803 | o.reverse() | |||
|
804 | ||||
|
805 | toupload = set() | |||
|
806 | for n in o: | |||
|
807 | parents = [p for p in repo.changelog.parents(n) if p != node.nullid] | |||
|
808 | ctx = repo[n] | |||
|
809 | files = set(ctx.files()) | |||
|
810 | if len(parents) == 2: | |||
|
811 | mc = ctx.manifest() | |||
|
812 | mp1 = ctx.parents()[0].manifest() | |||
|
813 | mp2 = ctx.parents()[1].manifest() | |||
|
814 | for f in mp1: | |||
|
815 | if f not in mc: | |||
|
816 | files.add(f) | |||
|
817 | for f in mp2: | |||
|
818 | if f not in mc: | |||
|
819 | files.add(f) | |||
|
820 | for f in mc: | |||
|
821 | if mc[f] != mp1.get(f, None) or mc[f] != mp2.get(f, None): | |||
|
822 | files.add(f) | |||
|
823 | toupload = toupload.union(set([f for f in files if lfutil.isstandin(f)\ | |||
|
824 | and f in ctx])) | |||
|
825 | return toupload | |||
|
826 | ||||
|
827 | def override_outgoing(orig, ui, repo, dest=None, **opts): | |||
|
828 | orig(ui, repo, dest, **opts) | |||
|
829 | ||||
|
830 | if opts.pop('large', None): | |||
|
831 | toupload = getoutgoinglfiles(ui, repo, dest, **opts) | |||
|
832 | if toupload is None: | |||
|
833 | ui.status(_('largefiles: No remote repo\n')) | |||
|
834 | else: | |||
|
835 | ui.status(_('largefiles to upload:\n')) | |||
|
836 | for file in toupload: | |||
|
837 | ui.status(lfutil.splitstandin(file) + '\n') | |||
|
838 | ui.status('\n') | |||
|
839 | ||||
|
840 | def override_summary(orig, ui, repo, *pats, **opts): | |||
|
841 | orig(ui, repo, *pats, **opts) | |||
|
842 | ||||
|
843 | if opts.pop('large', None): | |||
|
844 | toupload = getoutgoinglfiles(ui, repo, None, **opts) | |||
|
845 | if toupload is None: | |||
|
846 | ui.status(_('largefiles: No remote repo\n')) | |||
|
847 | else: | |||
|
848 | ui.status(_('largefiles: %d to upload\n') % len(toupload)) | |||
|
849 | ||||
|
850 | def override_addremove(orig, ui, repo, *pats, **opts): | |||
|
851 | # Check if the parent or child has lfiles if they do don't allow it. If | |||
|
852 | # there is a symlink in the manifest then getting the manifest throws an | |||
|
853 | # exception catch it and let addremove deal with it. This happens in | |||
|
854 | # Mercurial's test test-addremove-symlink | |||
|
855 | try: | |||
|
856 | manifesttip = set(repo['tip'].manifest()) | |||
|
857 | except util.Abort: | |||
|
858 | manifesttip = set() | |||
|
859 | try: | |||
|
860 | manifestworking = set(repo[None].manifest()) | |||
|
861 | except util.Abort: | |||
|
862 | manifestworking = set() | |||
|
863 | ||||
|
864 | # Manifests are only iterable so turn them into sets then union | |||
|
865 | for file in manifesttip.union(manifestworking): | |||
|
866 | if file.startswith(lfutil.shortname): | |||
|
867 | raise util.Abort( | |||
|
868 | _('addremove cannot be run on a repo with largefiles')) | |||
|
869 | ||||
|
870 | return orig(ui, repo, *pats, **opts) | |||
|
871 | ||||
|
872 | # Calling purge with --all will cause the lfiles to be deleted. | |||
|
873 | # Override repo.status to prevent this from happening. | |||
|
874 | def override_purge(orig, ui, repo, *dirs, **opts): | |||
|
875 | oldstatus = repo.status | |||
|
876 | def override_status(node1='.', node2=None, match=None, ignored=False, | |||
|
877 | clean=False, unknown=False, listsubrepos=False): | |||
|
878 | r = oldstatus(node1, node2, match, ignored, clean, unknown, | |||
|
879 | listsubrepos) | |||
|
880 | lfdirstate = lfutil.openlfdirstate(ui, repo) | |||
|
881 | modified, added, removed, deleted, unknown, ignored, clean = r | |||
|
882 | unknown = [f for f in unknown if lfdirstate[f] == '?'] | |||
|
883 | ignored = [f for f in ignored if lfdirstate[f] == '?'] | |||
|
884 | return modified, added, removed, deleted, unknown, ignored, clean | |||
|
885 | repo.status = override_status | |||
|
886 | orig(ui, repo, *dirs, **opts) | |||
|
887 | repo.status = oldstatus | |||
|
888 | ||||
|
889 | def override_rollback(orig, ui, repo, **opts): | |||
|
890 | result = orig(ui, repo, **opts) | |||
|
891 | merge.update(repo, node=None, branchmerge=False, force=True, | |||
|
892 | partial=lfutil.isstandin) | |||
|
893 | lfdirstate = lfutil.openlfdirstate(ui, repo) | |||
|
894 | lfiles = lfutil.listlfiles(repo) | |||
|
895 | oldlfiles = lfutil.listlfiles(repo, repo[None].parents()[0].rev()) | |||
|
896 | for file in lfiles: | |||
|
897 | if file in oldlfiles: | |||
|
898 | lfdirstate.normallookup(file) | |||
|
899 | else: | |||
|
900 | lfdirstate.add(file) | |||
|
901 | lfdirstate.write() | |||
|
902 | return result |
@@ -0,0 +1,161 b'' | |||||
|
1 | # Copyright 2011 Fog Creek Software | |||
|
2 | # | |||
|
3 | # This software may be used and distributed according to the terms of the | |||
|
4 | # GNU General Public License version 2 or any later version. | |||
|
5 | ||||
|
6 | import os | |||
|
7 | import tempfile | |||
|
8 | import urllib2 | |||
|
9 | ||||
|
10 | from mercurial import error, httprepo, util, wireproto | |||
|
11 | from mercurial.i18n import _ | |||
|
12 | ||||
|
13 | import lfutil | |||
|
14 | ||||
|
15 | LARGEFILES_REQUIRED_MSG = '\nThis repository uses the largefiles extension.' \ | |||
|
16 | '\n\nPlease enable it in your Mercurial config ' \ | |||
|
17 | 'file.\n' | |||
|
18 | ||||
|
19 | def putlfile(repo, proto, sha): | |||
|
20 | """putlfile puts a largefile into a repository's local cache and into the | |||
|
21 | system cache.""" | |||
|
22 | f = None | |||
|
23 | proto.redirect() | |||
|
24 | try: | |||
|
25 | try: | |||
|
26 | f = tempfile.NamedTemporaryFile(mode='wb+', prefix='hg-putlfile-') | |||
|
27 | proto.getfile(f) | |||
|
28 | f.seek(0) | |||
|
29 | if sha != lfutil.hexsha1(f): | |||
|
30 | return wireproto.pushres(1) | |||
|
31 | lfutil.copytocacheabsolute(repo, f.name, sha) | |||
|
32 | except IOError: | |||
|
33 | repo.ui.warn( | |||
|
34 | _('error: could not put received data into largefile store')) | |||
|
35 | return wireproto.pushres(1) | |||
|
36 | finally: | |||
|
37 | if f: | |||
|
38 | f.close() | |||
|
39 | ||||
|
40 | return wireproto.pushres(0) | |||
|
41 | ||||
|
42 | def getlfile(repo, proto, sha): | |||
|
43 | """getlfile retrieves a largefile from the repository-local cache or system | |||
|
44 | cache.""" | |||
|
45 | filename = lfutil.findfile(repo, sha) | |||
|
46 | if not filename: | |||
|
47 | raise util.Abort(_('requested largefile %s not present in cache') % sha) | |||
|
48 | f = open(filename, 'rb') | |||
|
49 | length = os.fstat(f.fileno())[6] | |||
|
50 | # since we can't set an HTTP content-length header here, and mercurial core | |||
|
51 | # provides no way to give the length of a streamres (and reading the entire | |||
|
52 | # file into RAM would be ill-advised), we just send the length on the first | |||
|
53 | # line of the response, like the ssh proto does for string responses. | |||
|
54 | def generator(): | |||
|
55 | yield '%d\n' % length | |||
|
56 | for chunk in f: | |||
|
57 | yield chunk | |||
|
58 | return wireproto.streamres(generator()) | |||
|
59 | ||||
|
60 | def statlfile(repo, proto, sha): | |||
|
61 | """statlfile sends '2\n' if the largefile is missing, '1\n' if it has a | |||
|
62 | mismatched checksum, or '0\n' if it is in good condition""" | |||
|
63 | filename = lfutil.findfile(repo, sha) | |||
|
64 | if not filename: | |||
|
65 | return '2\n' | |||
|
66 | fd = None | |||
|
67 | try: | |||
|
68 | fd = open(filename, 'rb') | |||
|
69 | return lfutil.hexsha1(fd) == sha and '0\n' or '1\n' | |||
|
70 | finally: | |||
|
71 | if fd: | |||
|
72 | fd.close() | |||
|
73 | ||||
|
74 | def wirereposetup(ui, repo): | |||
|
75 | class lfileswirerepository(repo.__class__): | |||
|
76 | def putlfile(self, sha, fd): | |||
|
77 | # unfortunately, httprepository._callpush tries to convert its | |||
|
78 | # input file-like into a bundle before sending it, so we can't use | |||
|
79 | # it ... | |||
|
80 | if issubclass(self.__class__, httprepo.httprepository): | |||
|
81 | try: | |||
|
82 | return int(self._call('putlfile', data=fd, sha=sha, | |||
|
83 | headers={'content-type':'application/mercurial-0.1'})) | |||
|
84 | except (ValueError, urllib2.HTTPError): | |||
|
85 | return 1 | |||
|
86 | # ... but we can't use sshrepository._call because the data= | |||
|
87 | # argument won't get sent, and _callpush does exactly what we want | |||
|
88 | # in this case: send the data straight through | |||
|
89 | else: | |||
|
90 | try: | |||
|
91 | ret, output = self._callpush("putlfile", fd, sha=sha) | |||
|
92 | if ret == "": | |||
|
93 | raise error.ResponseError(_('putlfile failed:'), | |||
|
94 | output) | |||
|
95 | return int(ret) | |||
|
96 | except IOError: | |||
|
97 | return 1 | |||
|
98 | except ValueError: | |||
|
99 | raise error.ResponseError( | |||
|
100 | _('putlfile failed (unexpected response):'), ret) | |||
|
101 | ||||
|
102 | def getlfile(self, sha): | |||
|
103 | stream = self._callstream("getlfile", sha=sha) | |||
|
104 | length = stream.readline() | |||
|
105 | try: | |||
|
106 | length = int(length) | |||
|
107 | except ValueError: | |||
|
108 | self._abort(error.ResponseError(_("unexpected response:"), length)) | |||
|
109 | return (length, stream) | |||
|
110 | ||||
|
111 | def statlfile(self, sha): | |||
|
112 | try: | |||
|
113 | return int(self._call("statlfile", sha=sha)) | |||
|
114 | except (ValueError, urllib2.HTTPError): | |||
|
115 | # if the server returns anything but an integer followed by a | |||
|
116 | # newline, newline, it's not speaking our language; if we get | |||
|
117 | # an HTTP error, we can't be sure the largefile is present; | |||
|
118 | # either way, consider it missing | |||
|
119 | return 2 | |||
|
120 | ||||
|
121 | repo.__class__ = lfileswirerepository | |||
|
122 | ||||
|
123 | # advertise the largefiles=serve capability | |||
|
124 | def capabilities(repo, proto): | |||
|
125 | return capabilities_orig(repo, proto) + ' largefiles=serve' | |||
|
126 | ||||
|
127 | # duplicate what Mercurial's new out-of-band errors mechanism does, because | |||
|
128 | # clients old and new alike both handle it well | |||
|
129 | def webproto_refuseclient(self, message): | |||
|
130 | self.req.header([('Content-Type', 'application/hg-error')]) | |||
|
131 | return message | |||
|
132 | ||||
|
133 | def sshproto_refuseclient(self, message): | |||
|
134 | self.ui.write_err('%s\n-\n' % message) | |||
|
135 | self.fout.write('\n') | |||
|
136 | self.fout.flush() | |||
|
137 | ||||
|
138 | return '' | |||
|
139 | ||||
|
140 | def heads(repo, proto): | |||
|
141 | if lfutil.islfilesrepo(repo): | |||
|
142 | try: | |||
|
143 | # Mercurial >= f4522df38c65 | |||
|
144 | return wireproto.ooberror(LARGEFILES_REQUIRED_MSG) | |||
|
145 | except AttributeError: | |||
|
146 | return proto.refuseclient(LARGEFILES_REQUIRED_MSG) | |||
|
147 | return wireproto.heads(repo, proto) | |||
|
148 | ||||
|
149 | def sshrepo_callstream(self, cmd, **args): | |||
|
150 | if cmd == 'heads' and self.capable('largefiles'): | |||
|
151 | cmd = 'lheads' | |||
|
152 | if cmd == 'batch' and self.capable('largefiles'): | |||
|
153 | args['cmds'] = args['cmds'].replace('heads ', 'lheads ') | |||
|
154 | return ssh_oldcallstream(self, cmd, **args) | |||
|
155 | ||||
|
156 | def httprepo_callstream(self, cmd, **args): | |||
|
157 | if cmd == 'heads' and self.capable('largefiles'): | |||
|
158 | cmd = 'lheads' | |||
|
159 | if cmd == 'batch' and self.capable('largefiles'): | |||
|
160 | args['cmds'] = args['cmds'].replace('heads ', 'lheads ') | |||
|
161 | return http_oldcallstream(self, cmd, **args) |
@@ -0,0 +1,106 b'' | |||||
|
1 | # Copyright 2010-2011 Fog Creek Software | |||
|
2 | # Copyright 2010-2011 Unity Technologies | |||
|
3 | # | |||
|
4 | # This software may be used and distributed according to the terms of the | |||
|
5 | # GNU General Public License version 2 or any later version. | |||
|
6 | ||||
|
7 | '''Remote largefile store; the base class for servestore''' | |||
|
8 | ||||
|
9 | import urllib2 | |||
|
10 | import HTTPError | |||
|
11 | ||||
|
12 | from mercurial import util | |||
|
13 | from mercurial.i18n import _ | |||
|
14 | ||||
|
15 | import lfutil | |||
|
16 | import basestore | |||
|
17 | ||||
|
18 | class remotestore(basestore.basestore): | |||
|
19 | """A largefile store accessed over a network""" | |||
|
20 | def __init__(self, ui, repo, url): | |||
|
21 | super(remotestore, self).__init__(ui, repo, url) | |||
|
22 | ||||
|
23 | def put(self, source, hash): | |||
|
24 | if self._verify(hash): | |||
|
25 | return | |||
|
26 | if self.sendfile(source, hash): | |||
|
27 | raise util.Abort( | |||
|
28 | _('remotestore: could not put %s to remote store %s') | |||
|
29 | % (source, self.url)) | |||
|
30 | self.ui.debug( | |||
|
31 | _('remotestore: put %s to remote store %s') % (source, self.url)) | |||
|
32 | ||||
|
33 | def exists(self, hash): | |||
|
34 | return self._verify(hash) | |||
|
35 | ||||
|
36 | def sendfile(self, filename, hash): | |||
|
37 | self.ui.debug('remotestore: sendfile(%s, %s)\n' % (filename, hash)) | |||
|
38 | fd = None | |||
|
39 | try: | |||
|
40 | try: | |||
|
41 | fd = lfutil.httpsendfile(self.ui, filename) | |||
|
42 | except IOError, e: | |||
|
43 | raise util.Abort( | |||
|
44 | _('remotestore: could not open file %s: %s') | |||
|
45 | % (filename, str(e))) | |||
|
46 | return self._put(hash, fd) | |||
|
47 | finally: | |||
|
48 | if fd: | |||
|
49 | fd.close() | |||
|
50 | ||||
|
51 | def _getfile(self, tmpfile, filename, hash): | |||
|
52 | # quit if the largefile isn't there | |||
|
53 | stat = self._stat(hash) | |||
|
54 | if stat: | |||
|
55 | raise util.Abort(_('remotestore: largefile %s is %s') % | |||
|
56 | (hash, stat == 1 and 'invalid' or 'missing')) | |||
|
57 | ||||
|
58 | try: | |||
|
59 | length, infile = self._get(hash) | |||
|
60 | except HTTPError, e: | |||
|
61 | # 401s get converted to util.Aborts; everything else is fine being | |||
|
62 | # turned into a StoreError | |||
|
63 | raise basestore.StoreError(filename, hash, self.url, str(e)) | |||
|
64 | except urllib2.URLError, e: | |||
|
65 | # This usually indicates a connection problem, so don't | |||
|
66 | # keep trying with the other files... they will probably | |||
|
67 | # all fail too. | |||
|
68 | raise util.Abort('%s: %s' % (self.url, str(e.reason))) | |||
|
69 | except IOError, e: | |||
|
70 | raise basestore.StoreError(filename, hash, self.url, str(e)) | |||
|
71 | ||||
|
72 | # Mercurial does not close its SSH connections after writing a stream | |||
|
73 | if length is not None: | |||
|
74 | infile = lfutil.limitreader(infile, length) | |||
|
75 | return lfutil.copyandhash(lfutil.blockstream(infile), tmpfile) | |||
|
76 | ||||
|
77 | def _verify(self, hash): | |||
|
78 | return not self._stat(hash) | |||
|
79 | ||||
|
80 | def _verifyfile(self, cctx, cset, contents, standin, verified): | |||
|
81 | filename = lfutil.splitstandin(standin) | |||
|
82 | if not filename: | |||
|
83 | return False | |||
|
84 | fctx = cctx[standin] | |||
|
85 | key = (filename, fctx.filenode()) | |||
|
86 | if key in verified: | |||
|
87 | return False | |||
|
88 | ||||
|
89 | verified.add(key) | |||
|
90 | ||||
|
91 | stat = self._stat(hash) | |||
|
92 | if not stat: | |||
|
93 | return False | |||
|
94 | elif stat == 1: | |||
|
95 | self.ui.warn( | |||
|
96 | _('changeset %s: %s: contents differ\n') | |||
|
97 | % (cset, filename)) | |||
|
98 | return True # failed | |||
|
99 | elif stat == 2: | |||
|
100 | self.ui.warn( | |||
|
101 | _('changeset %s: %s missing\n') | |||
|
102 | % (cset, filename)) | |||
|
103 | return True # failed | |||
|
104 | else: | |||
|
105 | raise util.Abort(_('check failed, unexpected response' | |||
|
106 | 'statlfile: %d') % stat) |
@@ -0,0 +1,411 b'' | |||||
|
1 | # Copyright 2009-2010 Gregory P. Ward | |||
|
2 | # Copyright 2009-2010 Intelerad Medical Systems Incorporated | |||
|
3 | # Copyright 2010-2011 Fog Creek Software | |||
|
4 | # Copyright 2010-2011 Unity Technologies | |||
|
5 | # | |||
|
6 | # This software may be used and distributed according to the terms of the | |||
|
7 | # GNU General Public License version 2 or any later version. | |||
|
8 | ||||
|
9 | '''setup for largefiles repositories: reposetup''' | |||
|
10 | import copy | |||
|
11 | import types | |||
|
12 | import os | |||
|
13 | import re | |||
|
14 | ||||
|
15 | from mercurial import context, error, manifest, match as match_, \ | |||
|
16 | node, util | |||
|
17 | from mercurial.i18n import _ | |||
|
18 | ||||
|
19 | import lfcommands | |||
|
20 | import proto | |||
|
21 | import lfutil | |||
|
22 | ||||
|
23 | def reposetup(ui, repo): | |||
|
24 | # wire repositories should be given new wireproto functions but not the | |||
|
25 | # other largefiles modifications | |||
|
26 | if not repo.local(): | |||
|
27 | return proto.wirereposetup(ui, repo) | |||
|
28 | ||||
|
29 | for name in ('status', 'commitctx', 'commit', 'push'): | |||
|
30 | method = getattr(repo, name) | |||
|
31 | #if not (isinstance(method, types.MethodType) and | |||
|
32 | # method.im_func is repo.__class__.commitctx.im_func): | |||
|
33 | if isinstance(method, types.FunctionType) and method.func_name == \ | |||
|
34 | 'wrap': | |||
|
35 | ui.warn(_('largefiles: repo method %r appears to have already been' | |||
|
36 | ' wrapped by another extension: ' | |||
|
37 | 'largefiles may behave incorrectly\n') | |||
|
38 | % name) | |||
|
39 | ||||
|
40 | class lfiles_repo(repo.__class__): | |||
|
41 | lfstatus = False | |||
|
42 | def status_nolfiles(self, *args, **kwargs): | |||
|
43 | return super(lfiles_repo, self).status(*args, **kwargs) | |||
|
44 | ||||
|
45 | # When lfstatus is set, return a context that gives the names of lfiles | |||
|
46 | # instead of their corresponding standins and identifies the lfiles as | |||
|
47 | # always binary, regardless of their actual contents. | |||
|
48 | def __getitem__(self, changeid): | |||
|
49 | ctx = super(lfiles_repo, self).__getitem__(changeid) | |||
|
50 | if self.lfstatus: | |||
|
51 | class lfiles_manifestdict(manifest.manifestdict): | |||
|
52 | def __contains__(self, filename): | |||
|
53 | if super(lfiles_manifestdict, | |||
|
54 | self).__contains__(filename): | |||
|
55 | return True | |||
|
56 | return super(lfiles_manifestdict, | |||
|
57 | self).__contains__(lfutil.shortname+'/' + filename) | |||
|
58 | class lfiles_ctx(ctx.__class__): | |||
|
59 | def files(self): | |||
|
60 | filenames = super(lfiles_ctx, self).files() | |||
|
61 | return [re.sub('^\\'+lfutil.shortname+'/', '', filename) for filename | |||
|
62 | in filenames] | |||
|
63 | def manifest(self): | |||
|
64 | man1 = super(lfiles_ctx, self).manifest() | |||
|
65 | man1.__class__ = lfiles_manifestdict | |||
|
66 | return man1 | |||
|
67 | def filectx(self, path, fileid=None, filelog=None): | |||
|
68 | try: | |||
|
69 | result = super(lfiles_ctx, self).filectx(path, | |||
|
70 | fileid, filelog) | |||
|
71 | except error.LookupError: | |||
|
72 | # Adding a null character will cause Mercurial to | |||
|
73 | # identify this as a binary file. | |||
|
74 | result = super(lfiles_ctx, self).filectx( | |||
|
75 | lfutil.shortname + '/' + path, fileid, | |||
|
76 | filelog) | |||
|
77 | olddata = result.data | |||
|
78 | result.data = lambda: olddata() + '\0' | |||
|
79 | return result | |||
|
80 | ctx.__class__ = lfiles_ctx | |||
|
81 | return ctx | |||
|
82 | ||||
|
83 | # Figure out the status of big files and insert them into the | |||
|
84 | # appropriate list in the result. Also removes standin files from | |||
|
85 | # the listing. This function reverts to the original status if | |||
|
86 | # self.lfstatus is False | |||
|
87 | def status(self, node1='.', node2=None, match=None, ignored=False, | |||
|
88 | clean=False, unknown=False, listsubrepos=False): | |||
|
89 | listignored, listclean, listunknown = ignored, clean, unknown | |||
|
90 | if not self.lfstatus: | |||
|
91 | try: | |||
|
92 | return super(lfiles_repo, self).status(node1, node2, match, | |||
|
93 | listignored, listclean, listunknown, listsubrepos) | |||
|
94 | except TypeError: | |||
|
95 | return super(lfiles_repo, self).status(node1, node2, match, | |||
|
96 | listignored, listclean, listunknown) | |||
|
97 | else: | |||
|
98 | # some calls in this function rely on the old version of status | |||
|
99 | self.lfstatus = False | |||
|
100 | if isinstance(node1, context.changectx): | |||
|
101 | ctx1 = node1 | |||
|
102 | else: | |||
|
103 | ctx1 = repo[node1] | |||
|
104 | if isinstance(node2, context.changectx): | |||
|
105 | ctx2 = node2 | |||
|
106 | else: | |||
|
107 | ctx2 = repo[node2] | |||
|
108 | working = ctx2.rev() is None | |||
|
109 | parentworking = working and ctx1 == self['.'] | |||
|
110 | ||||
|
111 | def inctx(file, ctx): | |||
|
112 | try: | |||
|
113 | if ctx.rev() is None: | |||
|
114 | return file in ctx.manifest() | |||
|
115 | ctx[file] | |||
|
116 | return True | |||
|
117 | except: | |||
|
118 | return False | |||
|
119 | ||||
|
120 | # create a copy of match that matches standins instead of | |||
|
121 | # lfiles if matcher not set then it is the always matcher so | |||
|
122 | # overwrite that | |||
|
123 | if match is None: | |||
|
124 | match = match_.always(self.root, self.getcwd()) | |||
|
125 | ||||
|
126 | def tostandin(file): | |||
|
127 | if inctx(lfutil.standin(file), ctx2): | |||
|
128 | return lfutil.standin(file) | |||
|
129 | return file | |||
|
130 | ||||
|
131 | m = copy.copy(match) | |||
|
132 | m._files = [tostandin(f) for f in m._files] | |||
|
133 | ||||
|
134 | # get ignored clean and unknown but remove them later if they | |||
|
135 | # were not asked for | |||
|
136 | try: | |||
|
137 | result = super(lfiles_repo, self).status(node1, node2, m, | |||
|
138 | True, True, True, listsubrepos) | |||
|
139 | except TypeError: | |||
|
140 | result = super(lfiles_repo, self).status(node1, node2, m, | |||
|
141 | True, True, True) | |||
|
142 | if working: | |||
|
143 | # Hold the wlock while we read lfiles and update the | |||
|
144 | # lfdirstate | |||
|
145 | wlock = repo.wlock() | |||
|
146 | try: | |||
|
147 | # Any non lfiles that were explicitly listed must be | |||
|
148 | # taken out or lfdirstate.status will report an error. | |||
|
149 | # The status of these files was already computed using | |||
|
150 | # super's status. | |||
|
151 | lfdirstate = lfutil.openlfdirstate(ui, self) | |||
|
152 | match._files = [f for f in match._files if f in | |||
|
153 | lfdirstate] | |||
|
154 | s = lfdirstate.status(match, [], listignored, | |||
|
155 | listclean, listunknown) | |||
|
156 | (unsure, modified, added, removed, missing, unknown, | |||
|
157 | ignored, clean) = s | |||
|
158 | if parentworking: | |||
|
159 | for lfile in unsure: | |||
|
160 | if ctx1[lfutil.standin(lfile)].data().strip() \ | |||
|
161 | != lfutil.hashfile(self.wjoin(lfile)): | |||
|
162 | modified.append(lfile) | |||
|
163 | else: | |||
|
164 | clean.append(lfile) | |||
|
165 | lfdirstate.normal(lfile) | |||
|
166 | lfdirstate.write() | |||
|
167 | else: | |||
|
168 | tocheck = unsure + modified + added + clean | |||
|
169 | modified, added, clean = [], [], [] | |||
|
170 | ||||
|
171 | for lfile in tocheck: | |||
|
172 | standin = lfutil.standin(lfile) | |||
|
173 | if inctx(standin, ctx1): | |||
|
174 | if ctx1[standin].data().strip() != \ | |||
|
175 | lfutil.hashfile(self.wjoin(lfile)): | |||
|
176 | modified.append(lfile) | |||
|
177 | else: | |||
|
178 | clean.append(lfile) | |||
|
179 | else: | |||
|
180 | added.append(lfile) | |||
|
181 | finally: | |||
|
182 | wlock.release() | |||
|
183 | ||||
|
184 | for standin in ctx1.manifest(): | |||
|
185 | if not lfutil.isstandin(standin): | |||
|
186 | continue | |||
|
187 | lfile = lfutil.splitstandin(standin) | |||
|
188 | if not match(lfile): | |||
|
189 | continue | |||
|
190 | if lfile not in lfdirstate: | |||
|
191 | removed.append(lfile) | |||
|
192 | # Handle unknown and ignored differently | |||
|
193 | lfiles = (modified, added, removed, missing, [], [], clean) | |||
|
194 | result = list(result) | |||
|
195 | # Unknown files | |||
|
196 | result[4] = [f for f in unknown if repo.dirstate[f] == '?'\ | |||
|
197 | and not lfutil.isstandin(f)] | |||
|
198 | # Ignored files must be ignored by both the dirstate and | |||
|
199 | # lfdirstate | |||
|
200 | result[5] = set(ignored).intersection(set(result[5])) | |||
|
201 | # combine normal files and lfiles | |||
|
202 | normals = [[fn for fn in filelist if not \ | |||
|
203 | lfutil.isstandin(fn)] for filelist in result] | |||
|
204 | result = [sorted(list1 + list2) for (list1, list2) in \ | |||
|
205 | zip(normals, lfiles)] | |||
|
206 | else: | |||
|
207 | def toname(f): | |||
|
208 | if lfutil.isstandin(f): | |||
|
209 | return lfutil.splitstandin(f) | |||
|
210 | return f | |||
|
211 | result = [[toname(f) for f in items] for items in result] | |||
|
212 | ||||
|
213 | if not listunknown: | |||
|
214 | result[4] = [] | |||
|
215 | if not listignored: | |||
|
216 | result[5] = [] | |||
|
217 | if not listclean: | |||
|
218 | result[6] = [] | |||
|
219 | self.lfstatus = True | |||
|
220 | return result | |||
|
221 | ||||
|
222 | # This call happens after a commit has occurred. Copy all of the lfiles | |||
|
223 | # into the cache | |||
|
224 | def commitctx(self, *args, **kwargs): | |||
|
225 | node = super(lfiles_repo, self).commitctx(*args, **kwargs) | |||
|
226 | ctx = self[node] | |||
|
227 | for filename in ctx.files(): | |||
|
228 | if lfutil.isstandin(filename) and filename in ctx.manifest(): | |||
|
229 | realfile = lfutil.splitstandin(filename) | |||
|
230 | lfutil.copytocache(self, ctx.node(), realfile) | |||
|
231 | ||||
|
232 | return node | |||
|
233 | ||||
|
234 | # This call happens before a commit has occurred. The lfile standins | |||
|
235 | # have not had their contents updated (to reflect the hash of their | |||
|
236 | # lfile). Do that here. | |||
|
237 | def commit(self, text="", user=None, date=None, match=None, | |||
|
238 | force=False, editor=False, extra={}): | |||
|
239 | orig = super(lfiles_repo, self).commit | |||
|
240 | ||||
|
241 | wlock = repo.wlock() | |||
|
242 | try: | |||
|
243 | if getattr(repo, "_isrebasing", False): | |||
|
244 | # We have to take the time to pull down the new lfiles now. | |||
|
245 | # Otherwise if we are rebasing, any lfiles that were | |||
|
246 | # modified in the changesets we are rebasing on top of get | |||
|
247 | # overwritten either by the rebase or in the first commit | |||
|
248 | # after the rebase. | |||
|
249 | lfcommands.updatelfiles(repo.ui, repo) | |||
|
250 | # Case 1: user calls commit with no specific files or | |||
|
251 | # include/exclude patterns: refresh and commit everything. | |||
|
252 | if (match is None) or (not match.anypats() and not \ | |||
|
253 | match.files()): | |||
|
254 | lfiles = lfutil.listlfiles(self) | |||
|
255 | lfdirstate = lfutil.openlfdirstate(ui, self) | |||
|
256 | # this only loops through lfiles that exist (not | |||
|
257 | # removed/renamed) | |||
|
258 | for lfile in lfiles: | |||
|
259 | if os.path.exists(self.wjoin(lfutil.standin(lfile))): | |||
|
260 | # this handles the case where a rebase is being | |||
|
261 | # performed and the working copy is not updated | |||
|
262 | # yet. | |||
|
263 | if os.path.exists(self.wjoin(lfile)): | |||
|
264 | lfutil.updatestandin(self, | |||
|
265 | lfutil.standin(lfile)) | |||
|
266 | lfdirstate.normal(lfile) | |||
|
267 | for lfile in lfdirstate: | |||
|
268 | if not os.path.exists( | |||
|
269 | repo.wjoin(lfutil.standin(lfile))): | |||
|
270 | try: | |||
|
271 | # Mercurial >= 1.9 | |||
|
272 | lfdirstate.drop(lfile) | |||
|
273 | except AttributeError: | |||
|
274 | # Mercurial <= 1.8 | |||
|
275 | lfdirstate.forget(lfile) | |||
|
276 | lfdirstate.write() | |||
|
277 | ||||
|
278 | return orig(text=text, user=user, date=date, match=match, | |||
|
279 | force=force, editor=editor, extra=extra) | |||
|
280 | ||||
|
281 | for file in match.files(): | |||
|
282 | if lfutil.isstandin(file): | |||
|
283 | raise util.Abort( | |||
|
284 | "Don't commit largefile standin. Commit largefile.") | |||
|
285 | ||||
|
286 | # Case 2: user calls commit with specified patterns: refresh | |||
|
287 | # any matching big files. | |||
|
288 | smatcher = lfutil.composestandinmatcher(self, match) | |||
|
289 | standins = lfutil.dirstate_walk(self.dirstate, smatcher) | |||
|
290 | ||||
|
291 | # No matching big files: get out of the way and pass control to | |||
|
292 | # the usual commit() method. | |||
|
293 | if not standins: | |||
|
294 | return orig(text=text, user=user, date=date, match=match, | |||
|
295 | force=force, editor=editor, extra=extra) | |||
|
296 | ||||
|
297 | # Refresh all matching big files. It's possible that the | |||
|
298 | # commit will end up failing, in which case the big files will | |||
|
299 | # stay refreshed. No harm done: the user modified them and | |||
|
300 | # asked to commit them, so sooner or later we're going to | |||
|
301 | # refresh the standins. Might as well leave them refreshed. | |||
|
302 | lfdirstate = lfutil.openlfdirstate(ui, self) | |||
|
303 | for standin in standins: | |||
|
304 | lfile = lfutil.splitstandin(standin) | |||
|
305 | if lfdirstate[lfile] <> 'r': | |||
|
306 | lfutil.updatestandin(self, standin) | |||
|
307 | lfdirstate.normal(lfile) | |||
|
308 | else: | |||
|
309 | try: | |||
|
310 | # Mercurial >= 1.9 | |||
|
311 | lfdirstate.drop(lfile) | |||
|
312 | except AttributeError: | |||
|
313 | # Mercurial <= 1.8 | |||
|
314 | lfdirstate.forget(lfile) | |||
|
315 | lfdirstate.write() | |||
|
316 | ||||
|
317 | # Cook up a new matcher that only matches regular files or | |||
|
318 | # standins corresponding to the big files requested by the | |||
|
319 | # user. Have to modify _files to prevent commit() from | |||
|
320 | # complaining "not tracked" for big files. | |||
|
321 | lfiles = lfutil.listlfiles(repo) | |||
|
322 | match = copy.copy(match) | |||
|
323 | orig_matchfn = match.matchfn | |||
|
324 | ||||
|
325 | # Check both the list of lfiles and the list of standins | |||
|
326 | # because if a lfile was removed, it won't be in the list of | |||
|
327 | # lfiles at this point | |||
|
328 | match._files += sorted(standins) | |||
|
329 | ||||
|
330 | actualfiles = [] | |||
|
331 | for f in match._files: | |||
|
332 | fstandin = lfutil.standin(f) | |||
|
333 | ||||
|
334 | # Ignore known lfiles and standins | |||
|
335 | if f in lfiles or fstandin in standins: | |||
|
336 | continue | |||
|
337 | ||||
|
338 | # Append directory separator to avoid collisions | |||
|
339 | if not fstandin.endswith(os.sep): | |||
|
340 | fstandin += os.sep | |||
|
341 | ||||
|
342 | # Prevalidate matching standin directories | |||
|
343 | if lfutil.any_(st for st in match._files if \ | |||
|
344 | st.startswith(fstandin)): | |||
|
345 | continue | |||
|
346 | actualfiles.append(f) | |||
|
347 | match._files = actualfiles | |||
|
348 | ||||
|
349 | def matchfn(f): | |||
|
350 | if orig_matchfn(f): | |||
|
351 | return f not in lfiles | |||
|
352 | else: | |||
|
353 | return f in standins | |||
|
354 | ||||
|
355 | match.matchfn = matchfn | |||
|
356 | return orig(text=text, user=user, date=date, match=match, | |||
|
357 | force=force, editor=editor, extra=extra) | |||
|
358 | finally: | |||
|
359 | wlock.release() | |||
|
360 | ||||
|
361 | def push(self, remote, force=False, revs=None, newbranch=False): | |||
|
362 | o = lfutil.findoutgoing(repo, remote, force) | |||
|
363 | if o: | |||
|
364 | toupload = set() | |||
|
365 | o = repo.changelog.nodesbetween(o, revs)[0] | |||
|
366 | for n in o: | |||
|
367 | parents = [p for p in repo.changelog.parents(n) if p != \ | |||
|
368 | node.nullid] | |||
|
369 | ctx = repo[n] | |||
|
370 | files = set(ctx.files()) | |||
|
371 | if len(parents) == 2: | |||
|
372 | mc = ctx.manifest() | |||
|
373 | mp1 = ctx.parents()[0].manifest() | |||
|
374 | mp2 = ctx.parents()[1].manifest() | |||
|
375 | for f in mp1: | |||
|
376 | if f not in mc: | |||
|
377 | files.add(f) | |||
|
378 | for f in mp2: | |||
|
379 | if f not in mc: | |||
|
380 | files.add(f) | |||
|
381 | for f in mc: | |||
|
382 | if mc[f] != mp1.get(f, None) or mc[f] != mp2.get(f, | |||
|
383 | None): | |||
|
384 | files.add(f) | |||
|
385 | ||||
|
386 | toupload = toupload.union(set([ctx[f].data().strip() for f\ | |||
|
387 | in files if lfutil.isstandin(f) and f in ctx])) | |||
|
388 | lfcommands.uploadlfiles(ui, self, remote, toupload) | |||
|
389 | # Mercurial >= 1.6 takes the newbranch argument, try that first. | |||
|
390 | try: | |||
|
391 | return super(lfiles_repo, self).push(remote, force, revs, | |||
|
392 | newbranch) | |||
|
393 | except TypeError: | |||
|
394 | return super(lfiles_repo, self).push(remote, force, revs) | |||
|
395 | ||||
|
396 | repo.__class__ = lfiles_repo | |||
|
397 | ||||
|
398 | def checkrequireslfiles(ui, repo, **kwargs): | |||
|
399 | if 'largefiles' not in repo.requirements and lfutil.any_( | |||
|
400 | lfutil.shortname+'/' in f[0] for f in repo.store.datafiles()): | |||
|
401 | # work around bug in mercurial 1.9 whereby requirements is a list | |||
|
402 | # on newly-cloned repos | |||
|
403 | repo.requirements = set(repo.requirements) | |||
|
404 | ||||
|
405 | repo.requirements |= set(['largefiles']) | |||
|
406 | repo._writerequirements() | |||
|
407 | ||||
|
408 | checkrequireslfiles(ui, repo) | |||
|
409 | ||||
|
410 | ui.setconfig('hooks', 'changegroup.lfiles', checkrequireslfiles) | |||
|
411 | ui.setconfig('hooks', 'commit.lfiles', checkrequireslfiles) |
@@ -0,0 +1,125 b'' | |||||
|
1 | # Copyright 2009-2010 Gregory P. Ward | |||
|
2 | # Copyright 2009-2010 Intelerad Medical Systems Incorporated | |||
|
3 | # Copyright 2010-2011 Fog Creek Software | |||
|
4 | # Copyright 2010-2011 Unity Technologies | |||
|
5 | # | |||
|
6 | # This software may be used and distributed according to the terms of the | |||
|
7 | # GNU General Public License version 2 or any later version. | |||
|
8 | ||||
|
9 | '''setup for largefiles extension: uisetup''' | |||
|
10 | ||||
|
11 | from mercurial import archival, cmdutil, commands, extensions, filemerge, hg, \ | |||
|
12 | httprepo, localrepo, sshrepo, sshserver, wireproto | |||
|
13 | from mercurial.i18n import _ | |||
|
14 | from mercurial.hgweb import hgweb_mod, protocol | |||
|
15 | ||||
|
16 | import overrides | |||
|
17 | import proto | |||
|
18 | ||||
|
19 | def uisetup(ui): | |||
|
20 | # Disable auto-status for some commands which assume that all | |||
|
21 | # files in the result are under Mercurial's control | |||
|
22 | ||||
|
23 | entry = extensions.wrapcommand(commands.table, 'add', overrides.override_add) | |||
|
24 | addopt = [('', 'large', None, _('add as largefile')), | |||
|
25 | ('', 'lfsize', '', _('add all files above this size (in megabytes)' | |||
|
26 | 'as largefiles (default: 10)'))] | |||
|
27 | entry[1].extend(addopt) | |||
|
28 | ||||
|
29 | entry = extensions.wrapcommand(commands.table, 'addremove', | |||
|
30 | overrides.override_addremove) | |||
|
31 | entry = extensions.wrapcommand(commands.table, 'remove', overrides.override_remove) | |||
|
32 | entry = extensions.wrapcommand(commands.table, 'forget', overrides.override_forget) | |||
|
33 | entry = extensions.wrapcommand(commands.table, 'status', overrides.override_status) | |||
|
34 | entry = extensions.wrapcommand(commands.table, 'log', overrides.override_log) | |||
|
35 | entry = extensions.wrapcommand(commands.table, 'rollback', | |||
|
36 | overrides.override_rollback) | |||
|
37 | ||||
|
38 | entry = extensions.wrapcommand(commands.table, 'verify', overrides.override_verify) | |||
|
39 | verifyopt = [('', 'large', None, _('verify largefiles')), | |||
|
40 | ('', 'lfa', None, | |||
|
41 | _('verify all revisions of largefiles not just current')), | |||
|
42 | ('', 'lfc', None, | |||
|
43 | _('verify largefile contents not just existence'))] | |||
|
44 | entry[1].extend(verifyopt) | |||
|
45 | ||||
|
46 | entry = extensions.wrapcommand(commands.table, 'outgoing', | |||
|
47 | overrides.override_outgoing) | |||
|
48 | outgoingopt = [('', 'large', None, _('display outgoing largefiles'))] | |||
|
49 | entry[1].extend(outgoingopt) | |||
|
50 | entry = extensions.wrapcommand(commands.table, 'summary', overrides.override_summary) | |||
|
51 | summaryopt = [('', 'large', None, _('display outgoing largefiles'))] | |||
|
52 | entry[1].extend(summaryopt) | |||
|
53 | ||||
|
54 | entry = extensions.wrapcommand(commands.table, 'update', overrides.override_update) | |||
|
55 | entry = extensions.wrapcommand(commands.table, 'pull', overrides.override_pull) | |||
|
56 | entry = extensions.wrapfunction(filemerge, 'filemerge', overrides.override_filemerge) | |||
|
57 | entry = extensions.wrapfunction(cmdutil, 'copy', overrides.override_copy) | |||
|
58 | ||||
|
59 | # Backout calls revert so we need to override both the command and the | |||
|
60 | # function | |||
|
61 | entry = extensions.wrapcommand(commands.table, 'revert', overrides.override_revert) | |||
|
62 | entry = extensions.wrapfunction(commands, 'revert', overrides.override_revert) | |||
|
63 | ||||
|
64 | # clone uses hg._update instead of hg.update even though they are the | |||
|
65 | # same function... so wrap both of them) | |||
|
66 | extensions.wrapfunction(hg, 'update', overrides.hg_update) | |||
|
67 | extensions.wrapfunction(hg, '_update', overrides.hg_update) | |||
|
68 | extensions.wrapfunction(hg, 'clean', overrides.hg_clean) | |||
|
69 | extensions.wrapfunction(hg, 'merge', overrides.hg_merge) | |||
|
70 | ||||
|
71 | extensions.wrapfunction(archival, 'archive', overrides.override_archive) | |||
|
72 | if hasattr(cmdutil, 'bailifchanged'): | |||
|
73 | extensions.wrapfunction(cmdutil, 'bailifchanged', | |||
|
74 | overrides.override_bailifchanged) | |||
|
75 | else: | |||
|
76 | extensions.wrapfunction(cmdutil, 'bail_if_changed', | |||
|
77 | overrides.override_bailifchanged) | |||
|
78 | ||||
|
79 | # create the new wireproto commands ... | |||
|
80 | wireproto.commands['putlfile'] = (proto.putlfile, 'sha') | |||
|
81 | wireproto.commands['getlfile'] = (proto.getlfile, 'sha') | |||
|
82 | wireproto.commands['statlfile'] = (proto.statlfile, 'sha') | |||
|
83 | ||||
|
84 | # ... and wrap some existing ones | |||
|
85 | wireproto.commands['capabilities'] = (proto.capabilities, '') | |||
|
86 | wireproto.commands['heads'] = (proto.heads, '') | |||
|
87 | wireproto.commands['lheads'] = (wireproto.heads, '') | |||
|
88 | ||||
|
89 | # make putlfile behave the same as push and {get,stat}lfile behave the same | |||
|
90 | # as pull w.r.t. permissions checks | |||
|
91 | hgweb_mod.perms['putlfile'] = 'push' | |||
|
92 | hgweb_mod.perms['getlfile'] = 'pull' | |||
|
93 | hgweb_mod.perms['statlfile'] = 'pull' | |||
|
94 | ||||
|
95 | # the hello wireproto command uses wireproto.capabilities, so it won't see | |||
|
96 | # our largefiles capability unless we replace the actual function as well. | |||
|
97 | proto.capabilities_orig = wireproto.capabilities | |||
|
98 | wireproto.capabilities = proto.capabilities | |||
|
99 | ||||
|
100 | # these let us reject non-lfiles clients and make them display our error | |||
|
101 | # messages | |||
|
102 | protocol.webproto.refuseclient = proto.webproto_refuseclient | |||
|
103 | sshserver.sshserver.refuseclient = proto.sshproto_refuseclient | |||
|
104 | ||||
|
105 | # can't do this in reposetup because it needs to have happened before | |||
|
106 | # wirerepo.__init__ is called | |||
|
107 | proto.ssh_oldcallstream = sshrepo.sshrepository._callstream | |||
|
108 | proto.http_oldcallstream = httprepo.httprepository._callstream | |||
|
109 | sshrepo.sshrepository._callstream = proto.sshrepo_callstream | |||
|
110 | httprepo.httprepository._callstream = proto.httprepo_callstream | |||
|
111 | ||||
|
112 | # don't die on seeing a repo with the largefiles requirement | |||
|
113 | localrepo.localrepository.supported |= set(['largefiles']) | |||
|
114 | ||||
|
115 | # override some extensions' stuff as well | |||
|
116 | for name, module in extensions.extensions(): | |||
|
117 | if name == 'fetch': | |||
|
118 | extensions.wrapcommand(getattr(module, 'cmdtable'), 'fetch', | |||
|
119 | overrides.override_fetch) | |||
|
120 | if name == 'purge': | |||
|
121 | extensions.wrapcommand(getattr(module, 'cmdtable'), 'purge', | |||
|
122 | overrides.override_purge) | |||
|
123 | if name == 'rebase': | |||
|
124 | extensions.wrapcommand(getattr(module, 'cmdtable'), 'rebase', | |||
|
125 | overrides.override_rebase) |
@@ -0,0 +1,51 b'' | |||||
|
1 | Largefiles allows for tracking large, incompressible binary files in Mercurial | |||
|
2 | without requiring excessive bandwidth for clones and pulls. Files added as | |||
|
3 | largefiles are not tracked directly by Mercurial; rather, their revisions are | |||
|
4 | identified by a checksum, and Mercurial tracks these checksums. This way, when | |||
|
5 | you clone a repository or pull in changesets, the large files in older | |||
|
6 | revisions of the repository are not needed, and only the ones needed to update | |||
|
7 | to the current version are downloaded. This saves both disk space and | |||
|
8 | bandwidth. | |||
|
9 | ||||
|
10 | If you are starting a new repository or adding new large binary files, using | |||
|
11 | largefiles for them is as easy as adding '--large' to your hg add command. For | |||
|
12 | example: | |||
|
13 | ||||
|
14 | $ dd if=/dev/urandom of=thisfileislarge count=2000 | |||
|
15 | $ hg add --large thisfileislarge | |||
|
16 | $ hg commit -m 'add thisfileislarge, which is large, as a largefile' | |||
|
17 | ||||
|
18 | When you push a changeset that affects largefiles to a remote repository, its | |||
|
19 | largefile revisions will be uploaded along with it. Note that the remote | |||
|
20 | Mercurial must also have the largefiles extension enabled for this to work. | |||
|
21 | ||||
|
22 | When you pull a changeset that affects largefiles from a remote repository, | |||
|
23 | nothing different from Mercurial's normal behavior happens. However, when you | |||
|
24 | update to such a revision, any largefiles needed by that revision are | |||
|
25 | downloaded and cached if they have never been downloaded before. This means | |||
|
26 | that network access is required to update to revision you have not yet updated | |||
|
27 | to. | |||
|
28 | ||||
|
29 | If you already have large files tracked by Mercurial without the largefiles | |||
|
30 | extension, you will need to convert your repository in order to benefit from | |||
|
31 | largefiles. This is done with the 'hg lfconvert' command: | |||
|
32 | ||||
|
33 | $ hg lfconvert --size 10 oldrepo newrepo | |||
|
34 | ||||
|
35 | By default, in repositories that already have largefiles in them, any new file | |||
|
36 | over 10MB will automatically be added as largefiles. To change this | |||
|
37 | threshhold, set [largefiles].size in your Mercurial config file to the minimum | |||
|
38 | size in megabytes to track as a largefile, or use the --lfsize option to the | |||
|
39 | add command (also in megabytes): | |||
|
40 | ||||
|
41 | [largefiles] | |||
|
42 | size = 2 | |||
|
43 | ||||
|
44 | $ hg add --lfsize 2 | |||
|
45 | ||||
|
46 | The [largefiles].patterns config option allows you to specify specific | |||
|
47 | space-separated filename patterns (in shell glob syntax) that should always be | |||
|
48 | tracked as largefiles: | |||
|
49 | ||||
|
50 | [largefiles] | |||
|
51 | pattens = *.jpg *.{png,bmp} library.zip content/audio/* |
@@ -0,0 +1,29 b'' | |||||
|
1 | # Copyright 2010-2011 Fog Creek Software | |||
|
2 | # | |||
|
3 | # This software may be used and distributed according to the terms of the | |||
|
4 | # GNU General Public License version 2 or any later version. | |||
|
5 | ||||
|
6 | '''largefile store working over mercurial's wire protocol''' | |||
|
7 | ||||
|
8 | import lfutil | |||
|
9 | import remotestore | |||
|
10 | ||||
|
11 | class wirestore(remotestore.remotestore): | |||
|
12 | def __init__(self, ui, repo, remote): | |||
|
13 | cap = remote.capable('largefiles') | |||
|
14 | if not cap: | |||
|
15 | raise lfutil.storeprotonotcapable([]) | |||
|
16 | storetypes = cap.split(',') | |||
|
17 | if not 'serve' in storetypes: | |||
|
18 | raise lfutil.storeprotonotcapable(storetypes) | |||
|
19 | self.remote = remote | |||
|
20 | super(wirestore, self).__init__(ui, repo, remote.url()) | |||
|
21 | ||||
|
22 | def _put(self, hash, fd): | |||
|
23 | return self.remote.putlfile(hash, fd) | |||
|
24 | ||||
|
25 | def _get(self, hash): | |||
|
26 | return self.remote.getlfile(hash) | |||
|
27 | ||||
|
28 | def _stat(self, hash): | |||
|
29 | return self.remote.statlfile(hash) |
General Comments 0
You need to be logged in to leave comments.
Login now