##// END OF EJS Templates
http client: support persistent connections....
Vadim Gelfer -
r2435:ff2bac73 default
parent child Browse files
Show More
This diff has been collapsed as it changes many lines, (587 lines changed) Show them Hide them
@@ -0,0 +1,587 b''
1 # This library is free software; you can redistribute it and/or
2 # modify it under the terms of the GNU Lesser General Public
3 # License as published by the Free Software Foundation; either
4 # version 2.1 of the License, or (at your option) any later version.
5 #
6 # This library is distributed in the hope that it will be useful,
7 # but WITHOUT ANY WARRANTY; without even the implied warranty of
8 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
9 # Lesser General Public License for more details.
10 #
11 # You should have received a copy of the GNU Lesser General Public
12 # License along with this library; if not, write to the
13 # Free Software Foundation, Inc.,
14 # 59 Temple Place, Suite 330,
15 # Boston, MA 02111-1307 USA
16
17 # This file is part of urlgrabber, a high-level cross-protocol url-grabber
18 # Copyright 2002-2004 Michael D. Stenner, Ryan Tomayko
19
20 """An HTTP handler for urllib2 that supports HTTP 1.1 and keepalive.
21
22 >>> import urllib2
23 >>> from keepalive import HTTPHandler
24 >>> keepalive_handler = HTTPHandler()
25 >>> opener = urllib2.build_opener(keepalive_handler)
26 >>> urllib2.install_opener(opener)
27 >>>
28 >>> fo = urllib2.urlopen('http://www.python.org')
29
30 If a connection to a given host is requested, and all of the existing
31 connections are still in use, another connection will be opened. If
32 the handler tries to use an existing connection but it fails in some
33 way, it will be closed and removed from the pool.
34
35 To remove the handler, simply re-run build_opener with no arguments, and
36 install that opener.
37
38 You can explicitly close connections by using the close_connection()
39 method of the returned file-like object (described below) or you can
40 use the handler methods:
41
42 close_connection(host)
43 close_all()
44 open_connections()
45
46 NOTE: using the close_connection and close_all methods of the handler
47 should be done with care when using multiple threads.
48 * there is nothing that prevents another thread from creating new
49 connections immediately after connections are closed
50 * no checks are done to prevent in-use connections from being closed
51
52 >>> keepalive_handler.close_all()
53
54 EXTRA ATTRIBUTES AND METHODS
55
56 Upon a status of 200, the object returned has a few additional
57 attributes and methods, which should not be used if you want to
58 remain consistent with the normal urllib2-returned objects:
59
60 close_connection() - close the connection to the host
61 readlines() - you know, readlines()
62 status - the return status (ie 404)
63 reason - english translation of status (ie 'File not found')
64
65 If you want the best of both worlds, use this inside an
66 AttributeError-catching try:
67
68 >>> try: status = fo.status
69 >>> except AttributeError: status = None
70
71 Unfortunately, these are ONLY there if status == 200, so it's not
72 easy to distinguish between non-200 responses. The reason is that
73 urllib2 tries to do clever things with error codes 301, 302, 401,
74 and 407, and it wraps the object upon return.
75
76 For python versions earlier than 2.4, you can avoid this fancy error
77 handling by setting the module-level global HANDLE_ERRORS to zero.
78 You see, prior to 2.4, it's the HTTP Handler's job to determine what
79 to handle specially, and what to just pass up. HANDLE_ERRORS == 0
80 means "pass everything up". In python 2.4, however, this job no
81 longer belongs to the HTTP Handler and is now done by a NEW handler,
82 HTTPErrorProcessor. Here's the bottom line:
83
84 python version < 2.4
85 HANDLE_ERRORS == 1 (default) pass up 200, treat the rest as
86 errors
87 HANDLE_ERRORS == 0 pass everything up, error processing is
88 left to the calling code
89 python version >= 2.4
90 HANDLE_ERRORS == 1 pass up 200, treat the rest as errors
91 HANDLE_ERRORS == 0 (default) pass everything up, let the
92 other handlers (specifically,
93 HTTPErrorProcessor) decide what to do
94
95 In practice, setting the variable either way makes little difference
96 in python 2.4, so for the most consistent behavior across versions,
97 you probably just want to use the defaults, which will give you
98 exceptions on errors.
99
100 """
101
102 # $Id: keepalive.py,v 1.13 2005/10/22 21:57:28 mstenner Exp $
103
104 import urllib2
105 import httplib
106 import socket
107 import thread
108
109 DEBUG = None
110
111 import sys
112 if sys.version_info < (2, 4): HANDLE_ERRORS = 1
113 else: HANDLE_ERRORS = 0
114
115 class ConnectionManager:
116 """
117 The connection manager must be able to:
118 * keep track of all existing
119 """
120 def __init__(self):
121 self._lock = thread.allocate_lock()
122 self._hostmap = {} # map hosts to a list of connections
123 self._connmap = {} # map connections to host
124 self._readymap = {} # map connection to ready state
125
126 def add(self, host, connection, ready):
127 self._lock.acquire()
128 try:
129 if not self._hostmap.has_key(host): self._hostmap[host] = []
130 self._hostmap[host].append(connection)
131 self._connmap[connection] = host
132 self._readymap[connection] = ready
133 finally:
134 self._lock.release()
135
136 def remove(self, connection):
137 self._lock.acquire()
138 try:
139 try:
140 host = self._connmap[connection]
141 except KeyError:
142 pass
143 else:
144 del self._connmap[connection]
145 del self._readymap[connection]
146 self._hostmap[host].remove(connection)
147 if not self._hostmap[host]: del self._hostmap[host]
148 finally:
149 self._lock.release()
150
151 def set_ready(self, connection, ready):
152 try: self._readymap[connection] = ready
153 except KeyError: pass
154
155 def get_ready_conn(self, host):
156 conn = None
157 self._lock.acquire()
158 try:
159 if self._hostmap.has_key(host):
160 for c in self._hostmap[host]:
161 if self._readymap[c]:
162 self._readymap[c] = 0
163 conn = c
164 break
165 finally:
166 self._lock.release()
167 return conn
168
169 def get_all(self, host=None):
170 if host:
171 return list(self._hostmap.get(host, []))
172 else:
173 return dict(self._hostmap)
174
175 class HTTPHandler(urllib2.HTTPHandler):
176 def __init__(self):
177 self._cm = ConnectionManager()
178
179 #### Connection Management
180 def open_connections(self):
181 """return a list of connected hosts and the number of connections
182 to each. [('foo.com:80', 2), ('bar.org', 1)]"""
183 return [(host, len(li)) for (host, li) in self._cm.get_all().items()]
184
185 def close_connection(self, host):
186 """close connection(s) to <host>
187 host is the host:port spec, as in 'www.cnn.com:8080' as passed in.
188 no error occurs if there is no connection to that host."""
189 for h in self._cm.get_all(host):
190 self._cm.remove(h)
191 h.close()
192
193 def close_all(self):
194 """close all open connections"""
195 for host, conns in self._cm.get_all().items():
196 for h in conns:
197 self._cm.remove(h)
198 h.close()
199
200 def _request_closed(self, request, host, connection):
201 """tells us that this request is now closed and the the
202 connection is ready for another request"""
203 self._cm.set_ready(connection, 1)
204
205 def _remove_connection(self, host, connection, close=0):
206 if close: connection.close()
207 self._cm.remove(connection)
208
209 #### Transaction Execution
210 def http_open(self, req):
211 return self.do_open(HTTPConnection, req)
212
213 def do_open(self, http_class, req):
214 host = req.get_host()
215 if not host:
216 raise urllib2.URLError('no host given')
217
218 try:
219 h = self._cm.get_ready_conn(host)
220 while h:
221 r = self._reuse_connection(h, req, host)
222
223 # if this response is non-None, then it worked and we're
224 # done. Break out, skipping the else block.
225 if r: break
226
227 # connection is bad - possibly closed by server
228 # discard it and ask for the next free connection
229 h.close()
230 self._cm.remove(h)
231 h = self._cm.get_ready_conn(host)
232 else:
233 # no (working) free connections were found. Create a new one.
234 h = http_class(host)
235 if DEBUG: DEBUG.info("creating new connection to %s (%d)",
236 host, id(h))
237 self._cm.add(host, h, 0)
238 self._start_transaction(h, req)
239 r = h.getresponse()
240 except (socket.error, httplib.HTTPException), err:
241 raise urllib2.URLError(err)
242
243 # if not a persistent connection, don't try to reuse it
244 if r.will_close: self._cm.remove(h)
245
246 if DEBUG: DEBUG.info("STATUS: %s, %s", r.status, r.reason)
247 r._handler = self
248 r._host = host
249 r._url = req.get_full_url()
250 r._connection = h
251 r.code = r.status
252
253 if r.status == 200 or not HANDLE_ERRORS:
254 return r
255 else:
256 return self.parent.error('http', req, r, r.status, r.reason, r.msg)
257
258
259 def _reuse_connection(self, h, req, host):
260 """start the transaction with a re-used connection
261 return a response object (r) upon success or None on failure.
262 This DOES not close or remove bad connections in cases where
263 it returns. However, if an unexpected exception occurs, it
264 will close and remove the connection before re-raising.
265 """
266 try:
267 self._start_transaction(h, req)
268 r = h.getresponse()
269 # note: just because we got something back doesn't mean it
270 # worked. We'll check the version below, too.
271 except (socket.error, httplib.HTTPException):
272 r = None
273 except:
274 # adding this block just in case we've missed
275 # something we will still raise the exception, but
276 # lets try and close the connection and remove it
277 # first. We previously got into a nasty loop
278 # where an exception was uncaught, and so the
279 # connection stayed open. On the next try, the
280 # same exception was raised, etc. The tradeoff is
281 # that it's now possible this call will raise
282 # a DIFFERENT exception
283 if DEBUG: DEBUG.error("unexpected exception - closing " + \
284 "connection to %s (%d)", host, id(h))
285 self._cm.remove(h)
286 h.close()
287 raise
288
289 if r is None or r.version == 9:
290 # httplib falls back to assuming HTTP 0.9 if it gets a
291 # bad header back. This is most likely to happen if
292 # the socket has been closed by the server since we
293 # last used the connection.
294 if DEBUG: DEBUG.info("failed to re-use connection to %s (%d)",
295 host, id(h))
296 r = None
297 else:
298 if DEBUG: DEBUG.info("re-using connection to %s (%d)", host, id(h))
299
300 return r
301
302 def _start_transaction(self, h, req):
303 try:
304 if req.has_data():
305 data = req.get_data()
306 h.putrequest('POST', req.get_selector())
307 if not req.headers.has_key('Content-type'):
308 h.putheader('Content-type',
309 'application/x-www-form-urlencoded')
310 if not req.headers.has_key('Content-length'):
311 h.putheader('Content-length', '%d' % len(data))
312 else:
313 h.putrequest('GET', req.get_selector())
314 except (socket.error, httplib.HTTPException), err:
315 raise urllib2.URLError(err)
316
317 for args in self.parent.addheaders:
318 h.putheader(*args)
319 for k, v in req.headers.items():
320 h.putheader(k, v)
321 h.endheaders()
322 if req.has_data():
323 h.send(data)
324
325 class HTTPResponse(httplib.HTTPResponse):
326 # we need to subclass HTTPResponse in order to
327 # 1) add readline() and readlines() methods
328 # 2) add close_connection() methods
329 # 3) add info() and geturl() methods
330
331 # in order to add readline(), read must be modified to deal with a
332 # buffer. example: readline must read a buffer and then spit back
333 # one line at a time. The only real alternative is to read one
334 # BYTE at a time (ick). Once something has been read, it can't be
335 # put back (ok, maybe it can, but that's even uglier than this),
336 # so if you THEN do a normal read, you must first take stuff from
337 # the buffer.
338
339 # the read method wraps the original to accomodate buffering,
340 # although read() never adds to the buffer.
341 # Both readline and readlines have been stolen with almost no
342 # modification from socket.py
343
344
345 def __init__(self, sock, debuglevel=0, strict=0, method=None):
346 if method: # the httplib in python 2.3 uses the method arg
347 httplib.HTTPResponse.__init__(self, sock, debuglevel, method)
348 else: # 2.2 doesn't
349 httplib.HTTPResponse.__init__(self, sock, debuglevel)
350 self.fileno = sock.fileno
351 self.code = None
352 self._rbuf = ''
353 self._rbufsize = 8096
354 self._handler = None # inserted by the handler later
355 self._host = None # (same)
356 self._url = None # (same)
357 self._connection = None # (same)
358
359 _raw_read = httplib.HTTPResponse.read
360
361 def close(self):
362 if self.fp:
363 self.fp.close()
364 self.fp = None
365 if self._handler:
366 self._handler._request_closed(self, self._host,
367 self._connection)
368
369 def close_connection(self):
370 self._handler._remove_connection(self._host, self._connection, close=1)
371 self.close()
372
373 def info(self):
374 return self.msg
375
376 def geturl(self):
377 return self._url
378
379 def read(self, amt=None):
380 # the _rbuf test is only in this first if for speed. It's not
381 # logically necessary
382 if self._rbuf and not amt is None:
383 L = len(self._rbuf)
384 if amt > L:
385 amt -= L
386 else:
387 s = self._rbuf[:amt]
388 self._rbuf = self._rbuf[amt:]
389 return s
390
391 s = self._rbuf + self._raw_read(amt)
392 self._rbuf = ''
393 return s
394
395 def readline(self, limit=-1):
396 data = ""
397 i = self._rbuf.find('\n')
398 while i < 0 and not (0 < limit <= len(self._rbuf)):
399 new = self._raw_read(self._rbufsize)
400 if not new: break
401 i = new.find('\n')
402 if i >= 0: i = i + len(self._rbuf)
403 self._rbuf = self._rbuf + new
404 if i < 0: i = len(self._rbuf)
405 else: i = i+1
406 if 0 <= limit < len(self._rbuf): i = limit
407 data, self._rbuf = self._rbuf[:i], self._rbuf[i:]
408 return data
409
410 def readlines(self, sizehint = 0):
411 total = 0
412 list = []
413 while 1:
414 line = self.readline()
415 if not line: break
416 list.append(line)
417 total += len(line)
418 if sizehint and total >= sizehint:
419 break
420 return list
421
422
423 class HTTPConnection(httplib.HTTPConnection):
424 # use the modified response class
425 response_class = HTTPResponse
426
427 #########################################################################
428 ##### TEST FUNCTIONS
429 #########################################################################
430
431 def error_handler(url):
432 global HANDLE_ERRORS
433 orig = HANDLE_ERRORS
434 keepalive_handler = HTTPHandler()
435 opener = urllib2.build_opener(keepalive_handler)
436 urllib2.install_opener(opener)
437 pos = {0: 'off', 1: 'on'}
438 for i in (0, 1):
439 print " fancy error handling %s (HANDLE_ERRORS = %i)" % (pos[i], i)
440 HANDLE_ERRORS = i
441 try:
442 fo = urllib2.urlopen(url)
443 foo = fo.read()
444 fo.close()
445 try: status, reason = fo.status, fo.reason
446 except AttributeError: status, reason = None, None
447 except IOError, e:
448 print " EXCEPTION: %s" % e
449 raise
450 else:
451 print " status = %s, reason = %s" % (status, reason)
452 HANDLE_ERRORS = orig
453 hosts = keepalive_handler.open_connections()
454 print "open connections:", hosts
455 keepalive_handler.close_all()
456
457 def continuity(url):
458 import md5
459 format = '%25s: %s'
460
461 # first fetch the file with the normal http handler
462 opener = urllib2.build_opener()
463 urllib2.install_opener(opener)
464 fo = urllib2.urlopen(url)
465 foo = fo.read()
466 fo.close()
467 m = md5.new(foo)
468 print format % ('normal urllib', m.hexdigest())
469
470 # now install the keepalive handler and try again
471 opener = urllib2.build_opener(HTTPHandler())
472 urllib2.install_opener(opener)
473
474 fo = urllib2.urlopen(url)
475 foo = fo.read()
476 fo.close()
477 m = md5.new(foo)
478 print format % ('keepalive read', m.hexdigest())
479
480 fo = urllib2.urlopen(url)
481 foo = ''
482 while 1:
483 f = fo.readline()
484 if f: foo = foo + f
485 else: break
486 fo.close()
487 m = md5.new(foo)
488 print format % ('keepalive readline', m.hexdigest())
489
490 def comp(N, url):
491 print ' making %i connections to:\n %s' % (N, url)
492
493 sys.stdout.write(' first using the normal urllib handlers')
494 # first use normal opener
495 opener = urllib2.build_opener()
496 urllib2.install_opener(opener)
497 t1 = fetch(N, url)
498 print ' TIME: %.3f s' % t1
499
500 sys.stdout.write(' now using the keepalive handler ')
501 # now install the keepalive handler and try again
502 opener = urllib2.build_opener(HTTPHandler())
503 urllib2.install_opener(opener)
504 t2 = fetch(N, url)
505 print ' TIME: %.3f s' % t2
506 print ' improvement factor: %.2f' % (t1/t2, )
507
508 def fetch(N, url, delay=0):
509 import time
510 lens = []
511 starttime = time.time()
512 for i in range(N):
513 if delay and i > 0: time.sleep(delay)
514 fo = urllib2.urlopen(url)
515 foo = fo.read()
516 fo.close()
517 lens.append(len(foo))
518 diff = time.time() - starttime
519
520 j = 0
521 for i in lens[1:]:
522 j = j + 1
523 if not i == lens[0]:
524 print "WARNING: inconsistent length on read %i: %i" % (j, i)
525
526 return diff
527
528 def test_timeout(url):
529 global DEBUG
530 dbbackup = DEBUG
531 class FakeLogger:
532 def debug(self, msg, *args): print msg % args
533 info = warning = error = debug
534 DEBUG = FakeLogger()
535 print " fetching the file to establish a connection"
536 fo = urllib2.urlopen(url)
537 data1 = fo.read()
538 fo.close()
539
540 i = 20
541 print " waiting %i seconds for the server to close the connection" % i
542 while i > 0:
543 sys.stdout.write('\r %2i' % i)
544 sys.stdout.flush()
545 time.sleep(1)
546 i -= 1
547 sys.stderr.write('\r')
548
549 print " fetching the file a second time"
550 fo = urllib2.urlopen(url)
551 data2 = fo.read()
552 fo.close()
553
554 if data1 == data2:
555 print ' data are identical'
556 else:
557 print ' ERROR: DATA DIFFER'
558
559 DEBUG = dbbackup
560
561
562 def test(url, N=10):
563 print "checking error hander (do this on a non-200)"
564 try: error_handler(url)
565 except IOError, e:
566 print "exiting - exception will prevent further tests"
567 sys.exit()
568 print
569 print "performing continuity test (making sure stuff isn't corrupted)"
570 continuity(url)
571 print
572 print "performing speed comparison"
573 comp(N, url)
574 print
575 print "performing dropped-connection check"
576 test_timeout(url)
577
578 if __name__ == '__main__':
579 import time
580 import sys
581 try:
582 N = int(sys.argv[1])
583 url = sys.argv[2]
584 except:
585 print "%s <integer> <url>" % sys.argv[0]
586 else:
587 test(url, N)
@@ -1,226 +1,238 b''
1 1 # httprepo.py - HTTP repository proxy classes for mercurial
2 2 #
3 3 # Copyright 2005 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms
6 6 # of the GNU General Public License, incorporated herein by reference.
7 7
8 8 from node import *
9 9 from remoterepo import *
10 10 from i18n import gettext as _
11 11 from demandload import *
12 12 demandload(globals(), "hg os urllib urllib2 urlparse zlib util httplib")
13 demandload(globals(), "keepalive")
13 14
14 15 class passwordmgr(urllib2.HTTPPasswordMgr):
15 16 def __init__(self, ui):
16 17 urllib2.HTTPPasswordMgr.__init__(self)
17 18 self.ui = ui
18 19
19 20 def find_user_password(self, realm, authuri):
20 21 authinfo = urllib2.HTTPPasswordMgr.find_user_password(
21 22 self, realm, authuri)
22 23 if authinfo != (None, None):
23 24 return authinfo
24 25
25 26 if not ui.interactive:
26 27 raise util.Abort(_('http authorization required'))
27 28
28 29 self.ui.write(_("http authorization required\n"))
29 30 self.ui.status(_("realm: %s\n") % realm)
30 31 user = self.ui.prompt(_("user:"), default=None)
31 32 passwd = self.ui.getpass()
32 33
33 34 self.add_password(realm, authuri, user, passwd)
34 35 return (user, passwd)
35 36
36 37 def netlocsplit(netloc):
37 38 '''split [user[:passwd]@]host[:port] into 4-tuple.'''
38 39
39 40 a = netloc.find('@')
40 41 if a == -1:
41 42 user, passwd = None, None
42 43 else:
43 44 userpass, netloc = netloc[:a], netloc[a+1:]
44 45 c = userpass.find(':')
45 46 if c == -1:
46 47 user, passwd = urllib.unquote(userpass), None
47 48 else:
48 49 user = urllib.unquote(userpass[:c])
49 50 passwd = urllib.unquote(userpass[c+1:])
50 51 c = netloc.find(':')
51 52 if c == -1:
52 53 host, port = netloc, None
53 54 else:
54 55 host, port = netloc[:c], netloc[c+1:]
55 56 return host, port, user, passwd
56 57
57 58 def netlocunsplit(host, port, user=None, passwd=None):
58 59 '''turn host, port, user, passwd into [user[:passwd]@]host[:port].'''
59 60 if port:
60 61 hostport = host + ':' + port
61 62 else:
62 63 hostport = host
63 64 if user:
64 65 if passwd:
65 66 userpass = urllib.quote(user) + ':' + urllib.quote(passwd)
66 67 else:
67 68 userpass = urllib.quote(user)
68 69 return userpass + '@' + hostport
69 70 return hostport
70 71
71 72 class httprepository(remoterepository):
72 73 def __init__(self, ui, path):
73 74 scheme, netloc, urlpath, query, frag = urlparse.urlsplit(path)
74 75 if query or frag:
75 76 raise util.Abort(_('unsupported URL component: "%s"') %
76 77 (query or frag))
77 78 if not urlpath: urlpath = '/'
78 79 host, port, user, passwd = netlocsplit(netloc)
79 80
80 81 # urllib cannot handle URLs with embedded user or passwd
81 82 self.url = urlparse.urlunsplit((scheme, netlocunsplit(host, port),
82 83 urlpath, '', ''))
83 84 self.ui = ui
84 85
85 86 proxyurl = ui.config("http_proxy", "host") or os.getenv('http_proxy')
86 87 proxyauthinfo = None
87 handler = urllib2.BaseHandler()
88 handler = keepalive.HTTPHandler()
88 89
89 90 if proxyurl:
90 91 # proxy can be proper url or host[:port]
91 92 if not (proxyurl.startswith('http:') or
92 93 proxyurl.startswith('https:')):
93 94 proxyurl = 'http://' + proxyurl + '/'
94 95 snpqf = urlparse.urlsplit(proxyurl)
95 96 proxyscheme, proxynetloc, proxypath, proxyquery, proxyfrag = snpqf
96 97 hpup = netlocsplit(proxynetloc)
97 98
98 99 proxyhost, proxyport, proxyuser, proxypasswd = hpup
99 100 if not proxyuser:
100 101 proxyuser = ui.config("http_proxy", "user")
101 102 proxypasswd = ui.config("http_proxy", "passwd")
102 103
103 104 # see if we should use a proxy for this url
104 105 no_list = [ "localhost", "127.0.0.1" ]
105 106 no_list.extend([p.strip().lower() for
106 107 p in ui.config("http_proxy", "no", '').split(',')
107 108 if p.strip()])
108 109 no_list.extend([p.strip().lower() for
109 110 p in os.getenv("no_proxy", '').split(',')
110 111 if p.strip()])
111 112 # "http_proxy.always" config is for running tests on localhost
112 113 if (not ui.configbool("http_proxy", "always") and
113 114 host.lower() in no_list):
114 115 ui.debug(_('disabling proxy for %s\n') % host)
115 116 else:
116 117 proxyurl = urlparse.urlunsplit((
117 118 proxyscheme, netlocunsplit(proxyhost, proxyport,
118 119 proxyuser, proxypasswd or ''),
119 120 proxypath, proxyquery, proxyfrag))
120 121 handler = urllib2.ProxyHandler({scheme: proxyurl})
121 122 ui.debug(_('proxying through %s\n') % proxyurl)
122 123
123 124 # urllib2 takes proxy values from the environment and those
124 125 # will take precedence if found, so drop them
125 126 for env in ["HTTP_PROXY", "http_proxy", "no_proxy"]:
126 127 try:
127 128 if os.environ.has_key(env):
128 129 del os.environ[env]
129 130 except OSError:
130 131 pass
131 132
132 133 passmgr = passwordmgr(ui)
133 134 if user:
134 135 ui.debug(_('will use user %s for http auth\n') % user)
135 136 passmgr.add_password(None, host, user, passwd or '')
136 137
137 138 opener = urllib2.build_opener(
138 139 handler,
139 140 urllib2.HTTPBasicAuthHandler(passmgr),
140 141 urllib2.HTTPDigestAuthHandler(passmgr))
141 142
142 143 # 1.0 here is the _protocol_ version
143 144 opener.addheaders = [('User-agent', 'mercurial/proto-1.0')]
144 145 urllib2.install_opener(opener)
145 146
146 147 def dev(self):
147 148 return -1
148 149
149 150 def lock(self):
150 151 raise util.Abort(_('operation not supported over http'))
151 152
152 153 def do_cmd(self, cmd, **args):
153 154 self.ui.debug(_("sending %s command\n") % cmd)
154 155 q = {"cmd": cmd}
155 156 q.update(args)
156 157 qs = urllib.urlencode(q)
157 158 cu = "%s?%s" % (self.url, qs)
158 159 try:
159 160 resp = urllib2.urlopen(cu)
160 161 except httplib.HTTPException, inst:
161 162 self.ui.debug(_('http error while sending %s command\n') % cmd)
162 163 self.ui.print_exc()
163 164 raise IOError(None, inst)
164 proto = resp.headers['content-type']
165 try:
166 proto = resp.getheader('content-type')
167 except AttributeError:
168 proto = resp.headers['content-type']
165 169
166 170 # accept old "text/plain" and "application/hg-changegroup" for now
167 171 if not proto.startswith('application/mercurial') and \
168 172 not proto.startswith('text/plain') and \
169 173 not proto.startswith('application/hg-changegroup'):
170 174 raise hg.RepoError(_("'%s' does not appear to be an hg repository") %
171 175 self.url)
172 176
173 177 if proto.startswith('application/mercurial'):
174 178 version = proto[22:]
175 179 if float(version) > 0.1:
176 180 raise hg.RepoError(_("'%s' uses newer protocol %s") %
177 181 (self.url, version))
178 182
179 183 return resp
180 184
185 def do_read(self, cmd, **args):
186 fp = self.do_cmd(cmd, **args)
187 try:
188 return fp.read()
189 finally:
190 # if using keepalive, allow connection to be reused
191 fp.close()
192
181 193 def heads(self):
182 d = self.do_cmd("heads").read()
194 d = self.do_read("heads")
183 195 try:
184 196 return map(bin, d[:-1].split(" "))
185 197 except:
186 198 self.ui.warn(_("unexpected response:\n") + d[:400] + "\n...\n")
187 199 raise
188 200
189 201 def branches(self, nodes):
190 202 n = " ".join(map(hex, nodes))
191 d = self.do_cmd("branches", nodes=n).read()
203 d = self.do_read("branches", nodes=n)
192 204 try:
193 205 br = [ tuple(map(bin, b.split(" "))) for b in d.splitlines() ]
194 206 return br
195 207 except:
196 208 self.ui.warn(_("unexpected response:\n") + d[:400] + "\n...\n")
197 209 raise
198 210
199 211 def between(self, pairs):
200 212 n = "\n".join(["-".join(map(hex, p)) for p in pairs])
201 d = self.do_cmd("between", pairs=n).read()
213 d = self.do_read("between", pairs=n)
202 214 try:
203 215 p = [ l and map(bin, l.split(" ")) or [] for l in d.splitlines() ]
204 216 return p
205 217 except:
206 218 self.ui.warn(_("unexpected response:\n") + d[:400] + "\n...\n")
207 219 raise
208 220
209 221 def changegroup(self, nodes, kind):
210 222 n = " ".join(map(hex, nodes))
211 223 f = self.do_cmd("changegroup", roots=n)
212 224 bytes = 0
213 225
214 226 def zgenerator(f):
215 227 zd = zlib.decompressobj()
216 228 try:
217 229 for chnk in f:
218 230 yield zd.decompress(chnk)
219 231 except httplib.HTTPException, inst:
220 232 raise IOError(None, _('connection ended unexpectedly'))
221 233 yield zd.flush()
222 234
223 235 return util.chunkbuffer(zgenerator(util.filechunkiter(f)))
224 236
225 237 class httpsrepository(httprepository):
226 238 pass
General Comments 0
You need to be logged in to leave comments. Login now