phabricator.py
1012 lines
| 36.7 KiB
| text/x-python
|
PythonLexer
/ hgext / phabricator.py
Augie Fackler
|
r39688 | # phabricator.py - simple Phabricator integration | ||
# | ||||
# Copyright 2017 Facebook, Inc. | ||||
# | ||||
# This software may be used and distributed according to the terms of the | ||||
# GNU General Public License version 2 or any later version. | ||||
Augie Fackler
|
r39691 | """simple Phabricator integration (EXPERIMENTAL) | ||
Augie Fackler
|
r39688 | |||
This extension provides a ``phabsend`` command which sends a stack of | ||||
changesets to Phabricator, and a ``phabread`` command which prints a stack of | ||||
revisions in a format suitable for :hg:`import`, and a ``phabupdate`` command | ||||
to update statuses in batch. | ||||
By default, Phabricator requires ``Test Plan`` which might prevent some | ||||
changeset from being sent. The requirement could be disabled by changing | ||||
``differential.require-test-plan-field`` config server side. | ||||
Config:: | ||||
[phabricator] | ||||
# Phabricator URL | ||||
url = https://phab.example.com/ | ||||
# Repo callsign. If a repo has a URL https://$HOST/diffusion/FOO, then its | ||||
# callsign is "FOO". | ||||
callsign = FOO | ||||
# curl command to use. If not set (default), use builtin HTTP library to | ||||
# communicate. If set, use the specified curl command. This could be useful | ||||
# if you need to specify advanced options that is not easily supported by | ||||
# the internal library. | ||||
curlcmd = curl --connect-timeout 2 --retry 3 --silent | ||||
[auth] | ||||
example.schemes = https | ||||
example.prefix = phab.example.com | ||||
# API token. Get it from https://$HOST/conduit/login/ | ||||
example.phabtoken = cli-xxxxxxxxxxxxxxxxxxxxxxxxxxxx | ||||
""" | ||||
from __future__ import absolute_import | ||||
Matt Harbison
|
r41111 | import contextlib | ||
Augie Fackler
|
r39688 | import itertools | ||
import json | ||||
import operator | ||||
import re | ||||
from mercurial.node import bin, nullid | ||||
from mercurial.i18n import _ | ||||
from mercurial import ( | ||||
cmdutil, | ||||
context, | ||||
encoding, | ||||
error, | ||||
httpconnection as httpconnectionmod, | ||||
mdiff, | ||||
obsutil, | ||||
parser, | ||||
patch, | ||||
Matt Harbison
|
r41198 | phases, | ||
Augie Fackler
|
r39688 | registrar, | ||
scmutil, | ||||
smartset, | ||||
tags, | ||||
Augie Fackler
|
r39690 | templateutil, | ||
Augie Fackler
|
r39688 | url as urlmod, | ||
util, | ||||
) | ||||
from mercurial.utils import ( | ||||
procutil, | ||||
stringutil, | ||||
) | ||||
Yuya Nishihara
|
r39771 | # Note for extension authors: ONLY specify testedwith = 'ships-with-hg-core' for | ||
# extensions which SHIP WITH MERCURIAL. Non-mainline extensions should | ||||
# be specifying the version(s) of Mercurial they are tested with, or | ||||
# leave the attribute unspecified. | ||||
testedwith = 'ships-with-hg-core' | ||||
Augie Fackler
|
r39688 | cmdtable = {} | ||
command = registrar.command(cmdtable) | ||||
configtable = {} | ||||
configitem = registrar.configitem(configtable) | ||||
# developer config: phabricator.batchsize | ||||
configitem(b'phabricator', b'batchsize', | ||||
default=12, | ||||
) | ||||
configitem(b'phabricator', b'callsign', | ||||
default=None, | ||||
) | ||||
configitem(b'phabricator', b'curlcmd', | ||||
default=None, | ||||
) | ||||
# developer config: phabricator.repophid | ||||
configitem(b'phabricator', b'repophid', | ||||
default=None, | ||||
) | ||||
configitem(b'phabricator', b'url', | ||||
default=None, | ||||
) | ||||
configitem(b'phabsend', b'confirm', | ||||
default=False, | ||||
) | ||||
colortable = { | ||||
b'phabricator.action.created': b'green', | ||||
b'phabricator.action.skipped': b'magenta', | ||||
b'phabricator.action.updated': b'magenta', | ||||
b'phabricator.desc': b'', | ||||
b'phabricator.drev': b'bold', | ||||
b'phabricator.node': b'', | ||||
} | ||||
_VCR_FLAGS = [ | ||||
(b'', b'test-vcr', b'', | ||||
_(b'Path to a vcr file. If nonexistent, will record a new vcr transcript' | ||||
b', otherwise will mock all http requests using the specified vcr file.' | ||||
b' (ADVANCED)' | ||||
)), | ||||
] | ||||
Matt Harbison
|
r41103 | def vcrcommand(name, flags, spec, helpcategory=None): | ||
Augie Fackler
|
r39688 | fullflags = flags + _VCR_FLAGS | ||
def decorate(fn): | ||||
def inner(*args, **kwargs): | ||||
cassette = kwargs.pop(r'test_vcr', None) | ||||
if cassette: | ||||
import hgdemandimport | ||||
with hgdemandimport.deactivated(): | ||||
import vcr as vcrmod | ||||
import vcr.stubs as stubs | ||||
Augie Fackler
|
r40414 | vcr = vcrmod.VCR( | ||
serializer=r'json', | ||||
custom_patches=[ | ||||
(urlmod, 'httpconnection', stubs.VCRHTTPConnection), | ||||
(urlmod, 'httpsconnection', | ||||
stubs.VCRHTTPSConnection), | ||||
]) | ||||
with vcr.use_cassette(cassette): | ||||
return fn(*args, **kwargs) | ||||
Augie Fackler
|
r39688 | return fn(*args, **kwargs) | ||
inner.__name__ = fn.__name__ | ||||
Matt Harbison
|
r40617 | inner.__doc__ = fn.__doc__ | ||
Matt Harbison
|
r41103 | return command(name, fullflags, spec, helpcategory=helpcategory)(inner) | ||
Augie Fackler
|
r39688 | return decorate | ||
def urlencodenested(params): | ||||
"""like urlencode, but works with nested parameters. | ||||
For example, if params is {'a': ['b', 'c'], 'd': {'e': 'f'}}, it will be | ||||
flattened to {'a[0]': 'b', 'a[1]': 'c', 'd[e]': 'f'} and then passed to | ||||
urlencode. Note: the encoding is consistent with PHP's http_build_query. | ||||
""" | ||||
flatparams = util.sortdict() | ||||
def process(prefix, obj): | ||||
Matt Harbison
|
r41073 | if isinstance(obj, bool): | ||
obj = {True: b'true', False: b'false'}[obj] # Python -> PHP form | ||||
Augie Fackler
|
r39688 | items = {list: enumerate, dict: lambda x: x.items()}.get(type(obj)) | ||
if items is None: | ||||
flatparams[prefix] = obj | ||||
else: | ||||
for k, v in items(obj): | ||||
if prefix: | ||||
process(b'%s[%s]' % (prefix, k), v) | ||||
else: | ||||
process(k, v) | ||||
process(b'', params) | ||||
return util.urlreq.urlencode(flatparams) | ||||
def readurltoken(repo): | ||||
"""return conduit url, token and make sure they exist | ||||
Currently read from [auth] config section. In the future, it might | ||||
make sense to read from .arcconfig and .arcrc as well. | ||||
""" | ||||
url = repo.ui.config(b'phabricator', b'url') | ||||
if not url: | ||||
raise error.Abort(_(b'config %s.%s is required') | ||||
% (b'phabricator', b'url')) | ||||
res = httpconnectionmod.readauthforuri(repo.ui, url, util.url(url).user) | ||||
token = None | ||||
if res: | ||||
group, auth = res | ||||
repo.ui.debug(b"using auth.%s.* for authentication\n" % group) | ||||
token = auth.get(b'phabtoken') | ||||
if not token: | ||||
Matt Harbison
|
r40187 | raise error.Abort(_(b'Can\'t find conduit token associated to %s') | ||
% (url,)) | ||||
Augie Fackler
|
r39688 | |||
return url, token | ||||
def callconduit(repo, name, params): | ||||
"""call Conduit API, params is a dict. return json.loads result, or None""" | ||||
host, token = readurltoken(repo) | ||||
url, authinfo = util.url(b'/'.join([host, b'api', name])).authinfo() | ||||
repo.ui.debug(b'Conduit Call: %s %s\n' % (url, params)) | ||||
params = params.copy() | ||||
params[b'api.token'] = token | ||||
data = urlencodenested(params) | ||||
curlcmd = repo.ui.config(b'phabricator', b'curlcmd') | ||||
if curlcmd: | ||||
sin, sout = procutil.popen2(b'%s -d @- %s' | ||||
% (curlcmd, procutil.shellquote(url))) | ||||
sin.write(data) | ||||
sin.close() | ||||
body = sout.read() | ||||
else: | ||||
urlopener = urlmod.opener(repo.ui, authinfo) | ||||
request = util.urlreq.request(url, data=data) | ||||
Matt Harbison
|
r41111 | with contextlib.closing(urlopener.open(request)) as rsp: | ||
body = rsp.read() | ||||
Augie Fackler
|
r39688 | repo.ui.debug(b'Conduit Response: %s\n' % body) | ||
parsed = json.loads(body) | ||||
if parsed.get(r'error_code'): | ||||
msg = (_(b'Conduit Error (%s): %s') | ||||
% (parsed[r'error_code'], parsed[r'error_info'])) | ||||
raise error.Abort(msg) | ||||
return parsed[r'result'] | ||||
@vcrcommand(b'debugcallconduit', [], _(b'METHOD')) | ||||
def debugcallconduit(ui, repo, name): | ||||
"""call Conduit API | ||||
Call parameters are read from stdin as a JSON blob. Result will be written | ||||
to stdout as a JSON blob. | ||||
""" | ||||
params = json.loads(ui.fin.read()) | ||||
result = callconduit(repo, name, params) | ||||
s = json.dumps(result, sort_keys=True, indent=2, separators=(b',', b': ')) | ||||
ui.write(b'%s\n' % s) | ||||
def getrepophid(repo): | ||||
"""given callsign, return repository PHID or None""" | ||||
# developer config: phabricator.repophid | ||||
repophid = repo.ui.config(b'phabricator', b'repophid') | ||||
if repophid: | ||||
return repophid | ||||
callsign = repo.ui.config(b'phabricator', b'callsign') | ||||
if not callsign: | ||||
return None | ||||
query = callconduit(repo, b'diffusion.repository.search', | ||||
{b'constraints': {b'callsigns': [callsign]}}) | ||||
if len(query[r'data']) == 0: | ||||
return None | ||||
repophid = encoding.strtolocal(query[r'data'][0][r'phid']) | ||||
repo.ui.setconfig(b'phabricator', b'repophid', repophid) | ||||
return repophid | ||||
_differentialrevisiontagre = re.compile(b'\AD([1-9][0-9]*)\Z') | ||||
_differentialrevisiondescre = re.compile( | ||||
b'^Differential Revision:\s*(?P<url>(?:.*)D(?P<id>[1-9][0-9]*))$', re.M) | ||||
def getoldnodedrevmap(repo, nodelist): | ||||
"""find previous nodes that has been sent to Phabricator | ||||
return {node: (oldnode, Differential diff, Differential Revision ID)} | ||||
for node in nodelist with known previous sent versions, or associated | ||||
Differential Revision IDs. ``oldnode`` and ``Differential diff`` could | ||||
be ``None``. | ||||
Examines commit messages like "Differential Revision:" to get the | ||||
association information. | ||||
If such commit message line is not found, examines all precursors and their | ||||
tags. Tags with format like "D1234" are considered a match and the node | ||||
with that tag, and the number after "D" (ex. 1234) will be returned. | ||||
The ``old node``, if not None, is guaranteed to be the last diff of | ||||
corresponding Differential Revision, and exist in the repo. | ||||
""" | ||||
unfi = repo.unfiltered() | ||||
nodemap = unfi.changelog.nodemap | ||||
result = {} # {node: (oldnode?, lastdiff?, drev)} | ||||
toconfirm = {} # {node: (force, {precnode}, drev)} | ||||
for node in nodelist: | ||||
ctx = unfi[node] | ||||
# For tags like "D123", put them into "toconfirm" to verify later | ||||
precnodes = list(obsutil.allpredecessors(unfi.obsstore, [node])) | ||||
for n in precnodes: | ||||
if n in nodemap: | ||||
for tag in unfi.nodetags(n): | ||||
m = _differentialrevisiontagre.match(tag) | ||||
if m: | ||||
toconfirm[node] = (0, set(precnodes), int(m.group(1))) | ||||
continue | ||||
# Check commit message | ||||
m = _differentialrevisiondescre.search(ctx.description()) | ||||
if m: | ||||
toconfirm[node] = (1, set(precnodes), int(m.group(b'id'))) | ||||
# Double check if tags are genuine by collecting all old nodes from | ||||
# Phabricator, and expect precursors overlap with it. | ||||
if toconfirm: | ||||
drevs = [drev for force, precs, drev in toconfirm.values()] | ||||
alldiffs = callconduit(unfi, b'differential.querydiffs', | ||||
{b'revisionIDs': drevs}) | ||||
getnode = lambda d: bin(encoding.unitolocal( | ||||
getdiffmeta(d).get(r'node', b''))) or None | ||||
for newnode, (force, precset, drev) in toconfirm.items(): | ||||
diffs = [d for d in alldiffs.values() | ||||
if int(d[r'revisionID']) == drev] | ||||
# "precursors" as known by Phabricator | ||||
phprecset = set(getnode(d) for d in diffs) | ||||
# Ignore if precursors (Phabricator and local repo) do not overlap, | ||||
# and force is not set (when commit message says nothing) | ||||
if not force and not bool(phprecset & precset): | ||||
tagname = b'D%d' % drev | ||||
tags.tag(repo, tagname, nullid, message=None, user=None, | ||||
date=None, local=True) | ||||
unfi.ui.warn(_(b'D%s: local tag removed - does not match ' | ||||
b'Differential history\n') % drev) | ||||
continue | ||||
# Find the last node using Phabricator metadata, and make sure it | ||||
# exists in the repo | ||||
oldnode = lastdiff = None | ||||
if diffs: | ||||
lastdiff = max(diffs, key=lambda d: int(d[r'id'])) | ||||
oldnode = getnode(lastdiff) | ||||
if oldnode and oldnode not in nodemap: | ||||
oldnode = None | ||||
result[newnode] = (oldnode, lastdiff, drev) | ||||
return result | ||||
def getdiff(ctx, diffopts): | ||||
"""plain-text diff without header (user, commit message, etc)""" | ||||
output = util.stringio() | ||||
for chunk, _label in patch.diffui(ctx.repo(), ctx.p1().node(), ctx.node(), | ||||
None, opts=diffopts): | ||||
output.write(chunk) | ||||
return output.getvalue() | ||||
def creatediff(ctx): | ||||
"""create a Differential Diff""" | ||||
repo = ctx.repo() | ||||
repophid = getrepophid(repo) | ||||
# Create a "Differential Diff" via "differential.createrawdiff" API | ||||
params = {b'diff': getdiff(ctx, mdiff.diffopts(git=True, context=32767))} | ||||
if repophid: | ||||
params[b'repositoryPHID'] = repophid | ||||
diff = callconduit(repo, b'differential.createrawdiff', params) | ||||
if not diff: | ||||
raise error.Abort(_(b'cannot create diff for %s') % ctx) | ||||
return diff | ||||
def writediffproperties(ctx, diff): | ||||
"""write metadata to diff so patches could be applied losslessly""" | ||||
params = { | ||||
b'diff_id': diff[r'id'], | ||||
b'name': b'hg:meta', | ||||
b'data': json.dumps({ | ||||
b'user': ctx.user(), | ||||
b'date': b'%d %d' % ctx.date(), | ||||
b'node': ctx.hex(), | ||||
b'parent': ctx.p1().hex(), | ||||
}), | ||||
} | ||||
callconduit(ctx.repo(), b'differential.setdiffproperty', params) | ||||
params = { | ||||
b'diff_id': diff[r'id'], | ||||
b'name': b'local:commits', | ||||
b'data': json.dumps({ | ||||
ctx.hex(): { | ||||
b'author': stringutil.person(ctx.user()), | ||||
b'authorEmail': stringutil.email(ctx.user()), | ||||
b'time': ctx.date()[0], | ||||
}, | ||||
}), | ||||
} | ||||
callconduit(ctx.repo(), b'differential.setdiffproperty', params) | ||||
def createdifferentialrevision(ctx, revid=None, parentrevid=None, oldnode=None, | ||||
olddiff=None, actions=None): | ||||
"""create or update a Differential Revision | ||||
If revid is None, create a new Differential Revision, otherwise update | ||||
revid. If parentrevid is not None, set it as a dependency. | ||||
If oldnode is not None, check if the patch content (without commit message | ||||
and metadata) has changed before creating another diff. | ||||
If actions is not None, they will be appended to the transaction. | ||||
""" | ||||
repo = ctx.repo() | ||||
if oldnode: | ||||
diffopts = mdiff.diffopts(git=True, context=32767) | ||||
oldctx = repo.unfiltered()[oldnode] | ||||
neednewdiff = (getdiff(ctx, diffopts) != getdiff(oldctx, diffopts)) | ||||
else: | ||||
neednewdiff = True | ||||
transactions = [] | ||||
if neednewdiff: | ||||
diff = creatediff(ctx) | ||||
transactions.append({b'type': b'update', b'value': diff[r'phid']}) | ||||
else: | ||||
# Even if we don't need to upload a new diff because the patch content | ||||
# does not change. We might still need to update its metadata so | ||||
# pushers could know the correct node metadata. | ||||
assert olddiff | ||||
diff = olddiff | ||||
writediffproperties(ctx, diff) | ||||
# Use a temporary summary to set dependency. There might be better ways but | ||||
# I cannot find them for now. But do not do that if we are updating an | ||||
# existing revision (revid is not None) since that introduces visible | ||||
# churns (someone edited "Summary" twice) on the web page. | ||||
if parentrevid and revid is None: | ||||
summary = b'Depends on D%s' % parentrevid | ||||
transactions += [{b'type': b'summary', b'value': summary}, | ||||
{b'type': b'summary', b'value': b' '}] | ||||
if actions: | ||||
transactions += actions | ||||
# Parse commit message and update related fields. | ||||
desc = ctx.description() | ||||
info = callconduit(repo, b'differential.parsecommitmessage', | ||||
{b'corpus': desc}) | ||||
for k, v in info[r'fields'].items(): | ||||
if k in [b'title', b'summary', b'testPlan']: | ||||
transactions.append({b'type': k, b'value': v}) | ||||
params = {b'transactions': transactions} | ||||
if revid is not None: | ||||
# Update an existing Differential Revision | ||||
params[b'objectIdentifier'] = revid | ||||
revision = callconduit(repo, b'differential.revision.edit', params) | ||||
if not revision: | ||||
raise error.Abort(_(b'cannot create revision for %s') % ctx) | ||||
return revision, diff | ||||
def userphids(repo, names): | ||||
"""convert user names to PHIDs""" | ||||
query = {b'constraints': {b'usernames': names}} | ||||
result = callconduit(repo, b'user.search', query) | ||||
# username not found is not an error of the API. So check if we have missed | ||||
# some names here. | ||||
data = result[r'data'] | ||||
resolved = set(entry[r'fields'][r'username'] for entry in data) | ||||
unresolved = set(names) - resolved | ||||
if unresolved: | ||||
raise error.Abort(_(b'unknown username: %s') | ||||
% b' '.join(sorted(unresolved))) | ||||
return [entry[r'phid'] for entry in data] | ||||
@vcrcommand(b'phabsend', | ||||
[(b'r', b'rev', [], _(b'revisions to send'), _(b'REV')), | ||||
(b'', b'amend', True, _(b'update commit messages')), | ||||
(b'', b'reviewer', [], _(b'specify reviewers')), | ||||
(b'', b'confirm', None, _(b'ask for confirmation before sending'))], | ||||
Matt Harbison
|
r41103 | _(b'REV [OPTIONS]'), | ||
helpcategory=command.CATEGORY_IMPORT_EXPORT) | ||||
Augie Fackler
|
r39688 | def phabsend(ui, repo, *revs, **opts): | ||
"""upload changesets to Phabricator | ||||
If there are multiple revisions specified, they will be send as a stack | ||||
with a linear dependencies relationship using the order specified by the | ||||
revset. | ||||
For the first time uploading changesets, local tags will be created to | ||||
maintain the association. After the first time, phabsend will check | ||||
obsstore and tags information so it can figure out whether to update an | ||||
existing Differential Revision, or create a new one. | ||||
If --amend is set, update commit messages so they have the | ||||
``Differential Revision`` URL, remove related tags. This is similar to what | ||||
arcanist will do, and is more desired in author-push workflows. Otherwise, | ||||
use local tags to record the ``Differential Revision`` association. | ||||
The --confirm option lets you confirm changesets before sending them. You | ||||
can also add following to your configuration file to make it default | ||||
behaviour:: | ||||
[phabsend] | ||||
confirm = true | ||||
phabsend will check obsstore and the above association to decide whether to | ||||
update an existing Differential Revision, or create a new one. | ||||
""" | ||||
revs = list(revs) + opts.get(b'rev', []) | ||||
revs = scmutil.revrange(repo, revs) | ||||
if not revs: | ||||
raise error.Abort(_(b'phabsend requires at least one changeset')) | ||||
if opts.get(b'amend'): | ||||
cmdutil.checkunfinished(repo) | ||||
# {newnode: (oldnode, olddiff, olddrev} | ||||
oldmap = getoldnodedrevmap(repo, [repo[r].node() for r in revs]) | ||||
confirm = ui.configbool(b'phabsend', b'confirm') | ||||
confirm |= bool(opts.get(b'confirm')) | ||||
if confirm: | ||||
confirmed = _confirmbeforesend(repo, revs, oldmap) | ||||
if not confirmed: | ||||
raise error.Abort(_(b'phabsend cancelled')) | ||||
actions = [] | ||||
reviewers = opts.get(b'reviewer', []) | ||||
if reviewers: | ||||
phids = userphids(repo, reviewers) | ||||
actions.append({b'type': b'reviewers.add', b'value': phids}) | ||||
drevids = [] # [int] | ||||
diffmap = {} # {newnode: diff} | ||||
# Send patches one by one so we know their Differential Revision IDs and | ||||
# can provide dependency relationship | ||||
lastrevid = None | ||||
for rev in revs: | ||||
ui.debug(b'sending rev %d\n' % rev) | ||||
ctx = repo[rev] | ||||
# Get Differential Revision ID | ||||
oldnode, olddiff, revid = oldmap.get(ctx.node(), (None, None, None)) | ||||
if oldnode != ctx.node() or opts.get(b'amend'): | ||||
# Create or update Differential Revision | ||||
revision, diff = createdifferentialrevision( | ||||
ctx, revid, lastrevid, oldnode, olddiff, actions) | ||||
diffmap[ctx.node()] = diff | ||||
newrevid = int(revision[r'object'][r'id']) | ||||
if revid: | ||||
action = b'updated' | ||||
else: | ||||
action = b'created' | ||||
# Create a local tag to note the association, if commit message | ||||
# does not have it already | ||||
m = _differentialrevisiondescre.search(ctx.description()) | ||||
if not m or int(m.group(b'id')) != newrevid: | ||||
tagname = b'D%d' % newrevid | ||||
tags.tag(repo, tagname, ctx.node(), message=None, user=None, | ||||
date=None, local=True) | ||||
else: | ||||
# Nothing changed. But still set "newrevid" so the next revision | ||||
# could depend on this one. | ||||
newrevid = revid | ||||
action = b'skipped' | ||||
actiondesc = ui.label( | ||||
{b'created': _(b'created'), | ||||
b'skipped': _(b'skipped'), | ||||
b'updated': _(b'updated')}[action], | ||||
b'phabricator.action.%s' % action) | ||||
drevdesc = ui.label(b'D%s' % newrevid, b'phabricator.drev') | ||||
nodedesc = ui.label(bytes(ctx), b'phabricator.node') | ||||
desc = ui.label(ctx.description().split(b'\n')[0], b'phabricator.desc') | ||||
ui.write(_(b'%s - %s - %s: %s\n') % (drevdesc, actiondesc, nodedesc, | ||||
desc)) | ||||
drevids.append(newrevid) | ||||
lastrevid = newrevid | ||||
# Update commit messages and remove tags | ||||
if opts.get(b'amend'): | ||||
unfi = repo.unfiltered() | ||||
drevs = callconduit(repo, b'differential.query', {b'ids': drevids}) | ||||
with repo.wlock(), repo.lock(), repo.transaction(b'phabsend'): | ||||
wnode = unfi[b'.'].node() | ||||
mapping = {} # {oldnode: [newnode]} | ||||
for i, rev in enumerate(revs): | ||||
old = unfi[rev] | ||||
drevid = drevids[i] | ||||
drev = [d for d in drevs if int(d[r'id']) == drevid][0] | ||||
newdesc = getdescfromdrev(drev) | ||||
newdesc = encoding.unitolocal(newdesc) | ||||
# Make sure commit message contain "Differential Revision" | ||||
if old.description() != newdesc: | ||||
Matt Harbison
|
r41198 | if old.phase() == phases.public: | ||
ui.warn(_("warning: not updating public commit %s\n") | ||||
% scmutil.formatchangeid(old)) | ||||
continue | ||||
Augie Fackler
|
r39688 | parents = [ | ||
mapping.get(old.p1().node(), (old.p1(),))[0], | ||||
mapping.get(old.p2().node(), (old.p2(),))[0], | ||||
] | ||||
new = context.metadataonlyctx( | ||||
repo, old, parents=parents, text=newdesc, | ||||
user=old.user(), date=old.date(), extra=old.extra()) | ||||
newnode = new.commit() | ||||
mapping[old.node()] = [newnode] | ||||
# Update diff property | ||||
writediffproperties(unfi[newnode], diffmap[old.node()]) | ||||
# Remove local tags since it's no longer necessary | ||||
tagname = b'D%d' % drevid | ||||
if tagname in repo.tags(): | ||||
tags.tag(repo, tagname, nullid, message=None, user=None, | ||||
date=None, local=True) | ||||
scmutil.cleanupnodes(repo, mapping, b'phabsend', fixphase=True) | ||||
if wnode in mapping: | ||||
unfi.setparents(mapping[wnode][0]) | ||||
# Map from "hg:meta" keys to header understood by "hg import". The order is | ||||
# consistent with "hg export" output. | ||||
_metanamemap = util.sortdict([(r'user', b'User'), (r'date', b'Date'), | ||||
(r'node', b'Node ID'), (r'parent', b'Parent ')]) | ||||
def _confirmbeforesend(repo, revs, oldmap): | ||||
url, token = readurltoken(repo) | ||||
ui = repo.ui | ||||
for rev in revs: | ||||
ctx = repo[rev] | ||||
desc = ctx.description().splitlines()[0] | ||||
oldnode, olddiff, drevid = oldmap.get(ctx.node(), (None, None, None)) | ||||
if drevid: | ||||
drevdesc = ui.label(b'D%s' % drevid, b'phabricator.drev') | ||||
else: | ||||
drevdesc = ui.label(_(b'NEW'), b'phabricator.drev') | ||||
ui.write(_(b'%s - %s: %s\n') | ||||
% (drevdesc, | ||||
ui.label(bytes(ctx), b'phabricator.node'), | ||||
ui.label(desc, b'phabricator.desc'))) | ||||
if ui.promptchoice(_(b'Send the above changes to %s (yn)?' | ||||
b'$$ &Yes $$ &No') % url): | ||||
return False | ||||
return True | ||||
_knownstatusnames = {b'accepted', b'needsreview', b'needsrevision', b'closed', | ||||
b'abandoned'} | ||||
def _getstatusname(drev): | ||||
"""get normalized status name from a Differential Revision""" | ||||
return drev[r'statusName'].replace(b' ', b'').lower() | ||||
# Small language to specify differential revisions. Support symbols: (), :X, | ||||
# +, and -. | ||||
_elements = { | ||||
# token-type: binding-strength, primary, prefix, infix, suffix | ||||
b'(': (12, None, (b'group', 1, b')'), None, None), | ||||
b':': (8, None, (b'ancestors', 8), None, None), | ||||
b'&': (5, None, None, (b'and_', 5), None), | ||||
b'+': (4, None, None, (b'add', 4), None), | ||||
b'-': (4, None, None, (b'sub', 4), None), | ||||
b')': (0, None, None, None, None), | ||||
b'symbol': (0, b'symbol', None, None, None), | ||||
b'end': (0, None, None, None, None), | ||||
} | ||||
def _tokenize(text): | ||||
view = memoryview(text) # zero-copy slice | ||||
special = b'():+-& ' | ||||
pos = 0 | ||||
length = len(text) | ||||
while pos < length: | ||||
symbol = b''.join(itertools.takewhile(lambda ch: ch not in special, | ||||
view[pos:])) | ||||
if symbol: | ||||
yield (b'symbol', symbol, pos) | ||||
pos += len(symbol) | ||||
else: # special char, ignore space | ||||
if text[pos] != b' ': | ||||
yield (text[pos], None, pos) | ||||
pos += 1 | ||||
yield (b'end', None, pos) | ||||
def _parse(text): | ||||
tree, pos = parser.parser(_elements).parse(_tokenize(text)) | ||||
if pos != len(text): | ||||
raise error.ParseError(b'invalid token', pos) | ||||
return tree | ||||
def _parsedrev(symbol): | ||||
"""str -> int or None, ex. 'D45' -> 45; '12' -> 12; 'x' -> None""" | ||||
if symbol.startswith(b'D') and symbol[1:].isdigit(): | ||||
return int(symbol[1:]) | ||||
if symbol.isdigit(): | ||||
return int(symbol) | ||||
def _prefetchdrevs(tree): | ||||
"""return ({single-drev-id}, {ancestor-drev-id}) to prefetch""" | ||||
drevs = set() | ||||
ancestordrevs = set() | ||||
op = tree[0] | ||||
if op == b'symbol': | ||||
r = _parsedrev(tree[1]) | ||||
if r: | ||||
drevs.add(r) | ||||
elif op == b'ancestors': | ||||
r, a = _prefetchdrevs(tree[1]) | ||||
drevs.update(r) | ||||
ancestordrevs.update(r) | ||||
ancestordrevs.update(a) | ||||
else: | ||||
for t in tree[1:]: | ||||
r, a = _prefetchdrevs(t) | ||||
drevs.update(r) | ||||
ancestordrevs.update(a) | ||||
return drevs, ancestordrevs | ||||
def querydrev(repo, spec): | ||||
"""return a list of "Differential Revision" dicts | ||||
spec is a string using a simple query language, see docstring in phabread | ||||
for details. | ||||
A "Differential Revision dict" looks like: | ||||
{ | ||||
"id": "2", | ||||
"phid": "PHID-DREV-672qvysjcczopag46qty", | ||||
"title": "example", | ||||
"uri": "https://phab.example.com/D2", | ||||
"dateCreated": "1499181406", | ||||
"dateModified": "1499182103", | ||||
"authorPHID": "PHID-USER-tv3ohwc4v4jeu34otlye", | ||||
"status": "0", | ||||
"statusName": "Needs Review", | ||||
"properties": [], | ||||
"branch": null, | ||||
"summary": "", | ||||
"testPlan": "", | ||||
"lineCount": "2", | ||||
"activeDiffPHID": "PHID-DIFF-xoqnjkobbm6k4dk6hi72", | ||||
"diffs": [ | ||||
"3", | ||||
"4", | ||||
], | ||||
"commits": [], | ||||
"reviewers": [], | ||||
"ccs": [], | ||||
"hashes": [], | ||||
"auxiliary": { | ||||
"phabricator:projects": [], | ||||
"phabricator:depends-on": [ | ||||
"PHID-DREV-gbapp366kutjebt7agcd" | ||||
] | ||||
}, | ||||
"repositoryPHID": "PHID-REPO-hub2hx62ieuqeheznasv", | ||||
"sourcePath": null | ||||
} | ||||
""" | ||||
def fetch(params): | ||||
"""params -> single drev or None""" | ||||
key = (params.get(r'ids') or params.get(r'phids') or [None])[0] | ||||
if key in prefetched: | ||||
return prefetched[key] | ||||
drevs = callconduit(repo, b'differential.query', params) | ||||
# Fill prefetched with the result | ||||
for drev in drevs: | ||||
prefetched[drev[r'phid']] = drev | ||||
prefetched[int(drev[r'id'])] = drev | ||||
if key not in prefetched: | ||||
raise error.Abort(_(b'cannot get Differential Revision %r') | ||||
% params) | ||||
return prefetched[key] | ||||
def getstack(topdrevids): | ||||
"""given a top, get a stack from the bottom, [id] -> [id]""" | ||||
visited = set() | ||||
result = [] | ||||
queue = [{r'ids': [i]} for i in topdrevids] | ||||
while queue: | ||||
params = queue.pop() | ||||
drev = fetch(params) | ||||
if drev[r'id'] in visited: | ||||
continue | ||||
visited.add(drev[r'id']) | ||||
result.append(int(drev[r'id'])) | ||||
auxiliary = drev.get(r'auxiliary', {}) | ||||
depends = auxiliary.get(r'phabricator:depends-on', []) | ||||
for phid in depends: | ||||
queue.append({b'phids': [phid]}) | ||||
result.reverse() | ||||
return smartset.baseset(result) | ||||
# Initialize prefetch cache | ||||
prefetched = {} # {id or phid: drev} | ||||
tree = _parse(spec) | ||||
drevs, ancestordrevs = _prefetchdrevs(tree) | ||||
# developer config: phabricator.batchsize | ||||
batchsize = repo.ui.configint(b'phabricator', b'batchsize') | ||||
# Prefetch Differential Revisions in batch | ||||
tofetch = set(drevs) | ||||
for r in ancestordrevs: | ||||
tofetch.update(range(max(1, r - batchsize), r + 1)) | ||||
if drevs: | ||||
fetch({r'ids': list(tofetch)}) | ||||
validids = sorted(set(getstack(list(ancestordrevs))) | set(drevs)) | ||||
# Walk through the tree, return smartsets | ||||
def walk(tree): | ||||
op = tree[0] | ||||
if op == b'symbol': | ||||
drev = _parsedrev(tree[1]) | ||||
if drev: | ||||
return smartset.baseset([drev]) | ||||
elif tree[1] in _knownstatusnames: | ||||
drevs = [r for r in validids | ||||
if _getstatusname(prefetched[r]) == tree[1]] | ||||
return smartset.baseset(drevs) | ||||
else: | ||||
raise error.Abort(_(b'unknown symbol: %s') % tree[1]) | ||||
elif op in {b'and_', b'add', b'sub'}: | ||||
assert len(tree) == 3 | ||||
return getattr(operator, op)(walk(tree[1]), walk(tree[2])) | ||||
elif op == b'group': | ||||
return walk(tree[1]) | ||||
elif op == b'ancestors': | ||||
return getstack(walk(tree[1])) | ||||
else: | ||||
raise error.ProgrammingError(b'illegal tree: %r' % tree) | ||||
return [prefetched[r] for r in walk(tree)] | ||||
def getdescfromdrev(drev): | ||||
"""get description (commit message) from "Differential Revision" | ||||
This is similar to differential.getcommitmessage API. But we only care | ||||
about limited fields: title, summary, test plan, and URL. | ||||
""" | ||||
title = drev[r'title'] | ||||
summary = drev[r'summary'].rstrip() | ||||
testplan = drev[r'testPlan'].rstrip() | ||||
if testplan: | ||||
testplan = b'Test Plan:\n%s' % testplan | ||||
uri = b'Differential Revision: %s' % drev[r'uri'] | ||||
return b'\n\n'.join(filter(None, [title, summary, testplan, uri])) | ||||
def getdiffmeta(diff): | ||||
"""get commit metadata (date, node, user, p1) from a diff object | ||||
The metadata could be "hg:meta", sent by phabsend, like: | ||||
"properties": { | ||||
"hg:meta": { | ||||
"date": "1499571514 25200", | ||||
"node": "98c08acae292b2faf60a279b4189beb6cff1414d", | ||||
"user": "Foo Bar <foo@example.com>", | ||||
"parent": "6d0abad76b30e4724a37ab8721d630394070fe16" | ||||
} | ||||
} | ||||
Or converted from "local:commits", sent by "arc", like: | ||||
"properties": { | ||||
"local:commits": { | ||||
"98c08acae292b2faf60a279b4189beb6cff1414d": { | ||||
"author": "Foo Bar", | ||||
"time": 1499546314, | ||||
"branch": "default", | ||||
"tag": "", | ||||
"commit": "98c08acae292b2faf60a279b4189beb6cff1414d", | ||||
"rev": "98c08acae292b2faf60a279b4189beb6cff1414d", | ||||
"local": "1000", | ||||
"parents": ["6d0abad76b30e4724a37ab8721d630394070fe16"], | ||||
"summary": "...", | ||||
"message": "...", | ||||
"authorEmail": "foo@example.com" | ||||
} | ||||
} | ||||
} | ||||
Note: metadata extracted from "local:commits" will lose time zone | ||||
information. | ||||
""" | ||||
props = diff.get(r'properties') or {} | ||||
meta = props.get(r'hg:meta') | ||||
if not meta and props.get(r'local:commits'): | ||||
commit = sorted(props[r'local:commits'].values())[0] | ||||
meta = { | ||||
r'date': r'%d 0' % commit[r'time'], | ||||
r'node': commit[r'rev'], | ||||
r'user': r'%s <%s>' % (commit[r'author'], commit[r'authorEmail']), | ||||
} | ||||
if len(commit.get(r'parents', ())) >= 1: | ||||
meta[r'parent'] = commit[r'parents'][0] | ||||
return meta or {} | ||||
def readpatch(repo, drevs, write): | ||||
"""generate plain-text patch readable by 'hg import' | ||||
write is usually ui.write. drevs is what "querydrev" returns, results of | ||||
"differential.query". | ||||
""" | ||||
# Prefetch hg:meta property for all diffs | ||||
diffids = sorted(set(max(int(v) for v in drev[r'diffs']) for drev in drevs)) | ||||
diffs = callconduit(repo, b'differential.querydiffs', {b'ids': diffids}) | ||||
# Generate patch for each drev | ||||
for drev in drevs: | ||||
repo.ui.note(_(b'reading D%s\n') % drev[r'id']) | ||||
diffid = max(int(v) for v in drev[r'diffs']) | ||||
body = callconduit(repo, b'differential.getrawdiff', | ||||
{b'diffID': diffid}) | ||||
desc = getdescfromdrev(drev) | ||||
header = b'# HG changeset patch\n' | ||||
# Try to preserve metadata from hg:meta property. Write hg patch | ||||
# headers that can be read by the "import" command. See patchheadermap | ||||
# and extract in mercurial/patch.py for supported headers. | ||||
meta = getdiffmeta(diffs[str(diffid)]) | ||||
for k in _metanamemap.keys(): | ||||
if k in meta: | ||||
header += b'# %s %s\n' % (_metanamemap[k], meta[k]) | ||||
content = b'%s%s\n%s' % (header, desc, body) | ||||
write(encoding.unitolocal(content)) | ||||
@vcrcommand(b'phabread', | ||||
[(b'', b'stack', False, _(b'read dependencies'))], | ||||
Matt Harbison
|
r41103 | _(b'DREVSPEC [OPTIONS]'), | ||
helpcategory=command.CATEGORY_IMPORT_EXPORT) | ||||
Augie Fackler
|
r39688 | def phabread(ui, repo, spec, **opts): | ||
"""print patches from Phabricator suitable for importing | ||||
DREVSPEC could be a Differential Revision identity, like ``D123``, or just | ||||
the number ``123``. It could also have common operators like ``+``, ``-``, | ||||
``&``, ``(``, ``)`` for complex queries. Prefix ``:`` could be used to | ||||
select a stack. | ||||
``abandoned``, ``accepted``, ``closed``, ``needsreview``, ``needsrevision`` | ||||
could be used to filter patches by status. For performance reason, they | ||||
only represent a subset of non-status selections and cannot be used alone. | ||||
For example, ``:D6+8-(2+D4)`` selects a stack up to D6, plus D8 and exclude | ||||
D2 and D4. ``:D9 & needsreview`` selects "Needs Review" revisions in a | ||||
stack up to D9. | ||||
If --stack is given, follow dependencies information and read all patches. | ||||
It is equivalent to the ``:`` operator. | ||||
""" | ||||
if opts.get(b'stack'): | ||||
spec = b':(%s)' % spec | ||||
drevs = querydrev(repo, spec) | ||||
readpatch(repo, drevs, ui.write) | ||||
@vcrcommand(b'phabupdate', | ||||
[(b'', b'accept', False, _(b'accept revisions')), | ||||
(b'', b'reject', False, _(b'reject revisions')), | ||||
(b'', b'abandon', False, _(b'abandon revisions')), | ||||
(b'', b'reclaim', False, _(b'reclaim revisions')), | ||||
(b'm', b'comment', b'', _(b'comment on the last revision')), | ||||
Matt Harbison
|
r41103 | ], _(b'DREVSPEC [OPTIONS]'), | ||
helpcategory=command.CATEGORY_IMPORT_EXPORT) | ||||
Augie Fackler
|
r39688 | def phabupdate(ui, repo, spec, **opts): | ||
"""update Differential Revision in batch | ||||
DREVSPEC selects revisions. See :hg:`help phabread` for its usage. | ||||
""" | ||||
flags = [n for n in b'accept reject abandon reclaim'.split() if opts.get(n)] | ||||
if len(flags) > 1: | ||||
raise error.Abort(_(b'%s cannot be used together') % b', '.join(flags)) | ||||
actions = [] | ||||
for f in flags: | ||||
actions.append({b'type': f, b'value': b'true'}) | ||||
drevs = querydrev(repo, spec) | ||||
for i, drev in enumerate(drevs): | ||||
if i + 1 == len(drevs) and opts.get(b'comment'): | ||||
actions.append({b'type': b'comment', b'value': opts[b'comment']}) | ||||
if actions: | ||||
params = {b'objectIdentifier': drev[r'phid'], | ||||
b'transactions': actions} | ||||
callconduit(repo, b'differential.revision.edit', params) | ||||
templatekeyword = registrar.templatekeyword() | ||||
@templatekeyword(b'phabreview', requires={b'ctx'}) | ||||
def template_review(context, mapping): | ||||
""":phabreview: Object describing the review for this changeset. | ||||
Has attributes `url` and `id`. | ||||
""" | ||||
ctx = context.resource(mapping, b'ctx') | ||||
m = _differentialrevisiondescre.search(ctx.description()) | ||||
if m: | ||||
Augie Fackler
|
r39690 | return templateutil.hybriddict({ | ||
Augie Fackler
|
r39688 | b'url': m.group(b'url'), | ||
b'id': b"D{}".format(m.group(b'id')), | ||||
Augie Fackler
|
r39690 | }) | ||
Matt Harbison
|
r41199 | else: | ||
tags = ctx.repo().nodetags(ctx.node()) | ||||
for t in tags: | ||||
if _differentialrevisiontagre.match(t): | ||||
url = ctx.repo().ui.config(b'phabricator', b'url') | ||||
if not url.endswith(b'/'): | ||||
url += b'/' | ||||
url += t | ||||
return templateutil.hybriddict({ | ||||
b'url': url, | ||||
b'id': t, | ||||
}) | ||||
return None | ||||