##// END OF EJS Templates
automation: install latest Python versions in Linux...
Gregory Szorc -
r45243:1b80e684 stable
parent child Browse files
Show More
@@ -1,595 +1,595 b''
1 # linux.py - Linux specific automation functionality
1 # linux.py - Linux specific automation functionality
2 #
2 #
3 # Copyright 2019 Gregory Szorc <gregory.szorc@gmail.com>
3 # Copyright 2019 Gregory Szorc <gregory.szorc@gmail.com>
4 #
4 #
5 # This software may be used and distributed according to the terms of the
5 # This software may be used and distributed according to the terms of the
6 # GNU General Public License version 2 or any later version.
6 # GNU General Public License version 2 or any later version.
7
7
8 # no-check-code because Python 3 native.
8 # no-check-code because Python 3 native.
9
9
10 import os
10 import os
11 import pathlib
11 import pathlib
12 import shlex
12 import shlex
13 import subprocess
13 import subprocess
14 import tempfile
14 import tempfile
15
15
16 from .ssh import exec_command
16 from .ssh import exec_command
17
17
18
18
19 # Linux distributions that are supported.
19 # Linux distributions that are supported.
20 DISTROS = {
20 DISTROS = {
21 'debian9',
21 'debian9',
22 'debian10',
22 'debian10',
23 'ubuntu18.04',
23 'ubuntu18.04',
24 'ubuntu19.04',
24 'ubuntu19.04',
25 }
25 }
26
26
27 INSTALL_PYTHONS = r'''
27 INSTALL_PYTHONS = r'''
28 PYENV2_VERSIONS="2.7.17 pypy2.7-7.2.0"
28 PYENV2_VERSIONS="2.7.17 pypy2.7-7.2.0"
29 PYENV3_VERSIONS="3.5.7 3.6.9 3.7.5 3.8.0 pypy3.5-7.0.0 pypy3.6-7.2.0"
29 PYENV3_VERSIONS="3.5.9 3.6.10 3.7.7 3.8.2 pypy3.5-7.0.0 pypy3.6-7.3.0"
30
30
31 git clone https://github.com/pyenv/pyenv.git /hgdev/pyenv
31 git clone https://github.com/pyenv/pyenv.git /hgdev/pyenv
32 pushd /hgdev/pyenv
32 pushd /hgdev/pyenv
33 git checkout 0e7cfc3b3d4eca46ad83d632e1505f5932cd179b
33 git checkout 3005c4664372ae13fbe376be699313eb428c8bdd
34 popd
34 popd
35
35
36 export PYENV_ROOT="/hgdev/pyenv"
36 export PYENV_ROOT="/hgdev/pyenv"
37 export PATH="$PYENV_ROOT/bin:$PATH"
37 export PATH="$PYENV_ROOT/bin:$PATH"
38
38
39 # pip 19.2.3.
39 # pip 19.2.3.
40 PIP_SHA256=57e3643ff19f018f8a00dfaa6b7e4620e3c1a7a2171fd218425366ec006b3bfe
40 PIP_SHA256=57e3643ff19f018f8a00dfaa6b7e4620e3c1a7a2171fd218425366ec006b3bfe
41 wget -O get-pip.py --progress dot:mega https://github.com/pypa/get-pip/raw/309a56c5fd94bd1134053a541cb4657a4e47e09d/get-pip.py
41 wget -O get-pip.py --progress dot:mega https://github.com/pypa/get-pip/raw/309a56c5fd94bd1134053a541cb4657a4e47e09d/get-pip.py
42 echo "${PIP_SHA256} get-pip.py" | sha256sum --check -
42 echo "${PIP_SHA256} get-pip.py" | sha256sum --check -
43
43
44 VIRTUALENV_SHA256=f78d81b62d3147396ac33fc9d77579ddc42cc2a98dd9ea38886f616b33bc7fb2
44 VIRTUALENV_SHA256=f78d81b62d3147396ac33fc9d77579ddc42cc2a98dd9ea38886f616b33bc7fb2
45 VIRTUALENV_TARBALL=virtualenv-16.7.5.tar.gz
45 VIRTUALENV_TARBALL=virtualenv-16.7.5.tar.gz
46 wget -O ${VIRTUALENV_TARBALL} --progress dot:mega https://files.pythonhosted.org/packages/66/f0/6867af06d2e2f511e4e1d7094ff663acdebc4f15d4a0cb0fed1007395124/${VIRTUALENV_TARBALL}
46 wget -O ${VIRTUALENV_TARBALL} --progress dot:mega https://files.pythonhosted.org/packages/66/f0/6867af06d2e2f511e4e1d7094ff663acdebc4f15d4a0cb0fed1007395124/${VIRTUALENV_TARBALL}
47 echo "${VIRTUALENV_SHA256} ${VIRTUALENV_TARBALL}" | sha256sum --check -
47 echo "${VIRTUALENV_SHA256} ${VIRTUALENV_TARBALL}" | sha256sum --check -
48
48
49 for v in ${PYENV2_VERSIONS}; do
49 for v in ${PYENV2_VERSIONS}; do
50 pyenv install -v ${v}
50 pyenv install -v ${v}
51 ${PYENV_ROOT}/versions/${v}/bin/python get-pip.py
51 ${PYENV_ROOT}/versions/${v}/bin/python get-pip.py
52 ${PYENV_ROOT}/versions/${v}/bin/pip install ${VIRTUALENV_TARBALL}
52 ${PYENV_ROOT}/versions/${v}/bin/pip install ${VIRTUALENV_TARBALL}
53 ${PYENV_ROOT}/versions/${v}/bin/pip install -r /hgdev/requirements-py2.txt
53 ${PYENV_ROOT}/versions/${v}/bin/pip install -r /hgdev/requirements-py2.txt
54 done
54 done
55
55
56 for v in ${PYENV3_VERSIONS}; do
56 for v in ${PYENV3_VERSIONS}; do
57 pyenv install -v ${v}
57 pyenv install -v ${v}
58 ${PYENV_ROOT}/versions/${v}/bin/python get-pip.py
58 ${PYENV_ROOT}/versions/${v}/bin/python get-pip.py
59 ${PYENV_ROOT}/versions/${v}/bin/pip install -r /hgdev/requirements-py3.txt
59 ${PYENV_ROOT}/versions/${v}/bin/pip install -r /hgdev/requirements-py3.txt
60 done
60 done
61
61
62 pyenv global ${PYENV2_VERSIONS} ${PYENV3_VERSIONS} system
62 pyenv global ${PYENV2_VERSIONS} ${PYENV3_VERSIONS} system
63 '''.lstrip().replace(
63 '''.lstrip().replace(
64 '\r\n', '\n'
64 '\r\n', '\n'
65 )
65 )
66
66
67
67
68 INSTALL_RUST = r'''
68 INSTALL_RUST = r'''
69 RUSTUP_INIT_SHA256=a46fe67199b7bcbbde2dcbc23ae08db6f29883e260e23899a88b9073effc9076
69 RUSTUP_INIT_SHA256=a46fe67199b7bcbbde2dcbc23ae08db6f29883e260e23899a88b9073effc9076
70 wget -O rustup-init --progress dot:mega https://static.rust-lang.org/rustup/archive/1.18.3/x86_64-unknown-linux-gnu/rustup-init
70 wget -O rustup-init --progress dot:mega https://static.rust-lang.org/rustup/archive/1.18.3/x86_64-unknown-linux-gnu/rustup-init
71 echo "${RUSTUP_INIT_SHA256} rustup-init" | sha256sum --check -
71 echo "${RUSTUP_INIT_SHA256} rustup-init" | sha256sum --check -
72
72
73 chmod +x rustup-init
73 chmod +x rustup-init
74 sudo -H -u hg -g hg ./rustup-init -y
74 sudo -H -u hg -g hg ./rustup-init -y
75 sudo -H -u hg -g hg /home/hg/.cargo/bin/rustup install 1.31.1 1.34.2
75 sudo -H -u hg -g hg /home/hg/.cargo/bin/rustup install 1.31.1 1.34.2
76 sudo -H -u hg -g hg /home/hg/.cargo/bin/rustup component add clippy
76 sudo -H -u hg -g hg /home/hg/.cargo/bin/rustup component add clippy
77 '''
77 '''
78
78
79
79
80 BOOTSTRAP_VIRTUALENV = r'''
80 BOOTSTRAP_VIRTUALENV = r'''
81 /usr/bin/virtualenv /hgdev/venv-bootstrap
81 /usr/bin/virtualenv /hgdev/venv-bootstrap
82
82
83 HG_SHA256=35fc8ba5e0379c1b3affa2757e83fb0509e8ac314cbd9f1fd133cf265d16e49f
83 HG_SHA256=35fc8ba5e0379c1b3affa2757e83fb0509e8ac314cbd9f1fd133cf265d16e49f
84 HG_TARBALL=mercurial-5.1.1.tar.gz
84 HG_TARBALL=mercurial-5.1.1.tar.gz
85
85
86 wget -O ${HG_TARBALL} --progress dot:mega https://www.mercurial-scm.org/release/${HG_TARBALL}
86 wget -O ${HG_TARBALL} --progress dot:mega https://www.mercurial-scm.org/release/${HG_TARBALL}
87 echo "${HG_SHA256} ${HG_TARBALL}" | sha256sum --check -
87 echo "${HG_SHA256} ${HG_TARBALL}" | sha256sum --check -
88
88
89 /hgdev/venv-bootstrap/bin/pip install ${HG_TARBALL}
89 /hgdev/venv-bootstrap/bin/pip install ${HG_TARBALL}
90 '''.lstrip().replace(
90 '''.lstrip().replace(
91 '\r\n', '\n'
91 '\r\n', '\n'
92 )
92 )
93
93
94
94
95 BOOTSTRAP_DEBIAN = (
95 BOOTSTRAP_DEBIAN = (
96 r'''
96 r'''
97 #!/bin/bash
97 #!/bin/bash
98
98
99 set -ex
99 set -ex
100
100
101 DISTRO=`grep DISTRIB_ID /etc/lsb-release | awk -F= '{{print $2}}'`
101 DISTRO=`grep DISTRIB_ID /etc/lsb-release | awk -F= '{{print $2}}'`
102 DEBIAN_VERSION=`cat /etc/debian_version`
102 DEBIAN_VERSION=`cat /etc/debian_version`
103 LSB_RELEASE=`lsb_release -cs`
103 LSB_RELEASE=`lsb_release -cs`
104
104
105 sudo /usr/sbin/groupadd hg
105 sudo /usr/sbin/groupadd hg
106 sudo /usr/sbin/groupadd docker
106 sudo /usr/sbin/groupadd docker
107 sudo /usr/sbin/useradd -g hg -G sudo,docker -d /home/hg -m -s /bin/bash hg
107 sudo /usr/sbin/useradd -g hg -G sudo,docker -d /home/hg -m -s /bin/bash hg
108 sudo mkdir /home/hg/.ssh
108 sudo mkdir /home/hg/.ssh
109 sudo cp ~/.ssh/authorized_keys /home/hg/.ssh/authorized_keys
109 sudo cp ~/.ssh/authorized_keys /home/hg/.ssh/authorized_keys
110 sudo chown -R hg:hg /home/hg/.ssh
110 sudo chown -R hg:hg /home/hg/.ssh
111 sudo chmod 700 /home/hg/.ssh
111 sudo chmod 700 /home/hg/.ssh
112 sudo chmod 600 /home/hg/.ssh/authorized_keys
112 sudo chmod 600 /home/hg/.ssh/authorized_keys
113
113
114 cat << EOF | sudo tee /etc/sudoers.d/90-hg
114 cat << EOF | sudo tee /etc/sudoers.d/90-hg
115 hg ALL=(ALL) NOPASSWD:ALL
115 hg ALL=(ALL) NOPASSWD:ALL
116 EOF
116 EOF
117
117
118 sudo apt-get update
118 sudo apt-get update
119 sudo DEBIAN_FRONTEND=noninteractive apt-get -yq dist-upgrade
119 sudo DEBIAN_FRONTEND=noninteractive apt-get -yq dist-upgrade
120
120
121 # Install packages necessary to set up Docker Apt repo.
121 # Install packages necessary to set up Docker Apt repo.
122 sudo DEBIAN_FRONTEND=noninteractive apt-get -yq install --no-install-recommends \
122 sudo DEBIAN_FRONTEND=noninteractive apt-get -yq install --no-install-recommends \
123 apt-transport-https \
123 apt-transport-https \
124 gnupg
124 gnupg
125
125
126 cat > docker-apt-key << EOF
126 cat > docker-apt-key << EOF
127 -----BEGIN PGP PUBLIC KEY BLOCK-----
127 -----BEGIN PGP PUBLIC KEY BLOCK-----
128
128
129 mQINBFit2ioBEADhWpZ8/wvZ6hUTiXOwQHXMAlaFHcPH9hAtr4F1y2+OYdbtMuth
129 mQINBFit2ioBEADhWpZ8/wvZ6hUTiXOwQHXMAlaFHcPH9hAtr4F1y2+OYdbtMuth
130 lqqwp028AqyY+PRfVMtSYMbjuQuu5byyKR01BbqYhuS3jtqQmljZ/bJvXqnmiVXh
130 lqqwp028AqyY+PRfVMtSYMbjuQuu5byyKR01BbqYhuS3jtqQmljZ/bJvXqnmiVXh
131 38UuLa+z077PxyxQhu5BbqntTPQMfiyqEiU+BKbq2WmANUKQf+1AmZY/IruOXbnq
131 38UuLa+z077PxyxQhu5BbqntTPQMfiyqEiU+BKbq2WmANUKQf+1AmZY/IruOXbnq
132 L4C1+gJ8vfmXQt99npCaxEjaNRVYfOS8QcixNzHUYnb6emjlANyEVlZzeqo7XKl7
132 L4C1+gJ8vfmXQt99npCaxEjaNRVYfOS8QcixNzHUYnb6emjlANyEVlZzeqo7XKl7
133 UrwV5inawTSzWNvtjEjj4nJL8NsLwscpLPQUhTQ+7BbQXAwAmeHCUTQIvvWXqw0N
133 UrwV5inawTSzWNvtjEjj4nJL8NsLwscpLPQUhTQ+7BbQXAwAmeHCUTQIvvWXqw0N
134 cmhh4HgeQscQHYgOJjjDVfoY5MucvglbIgCqfzAHW9jxmRL4qbMZj+b1XoePEtht
134 cmhh4HgeQscQHYgOJjjDVfoY5MucvglbIgCqfzAHW9jxmRL4qbMZj+b1XoePEtht
135 ku4bIQN1X5P07fNWzlgaRL5Z4POXDDZTlIQ/El58j9kp4bnWRCJW0lya+f8ocodo
135 ku4bIQN1X5P07fNWzlgaRL5Z4POXDDZTlIQ/El58j9kp4bnWRCJW0lya+f8ocodo
136 vZZ+Doi+fy4D5ZGrL4XEcIQP/Lv5uFyf+kQtl/94VFYVJOleAv8W92KdgDkhTcTD
136 vZZ+Doi+fy4D5ZGrL4XEcIQP/Lv5uFyf+kQtl/94VFYVJOleAv8W92KdgDkhTcTD
137 G7c0tIkVEKNUq48b3aQ64NOZQW7fVjfoKwEZdOqPE72Pa45jrZzvUFxSpdiNk2tZ
137 G7c0tIkVEKNUq48b3aQ64NOZQW7fVjfoKwEZdOqPE72Pa45jrZzvUFxSpdiNk2tZ
138 XYukHjlxxEgBdC/J3cMMNRE1F4NCA3ApfV1Y7/hTeOnmDuDYwr9/obA8t016Yljj
138 XYukHjlxxEgBdC/J3cMMNRE1F4NCA3ApfV1Y7/hTeOnmDuDYwr9/obA8t016Yljj
139 q5rdkywPf4JF8mXUW5eCN1vAFHxeg9ZWemhBtQmGxXnw9M+z6hWwc6ahmwARAQAB
139 q5rdkywPf4JF8mXUW5eCN1vAFHxeg9ZWemhBtQmGxXnw9M+z6hWwc6ahmwARAQAB
140 tCtEb2NrZXIgUmVsZWFzZSAoQ0UgZGViKSA8ZG9ja2VyQGRvY2tlci5jb20+iQI3
140 tCtEb2NrZXIgUmVsZWFzZSAoQ0UgZGViKSA8ZG9ja2VyQGRvY2tlci5jb20+iQI3
141 BBMBCgAhBQJYrefAAhsvBQsJCAcDBRUKCQgLBRYCAwEAAh4BAheAAAoJEI2BgDwO
141 BBMBCgAhBQJYrefAAhsvBQsJCAcDBRUKCQgLBRYCAwEAAh4BAheAAAoJEI2BgDwO
142 v82IsskP/iQZo68flDQmNvn8X5XTd6RRaUH33kXYXquT6NkHJciS7E2gTJmqvMqd
142 v82IsskP/iQZo68flDQmNvn8X5XTd6RRaUH33kXYXquT6NkHJciS7E2gTJmqvMqd
143 tI4mNYHCSEYxI5qrcYV5YqX9P6+Ko+vozo4nseUQLPH/ATQ4qL0Zok+1jkag3Lgk
143 tI4mNYHCSEYxI5qrcYV5YqX9P6+Ko+vozo4nseUQLPH/ATQ4qL0Zok+1jkag3Lgk
144 jonyUf9bwtWxFp05HC3GMHPhhcUSexCxQLQvnFWXD2sWLKivHp2fT8QbRGeZ+d3m
144 jonyUf9bwtWxFp05HC3GMHPhhcUSexCxQLQvnFWXD2sWLKivHp2fT8QbRGeZ+d3m
145 6fqcd5Fu7pxsqm0EUDK5NL+nPIgYhN+auTrhgzhK1CShfGccM/wfRlei9Utz6p9P
145 6fqcd5Fu7pxsqm0EUDK5NL+nPIgYhN+auTrhgzhK1CShfGccM/wfRlei9Utz6p9P
146 XRKIlWnXtT4qNGZNTN0tR+NLG/6Bqd8OYBaFAUcue/w1VW6JQ2VGYZHnZu9S8LMc
146 XRKIlWnXtT4qNGZNTN0tR+NLG/6Bqd8OYBaFAUcue/w1VW6JQ2VGYZHnZu9S8LMc
147 FYBa5Ig9PxwGQOgq6RDKDbV+PqTQT5EFMeR1mrjckk4DQJjbxeMZbiNMG5kGECA8
147 FYBa5Ig9PxwGQOgq6RDKDbV+PqTQT5EFMeR1mrjckk4DQJjbxeMZbiNMG5kGECA8
148 g383P3elhn03WGbEEa4MNc3Z4+7c236QI3xWJfNPdUbXRaAwhy/6rTSFbzwKB0Jm
148 g383P3elhn03WGbEEa4MNc3Z4+7c236QI3xWJfNPdUbXRaAwhy/6rTSFbzwKB0Jm
149 ebwzQfwjQY6f55MiI/RqDCyuPj3r3jyVRkK86pQKBAJwFHyqj9KaKXMZjfVnowLh
149 ebwzQfwjQY6f55MiI/RqDCyuPj3r3jyVRkK86pQKBAJwFHyqj9KaKXMZjfVnowLh
150 9svIGfNbGHpucATqREvUHuQbNnqkCx8VVhtYkhDb9fEP2xBu5VvHbR+3nfVhMut5
150 9svIGfNbGHpucATqREvUHuQbNnqkCx8VVhtYkhDb9fEP2xBu5VvHbR+3nfVhMut5
151 G34Ct5RS7Jt6LIfFdtcn8CaSas/l1HbiGeRgc70X/9aYx/V/CEJv0lIe8gP6uDoW
151 G34Ct5RS7Jt6LIfFdtcn8CaSas/l1HbiGeRgc70X/9aYx/V/CEJv0lIe8gP6uDoW
152 FPIZ7d6vH+Vro6xuWEGiuMaiznap2KhZmpkgfupyFmplh0s6knymuQINBFit2ioB
152 FPIZ7d6vH+Vro6xuWEGiuMaiznap2KhZmpkgfupyFmplh0s6knymuQINBFit2ioB
153 EADneL9S9m4vhU3blaRjVUUyJ7b/qTjcSylvCH5XUE6R2k+ckEZjfAMZPLpO+/tF
153 EADneL9S9m4vhU3blaRjVUUyJ7b/qTjcSylvCH5XUE6R2k+ckEZjfAMZPLpO+/tF
154 M2JIJMD4SifKuS3xck9KtZGCufGmcwiLQRzeHF7vJUKrLD5RTkNi23ydvWZgPjtx
154 M2JIJMD4SifKuS3xck9KtZGCufGmcwiLQRzeHF7vJUKrLD5RTkNi23ydvWZgPjtx
155 Q+DTT1Zcn7BrQFY6FgnRoUVIxwtdw1bMY/89rsFgS5wwuMESd3Q2RYgb7EOFOpnu
155 Q+DTT1Zcn7BrQFY6FgnRoUVIxwtdw1bMY/89rsFgS5wwuMESd3Q2RYgb7EOFOpnu
156 w6da7WakWf4IhnF5nsNYGDVaIHzpiqCl+uTbf1epCjrOlIzkZ3Z3Yk5CM/TiFzPk
156 w6da7WakWf4IhnF5nsNYGDVaIHzpiqCl+uTbf1epCjrOlIzkZ3Z3Yk5CM/TiFzPk
157 z2lLz89cpD8U+NtCsfagWWfjd2U3jDapgH+7nQnCEWpROtzaKHG6lA3pXdix5zG8
157 z2lLz89cpD8U+NtCsfagWWfjd2U3jDapgH+7nQnCEWpROtzaKHG6lA3pXdix5zG8
158 eRc6/0IbUSWvfjKxLLPfNeCS2pCL3IeEI5nothEEYdQH6szpLog79xB9dVnJyKJb
158 eRc6/0IbUSWvfjKxLLPfNeCS2pCL3IeEI5nothEEYdQH6szpLog79xB9dVnJyKJb
159 VfxXnseoYqVrRz2VVbUI5Blwm6B40E3eGVfUQWiux54DspyVMMk41Mx7QJ3iynIa
159 VfxXnseoYqVrRz2VVbUI5Blwm6B40E3eGVfUQWiux54DspyVMMk41Mx7QJ3iynIa
160 1N4ZAqVMAEruyXTRTxc9XW0tYhDMA/1GYvz0EmFpm8LzTHA6sFVtPm/ZlNCX6P1X
160 1N4ZAqVMAEruyXTRTxc9XW0tYhDMA/1GYvz0EmFpm8LzTHA6sFVtPm/ZlNCX6P1X
161 zJwrv7DSQKD6GGlBQUX+OeEJ8tTkkf8QTJSPUdh8P8YxDFS5EOGAvhhpMBYD42kQ
161 zJwrv7DSQKD6GGlBQUX+OeEJ8tTkkf8QTJSPUdh8P8YxDFS5EOGAvhhpMBYD42kQ
162 pqXjEC+XcycTvGI7impgv9PDY1RCC1zkBjKPa120rNhv/hkVk/YhuGoajoHyy4h7
162 pqXjEC+XcycTvGI7impgv9PDY1RCC1zkBjKPa120rNhv/hkVk/YhuGoajoHyy4h7
163 ZQopdcMtpN2dgmhEegny9JCSwxfQmQ0zK0g7m6SHiKMwjwARAQABiQQ+BBgBCAAJ
163 ZQopdcMtpN2dgmhEegny9JCSwxfQmQ0zK0g7m6SHiKMwjwARAQABiQQ+BBgBCAAJ
164 BQJYrdoqAhsCAikJEI2BgDwOv82IwV0gBBkBCAAGBQJYrdoqAAoJEH6gqcPyc/zY
164 BQJYrdoqAhsCAikJEI2BgDwOv82IwV0gBBkBCAAGBQJYrdoqAAoJEH6gqcPyc/zY
165 1WAP/2wJ+R0gE6qsce3rjaIz58PJmc8goKrir5hnElWhPgbq7cYIsW5qiFyLhkdp
165 1WAP/2wJ+R0gE6qsce3rjaIz58PJmc8goKrir5hnElWhPgbq7cYIsW5qiFyLhkdp
166 YcMmhD9mRiPpQn6Ya2w3e3B8zfIVKipbMBnke/ytZ9M7qHmDCcjoiSmwEXN3wKYI
166 YcMmhD9mRiPpQn6Ya2w3e3B8zfIVKipbMBnke/ytZ9M7qHmDCcjoiSmwEXN3wKYI
167 mD9VHONsl/CG1rU9Isw1jtB5g1YxuBA7M/m36XN6x2u+NtNMDB9P56yc4gfsZVES
167 mD9VHONsl/CG1rU9Isw1jtB5g1YxuBA7M/m36XN6x2u+NtNMDB9P56yc4gfsZVES
168 KA9v+yY2/l45L8d/WUkUi0YXomn6hyBGI7JrBLq0CX37GEYP6O9rrKipfz73XfO7
168 KA9v+yY2/l45L8d/WUkUi0YXomn6hyBGI7JrBLq0CX37GEYP6O9rrKipfz73XfO7
169 JIGzOKZlljb/D9RX/g7nRbCn+3EtH7xnk+TK/50euEKw8SMUg147sJTcpQmv6UzZ
169 JIGzOKZlljb/D9RX/g7nRbCn+3EtH7xnk+TK/50euEKw8SMUg147sJTcpQmv6UzZ
170 cM4JgL0HbHVCojV4C/plELwMddALOFeYQzTif6sMRPf+3DSj8frbInjChC3yOLy0
170 cM4JgL0HbHVCojV4C/plELwMddALOFeYQzTif6sMRPf+3DSj8frbInjChC3yOLy0
171 6br92KFom17EIj2CAcoeq7UPhi2oouYBwPxh5ytdehJkoo+sN7RIWua6P2WSmon5
171 6br92KFom17EIj2CAcoeq7UPhi2oouYBwPxh5ytdehJkoo+sN7RIWua6P2WSmon5
172 U888cSylXC0+ADFdgLX9K2zrDVYUG1vo8CX0vzxFBaHwN6Px26fhIT1/hYUHQR1z
172 U888cSylXC0+ADFdgLX9K2zrDVYUG1vo8CX0vzxFBaHwN6Px26fhIT1/hYUHQR1z
173 VfNDcyQmXqkOnZvvoMfz/Q0s9BhFJ/zU6AgQbIZE/hm1spsfgvtsD1frZfygXJ9f
173 VfNDcyQmXqkOnZvvoMfz/Q0s9BhFJ/zU6AgQbIZE/hm1spsfgvtsD1frZfygXJ9f
174 irP+MSAI80xHSf91qSRZOj4Pl3ZJNbq4yYxv0b1pkMqeGdjdCYhLU+LZ4wbQmpCk
174 irP+MSAI80xHSf91qSRZOj4Pl3ZJNbq4yYxv0b1pkMqeGdjdCYhLU+LZ4wbQmpCk
175 SVe2prlLureigXtmZfkqevRz7FrIZiu9ky8wnCAPwC7/zmS18rgP/17bOtL4/iIz
175 SVe2prlLureigXtmZfkqevRz7FrIZiu9ky8wnCAPwC7/zmS18rgP/17bOtL4/iIz
176 QhxAAoAMWVrGyJivSkjhSGx1uCojsWfsTAm11P7jsruIL61ZzMUVE2aM3Pmj5G+W
176 QhxAAoAMWVrGyJivSkjhSGx1uCojsWfsTAm11P7jsruIL61ZzMUVE2aM3Pmj5G+W
177 9AcZ58Em+1WsVnAXdUR//bMmhyr8wL/G1YO1V3JEJTRdxsSxdYa4deGBBY/Adpsw
177 9AcZ58Em+1WsVnAXdUR//bMmhyr8wL/G1YO1V3JEJTRdxsSxdYa4deGBBY/Adpsw
178 24jxhOJR+lsJpqIUeb999+R8euDhRHG9eFO7DRu6weatUJ6suupoDTRWtr/4yGqe
178 24jxhOJR+lsJpqIUeb999+R8euDhRHG9eFO7DRu6weatUJ6suupoDTRWtr/4yGqe
179 dKxV3qQhNLSnaAzqW/1nA3iUB4k7kCaKZxhdhDbClf9P37qaRW467BLCVO/coL3y
179 dKxV3qQhNLSnaAzqW/1nA3iUB4k7kCaKZxhdhDbClf9P37qaRW467BLCVO/coL3y
180 Vm50dwdrNtKpMBh3ZpbB1uJvgi9mXtyBOMJ3v8RZeDzFiG8HdCtg9RvIt/AIFoHR
180 Vm50dwdrNtKpMBh3ZpbB1uJvgi9mXtyBOMJ3v8RZeDzFiG8HdCtg9RvIt/AIFoHR
181 H3S+U79NT6i0KPzLImDfs8T7RlpyuMc4Ufs8ggyg9v3Ae6cN3eQyxcK3w0cbBwsh
181 H3S+U79NT6i0KPzLImDfs8T7RlpyuMc4Ufs8ggyg9v3Ae6cN3eQyxcK3w0cbBwsh
182 /nQNfsA6uu+9H7NhbehBMhYnpNZyrHzCmzyXkauwRAqoCbGCNykTRwsur9gS41TQ
182 /nQNfsA6uu+9H7NhbehBMhYnpNZyrHzCmzyXkauwRAqoCbGCNykTRwsur9gS41TQ
183 M8ssD1jFheOJf3hODnkKU+HKjvMROl1DK7zdmLdNzA1cvtZH/nCC9KPj1z8QC47S
183 M8ssD1jFheOJf3hODnkKU+HKjvMROl1DK7zdmLdNzA1cvtZH/nCC9KPj1z8QC47S
184 xx+dTZSx4ONAhwbS/LN3PoKtn8LPjY9NP9uDWI+TWYquS2U+KHDrBDlsgozDbs/O
184 xx+dTZSx4ONAhwbS/LN3PoKtn8LPjY9NP9uDWI+TWYquS2U+KHDrBDlsgozDbs/O
185 jCxcpDzNmXpWQHEtHU7649OXHP7UeNST1mCUCH5qdank0V1iejF6/CfTFU4MfcrG
185 jCxcpDzNmXpWQHEtHU7649OXHP7UeNST1mCUCH5qdank0V1iejF6/CfTFU4MfcrG
186 YT90qFF93M3v01BbxP+EIY2/9tiIPbrd
186 YT90qFF93M3v01BbxP+EIY2/9tiIPbrd
187 =0YYh
187 =0YYh
188 -----END PGP PUBLIC KEY BLOCK-----
188 -----END PGP PUBLIC KEY BLOCK-----
189 EOF
189 EOF
190
190
191 sudo apt-key add docker-apt-key
191 sudo apt-key add docker-apt-key
192
192
193 if [ "$LSB_RELEASE" = "stretch" ]; then
193 if [ "$LSB_RELEASE" = "stretch" ]; then
194 cat << EOF | sudo tee -a /etc/apt/sources.list
194 cat << EOF | sudo tee -a /etc/apt/sources.list
195 # Need backports for clang-format-6.0
195 # Need backports for clang-format-6.0
196 deb http://deb.debian.org/debian stretch-backports main
196 deb http://deb.debian.org/debian stretch-backports main
197 EOF
197 EOF
198 fi
198 fi
199
199
200 if [ "$LSB_RELEASE" = "stretch" -o "$LSB_RELEASE" = "buster" ]; then
200 if [ "$LSB_RELEASE" = "stretch" -o "$LSB_RELEASE" = "buster" ]; then
201 cat << EOF | sudo tee -a /etc/apt/sources.list
201 cat << EOF | sudo tee -a /etc/apt/sources.list
202 # Sources are useful if we want to compile things locally.
202 # Sources are useful if we want to compile things locally.
203 deb-src http://deb.debian.org/debian $LSB_RELEASE main
203 deb-src http://deb.debian.org/debian $LSB_RELEASE main
204 deb-src http://security.debian.org/debian-security $LSB_RELEASE/updates main
204 deb-src http://security.debian.org/debian-security $LSB_RELEASE/updates main
205 deb-src http://deb.debian.org/debian $LSB_RELEASE-updates main
205 deb-src http://deb.debian.org/debian $LSB_RELEASE-updates main
206 deb-src http://deb.debian.org/debian $LSB_RELEASE-backports main
206 deb-src http://deb.debian.org/debian $LSB_RELEASE-backports main
207
207
208 deb [arch=amd64] https://download.docker.com/linux/debian $LSB_RELEASE stable
208 deb [arch=amd64] https://download.docker.com/linux/debian $LSB_RELEASE stable
209 EOF
209 EOF
210
210
211 elif [ "$DISTRO" = "Ubuntu" ]; then
211 elif [ "$DISTRO" = "Ubuntu" ]; then
212 cat << EOF | sudo tee -a /etc/apt/sources.list
212 cat << EOF | sudo tee -a /etc/apt/sources.list
213 deb [arch=amd64] https://download.docker.com/linux/ubuntu $LSB_RELEASE stable
213 deb [arch=amd64] https://download.docker.com/linux/ubuntu $LSB_RELEASE stable
214 EOF
214 EOF
215
215
216 fi
216 fi
217
217
218 sudo apt-get update
218 sudo apt-get update
219
219
220 PACKAGES="\
220 PACKAGES="\
221 awscli \
221 awscli \
222 btrfs-progs \
222 btrfs-progs \
223 build-essential \
223 build-essential \
224 bzr \
224 bzr \
225 clang-format-6.0 \
225 clang-format-6.0 \
226 cvs \
226 cvs \
227 darcs \
227 darcs \
228 debhelper \
228 debhelper \
229 devscripts \
229 devscripts \
230 docker-ce \
230 docker-ce \
231 dpkg-dev \
231 dpkg-dev \
232 dstat \
232 dstat \
233 emacs \
233 emacs \
234 gettext \
234 gettext \
235 git \
235 git \
236 htop \
236 htop \
237 iotop \
237 iotop \
238 jfsutils \
238 jfsutils \
239 libbz2-dev \
239 libbz2-dev \
240 libexpat1-dev \
240 libexpat1-dev \
241 libffi-dev \
241 libffi-dev \
242 libgdbm-dev \
242 libgdbm-dev \
243 liblzma-dev \
243 liblzma-dev \
244 libncurses5-dev \
244 libncurses5-dev \
245 libnss3-dev \
245 libnss3-dev \
246 libreadline-dev \
246 libreadline-dev \
247 libsqlite3-dev \
247 libsqlite3-dev \
248 libssl-dev \
248 libssl-dev \
249 netbase \
249 netbase \
250 ntfs-3g \
250 ntfs-3g \
251 nvme-cli \
251 nvme-cli \
252 pyflakes \
252 pyflakes \
253 pyflakes3 \
253 pyflakes3 \
254 pylint \
254 pylint \
255 pylint3 \
255 pylint3 \
256 python-all-dev \
256 python-all-dev \
257 python-dev \
257 python-dev \
258 python-docutils \
258 python-docutils \
259 python-fuzzywuzzy \
259 python-fuzzywuzzy \
260 python-pygments \
260 python-pygments \
261 python-subversion \
261 python-subversion \
262 python-vcr \
262 python-vcr \
263 python3-boto3 \
263 python3-boto3 \
264 python3-dev \
264 python3-dev \
265 python3-docutils \
265 python3-docutils \
266 python3-fuzzywuzzy \
266 python3-fuzzywuzzy \
267 python3-pygments \
267 python3-pygments \
268 python3-vcr \
268 python3-vcr \
269 python3-venv \
269 python3-venv \
270 rsync \
270 rsync \
271 sqlite3 \
271 sqlite3 \
272 subversion \
272 subversion \
273 tcl-dev \
273 tcl-dev \
274 tk-dev \
274 tk-dev \
275 tla \
275 tla \
276 unzip \
276 unzip \
277 uuid-dev \
277 uuid-dev \
278 vim \
278 vim \
279 virtualenv \
279 virtualenv \
280 wget \
280 wget \
281 xfsprogs \
281 xfsprogs \
282 zip \
282 zip \
283 zlib1g-dev"
283 zlib1g-dev"
284
284
285 if [ "LSB_RELEASE" = "stretch" ]; then
285 if [ "LSB_RELEASE" = "stretch" ]; then
286 PACKAGES="$PACKAGES linux-perf"
286 PACKAGES="$PACKAGES linux-perf"
287 elif [ "$DISTRO" = "Ubuntu" ]; then
287 elif [ "$DISTRO" = "Ubuntu" ]; then
288 PACKAGES="$PACKAGES linux-tools-common"
288 PACKAGES="$PACKAGES linux-tools-common"
289 fi
289 fi
290
290
291 # Monotone only available in older releases.
291 # Monotone only available in older releases.
292 if [ "$LSB_RELEASE" = "stretch" -o "$LSB_RELEASE" = "xenial" ]; then
292 if [ "$LSB_RELEASE" = "stretch" -o "$LSB_RELEASE" = "xenial" ]; then
293 PACKAGES="$PACKAGES monotone"
293 PACKAGES="$PACKAGES monotone"
294 fi
294 fi
295
295
296 sudo DEBIAN_FRONTEND=noninteractive apt-get -yq install --no-install-recommends $PACKAGES
296 sudo DEBIAN_FRONTEND=noninteractive apt-get -yq install --no-install-recommends $PACKAGES
297
297
298 # Create clang-format symlink so test harness finds it.
298 # Create clang-format symlink so test harness finds it.
299 sudo update-alternatives --install /usr/bin/clang-format clang-format \
299 sudo update-alternatives --install /usr/bin/clang-format clang-format \
300 /usr/bin/clang-format-6.0 1000
300 /usr/bin/clang-format-6.0 1000
301
301
302 sudo mkdir /hgdev
302 sudo mkdir /hgdev
303 # Will be normalized to hg:hg later.
303 # Will be normalized to hg:hg later.
304 sudo chown `whoami` /hgdev
304 sudo chown `whoami` /hgdev
305
305
306 {install_rust}
306 {install_rust}
307
307
308 cp requirements-py2.txt /hgdev/requirements-py2.txt
308 cp requirements-py2.txt /hgdev/requirements-py2.txt
309 cp requirements-py3.txt /hgdev/requirements-py3.txt
309 cp requirements-py3.txt /hgdev/requirements-py3.txt
310
310
311 # Disable the pip version check because it uses the network and can
311 # Disable the pip version check because it uses the network and can
312 # be annoying.
312 # be annoying.
313 cat << EOF | sudo tee -a /etc/pip.conf
313 cat << EOF | sudo tee -a /etc/pip.conf
314 [global]
314 [global]
315 disable-pip-version-check = True
315 disable-pip-version-check = True
316 EOF
316 EOF
317
317
318 {install_pythons}
318 {install_pythons}
319 {bootstrap_virtualenv}
319 {bootstrap_virtualenv}
320
320
321 /hgdev/venv-bootstrap/bin/hg clone https://www.mercurial-scm.org/repo/hg /hgdev/src
321 /hgdev/venv-bootstrap/bin/hg clone https://www.mercurial-scm.org/repo/hg /hgdev/src
322
322
323 # Mark the repo as non-publishing.
323 # Mark the repo as non-publishing.
324 cat >> /hgdev/src/.hg/hgrc << EOF
324 cat >> /hgdev/src/.hg/hgrc << EOF
325 [phases]
325 [phases]
326 publish = false
326 publish = false
327 EOF
327 EOF
328
328
329 sudo chown -R hg:hg /hgdev
329 sudo chown -R hg:hg /hgdev
330 '''.lstrip()
330 '''.lstrip()
331 .format(
331 .format(
332 install_rust=INSTALL_RUST,
332 install_rust=INSTALL_RUST,
333 install_pythons=INSTALL_PYTHONS,
333 install_pythons=INSTALL_PYTHONS,
334 bootstrap_virtualenv=BOOTSTRAP_VIRTUALENV,
334 bootstrap_virtualenv=BOOTSTRAP_VIRTUALENV,
335 )
335 )
336 .replace('\r\n', '\n')
336 .replace('\r\n', '\n')
337 )
337 )
338
338
339
339
340 # Prepares /hgdev for operations.
340 # Prepares /hgdev for operations.
341 PREPARE_HGDEV = '''
341 PREPARE_HGDEV = '''
342 #!/bin/bash
342 #!/bin/bash
343
343
344 set -e
344 set -e
345
345
346 FS=$1
346 FS=$1
347
347
348 ensure_device() {
348 ensure_device() {
349 if [ -z "${DEVICE}" ]; then
349 if [ -z "${DEVICE}" ]; then
350 echo "could not find block device to format"
350 echo "could not find block device to format"
351 exit 1
351 exit 1
352 fi
352 fi
353 }
353 }
354
354
355 # Determine device to partition for extra filesystem.
355 # Determine device to partition for extra filesystem.
356 # If only 1 volume is present, it will be the root volume and
356 # If only 1 volume is present, it will be the root volume and
357 # should be /dev/nvme0. If multiple volumes are present, the
357 # should be /dev/nvme0. If multiple volumes are present, the
358 # root volume could be nvme0 or nvme1. Use whichever one doesn't have
358 # root volume could be nvme0 or nvme1. Use whichever one doesn't have
359 # a partition.
359 # a partition.
360 if [ -e /dev/nvme1n1 ]; then
360 if [ -e /dev/nvme1n1 ]; then
361 if [ -e /dev/nvme0n1p1 ]; then
361 if [ -e /dev/nvme0n1p1 ]; then
362 DEVICE=/dev/nvme1n1
362 DEVICE=/dev/nvme1n1
363 else
363 else
364 DEVICE=/dev/nvme0n1
364 DEVICE=/dev/nvme0n1
365 fi
365 fi
366 else
366 else
367 DEVICE=
367 DEVICE=
368 fi
368 fi
369
369
370 sudo mkdir /hgwork
370 sudo mkdir /hgwork
371
371
372 if [ "${FS}" != "default" -a "${FS}" != "tmpfs" ]; then
372 if [ "${FS}" != "default" -a "${FS}" != "tmpfs" ]; then
373 ensure_device
373 ensure_device
374 echo "creating ${FS} filesystem on ${DEVICE}"
374 echo "creating ${FS} filesystem on ${DEVICE}"
375 fi
375 fi
376
376
377 if [ "${FS}" = "default" ]; then
377 if [ "${FS}" = "default" ]; then
378 :
378 :
379
379
380 elif [ "${FS}" = "btrfs" ]; then
380 elif [ "${FS}" = "btrfs" ]; then
381 sudo mkfs.btrfs ${DEVICE}
381 sudo mkfs.btrfs ${DEVICE}
382 sudo mount ${DEVICE} /hgwork
382 sudo mount ${DEVICE} /hgwork
383
383
384 elif [ "${FS}" = "ext3" ]; then
384 elif [ "${FS}" = "ext3" ]; then
385 # lazy_journal_init speeds up filesystem creation at the expense of
385 # lazy_journal_init speeds up filesystem creation at the expense of
386 # integrity if things crash. We are an ephemeral instance, so we don't
386 # integrity if things crash. We are an ephemeral instance, so we don't
387 # care about integrity.
387 # care about integrity.
388 sudo mkfs.ext3 -E lazy_journal_init=1 ${DEVICE}
388 sudo mkfs.ext3 -E lazy_journal_init=1 ${DEVICE}
389 sudo mount ${DEVICE} /hgwork
389 sudo mount ${DEVICE} /hgwork
390
390
391 elif [ "${FS}" = "ext4" ]; then
391 elif [ "${FS}" = "ext4" ]; then
392 sudo mkfs.ext4 -E lazy_journal_init=1 ${DEVICE}
392 sudo mkfs.ext4 -E lazy_journal_init=1 ${DEVICE}
393 sudo mount ${DEVICE} /hgwork
393 sudo mount ${DEVICE} /hgwork
394
394
395 elif [ "${FS}" = "jfs" ]; then
395 elif [ "${FS}" = "jfs" ]; then
396 sudo mkfs.jfs ${DEVICE}
396 sudo mkfs.jfs ${DEVICE}
397 sudo mount ${DEVICE} /hgwork
397 sudo mount ${DEVICE} /hgwork
398
398
399 elif [ "${FS}" = "tmpfs" ]; then
399 elif [ "${FS}" = "tmpfs" ]; then
400 echo "creating tmpfs volume in /hgwork"
400 echo "creating tmpfs volume in /hgwork"
401 sudo mount -t tmpfs -o size=1024M tmpfs /hgwork
401 sudo mount -t tmpfs -o size=1024M tmpfs /hgwork
402
402
403 elif [ "${FS}" = "xfs" ]; then
403 elif [ "${FS}" = "xfs" ]; then
404 sudo mkfs.xfs ${DEVICE}
404 sudo mkfs.xfs ${DEVICE}
405 sudo mount ${DEVICE} /hgwork
405 sudo mount ${DEVICE} /hgwork
406
406
407 else
407 else
408 echo "unsupported filesystem: ${FS}"
408 echo "unsupported filesystem: ${FS}"
409 exit 1
409 exit 1
410 fi
410 fi
411
411
412 echo "/hgwork ready"
412 echo "/hgwork ready"
413
413
414 sudo chown hg:hg /hgwork
414 sudo chown hg:hg /hgwork
415 mkdir /hgwork/tmp
415 mkdir /hgwork/tmp
416 chown hg:hg /hgwork/tmp
416 chown hg:hg /hgwork/tmp
417
417
418 rsync -a /hgdev/src /hgwork/
418 rsync -a /hgdev/src /hgwork/
419 '''.lstrip().replace(
419 '''.lstrip().replace(
420 '\r\n', '\n'
420 '\r\n', '\n'
421 )
421 )
422
422
423
423
424 HG_UPDATE_CLEAN = '''
424 HG_UPDATE_CLEAN = '''
425 set -ex
425 set -ex
426
426
427 HG=/hgdev/venv-bootstrap/bin/hg
427 HG=/hgdev/venv-bootstrap/bin/hg
428
428
429 cd /hgwork/src
429 cd /hgwork/src
430 ${HG} --config extensions.purge= purge --all
430 ${HG} --config extensions.purge= purge --all
431 ${HG} update -C $1
431 ${HG} update -C $1
432 ${HG} log -r .
432 ${HG} log -r .
433 '''.lstrip().replace(
433 '''.lstrip().replace(
434 '\r\n', '\n'
434 '\r\n', '\n'
435 )
435 )
436
436
437
437
438 def prepare_exec_environment(ssh_client, filesystem='default'):
438 def prepare_exec_environment(ssh_client, filesystem='default'):
439 """Prepare an EC2 instance to execute things.
439 """Prepare an EC2 instance to execute things.
440
440
441 The AMI has an ``/hgdev`` bootstrapped with various Python installs
441 The AMI has an ``/hgdev`` bootstrapped with various Python installs
442 and a clone of the Mercurial repo.
442 and a clone of the Mercurial repo.
443
443
444 In EC2, EBS volumes launched from snapshots have wonky performance behavior.
444 In EC2, EBS volumes launched from snapshots have wonky performance behavior.
445 Notably, blocks have to be copied on first access, which makes volume
445 Notably, blocks have to be copied on first access, which makes volume
446 I/O extremely slow on fresh volumes.
446 I/O extremely slow on fresh volumes.
447
447
448 Furthermore, we may want to run operations, tests, etc on alternative
448 Furthermore, we may want to run operations, tests, etc on alternative
449 filesystems so we examine behavior on different filesystems.
449 filesystems so we examine behavior on different filesystems.
450
450
451 This function is used to facilitate executing operations on alternate
451 This function is used to facilitate executing operations on alternate
452 volumes.
452 volumes.
453 """
453 """
454 sftp = ssh_client.open_sftp()
454 sftp = ssh_client.open_sftp()
455
455
456 with sftp.open('/hgdev/prepare-hgdev', 'wb') as fh:
456 with sftp.open('/hgdev/prepare-hgdev', 'wb') as fh:
457 fh.write(PREPARE_HGDEV)
457 fh.write(PREPARE_HGDEV)
458 fh.chmod(0o0777)
458 fh.chmod(0o0777)
459
459
460 command = 'sudo /hgdev/prepare-hgdev %s' % filesystem
460 command = 'sudo /hgdev/prepare-hgdev %s' % filesystem
461 chan, stdin, stdout = exec_command(ssh_client, command)
461 chan, stdin, stdout = exec_command(ssh_client, command)
462 stdin.close()
462 stdin.close()
463
463
464 for line in stdout:
464 for line in stdout:
465 print(line, end='')
465 print(line, end='')
466
466
467 res = chan.recv_exit_status()
467 res = chan.recv_exit_status()
468
468
469 if res:
469 if res:
470 raise Exception('non-0 exit code updating working directory; %d' % res)
470 raise Exception('non-0 exit code updating working directory; %d' % res)
471
471
472
472
473 def synchronize_hg(
473 def synchronize_hg(
474 source_path: pathlib.Path, ec2_instance, revision: str = None
474 source_path: pathlib.Path, ec2_instance, revision: str = None
475 ):
475 ):
476 """Synchronize a local Mercurial source path to remote EC2 instance."""
476 """Synchronize a local Mercurial source path to remote EC2 instance."""
477
477
478 with tempfile.TemporaryDirectory() as temp_dir:
478 with tempfile.TemporaryDirectory() as temp_dir:
479 temp_dir = pathlib.Path(temp_dir)
479 temp_dir = pathlib.Path(temp_dir)
480
480
481 ssh_dir = temp_dir / '.ssh'
481 ssh_dir = temp_dir / '.ssh'
482 ssh_dir.mkdir()
482 ssh_dir.mkdir()
483 ssh_dir.chmod(0o0700)
483 ssh_dir.chmod(0o0700)
484
484
485 public_ip = ec2_instance.public_ip_address
485 public_ip = ec2_instance.public_ip_address
486
486
487 ssh_config = ssh_dir / 'config'
487 ssh_config = ssh_dir / 'config'
488
488
489 with ssh_config.open('w', encoding='utf-8') as fh:
489 with ssh_config.open('w', encoding='utf-8') as fh:
490 fh.write('Host %s\n' % public_ip)
490 fh.write('Host %s\n' % public_ip)
491 fh.write(' User hg\n')
491 fh.write(' User hg\n')
492 fh.write(' StrictHostKeyChecking no\n')
492 fh.write(' StrictHostKeyChecking no\n')
493 fh.write(' UserKnownHostsFile %s\n' % (ssh_dir / 'known_hosts'))
493 fh.write(' UserKnownHostsFile %s\n' % (ssh_dir / 'known_hosts'))
494 fh.write(' IdentityFile %s\n' % ec2_instance.ssh_private_key_path)
494 fh.write(' IdentityFile %s\n' % ec2_instance.ssh_private_key_path)
495
495
496 if not (source_path / '.hg').is_dir():
496 if not (source_path / '.hg').is_dir():
497 raise Exception(
497 raise Exception(
498 '%s is not a Mercurial repository; synchronization '
498 '%s is not a Mercurial repository; synchronization '
499 'not yet supported' % source_path
499 'not yet supported' % source_path
500 )
500 )
501
501
502 env = dict(os.environ)
502 env = dict(os.environ)
503 env['HGPLAIN'] = '1'
503 env['HGPLAIN'] = '1'
504 env['HGENCODING'] = 'utf-8'
504 env['HGENCODING'] = 'utf-8'
505
505
506 hg_bin = source_path / 'hg'
506 hg_bin = source_path / 'hg'
507
507
508 res = subprocess.run(
508 res = subprocess.run(
509 ['python2.7', str(hg_bin), 'log', '-r', revision, '-T', '{node}'],
509 ['python2.7', str(hg_bin), 'log', '-r', revision, '-T', '{node}'],
510 cwd=str(source_path),
510 cwd=str(source_path),
511 env=env,
511 env=env,
512 check=True,
512 check=True,
513 capture_output=True,
513 capture_output=True,
514 )
514 )
515
515
516 full_revision = res.stdout.decode('ascii')
516 full_revision = res.stdout.decode('ascii')
517
517
518 args = [
518 args = [
519 'python2.7',
519 'python2.7',
520 str(hg_bin),
520 str(hg_bin),
521 '--config',
521 '--config',
522 'ui.ssh=ssh -F %s' % ssh_config,
522 'ui.ssh=ssh -F %s' % ssh_config,
523 '--config',
523 '--config',
524 'ui.remotecmd=/hgdev/venv-bootstrap/bin/hg',
524 'ui.remotecmd=/hgdev/venv-bootstrap/bin/hg',
525 # Also ensure .hgtags changes are present so auto version
525 # Also ensure .hgtags changes are present so auto version
526 # calculation works.
526 # calculation works.
527 'push',
527 'push',
528 '-f',
528 '-f',
529 '-r',
529 '-r',
530 full_revision,
530 full_revision,
531 '-r',
531 '-r',
532 'file(.hgtags)',
532 'file(.hgtags)',
533 'ssh://%s//hgwork/src' % public_ip,
533 'ssh://%s//hgwork/src' % public_ip,
534 ]
534 ]
535
535
536 res = subprocess.run(args, cwd=str(source_path), env=env)
536 res = subprocess.run(args, cwd=str(source_path), env=env)
537
537
538 # Allow 1 (no-op) to not trigger error.
538 # Allow 1 (no-op) to not trigger error.
539 if res.returncode not in (0, 1):
539 if res.returncode not in (0, 1):
540 res.check_returncode()
540 res.check_returncode()
541
541
542 # TODO support synchronizing dirty working directory.
542 # TODO support synchronizing dirty working directory.
543
543
544 sftp = ec2_instance.ssh_client.open_sftp()
544 sftp = ec2_instance.ssh_client.open_sftp()
545
545
546 with sftp.open('/hgdev/hgup', 'wb') as fh:
546 with sftp.open('/hgdev/hgup', 'wb') as fh:
547 fh.write(HG_UPDATE_CLEAN)
547 fh.write(HG_UPDATE_CLEAN)
548 fh.chmod(0o0700)
548 fh.chmod(0o0700)
549
549
550 chan, stdin, stdout = exec_command(
550 chan, stdin, stdout = exec_command(
551 ec2_instance.ssh_client, '/hgdev/hgup %s' % full_revision
551 ec2_instance.ssh_client, '/hgdev/hgup %s' % full_revision
552 )
552 )
553 stdin.close()
553 stdin.close()
554
554
555 for line in stdout:
555 for line in stdout:
556 print(line, end='')
556 print(line, end='')
557
557
558 res = chan.recv_exit_status()
558 res = chan.recv_exit_status()
559
559
560 if res:
560 if res:
561 raise Exception(
561 raise Exception(
562 'non-0 exit code updating working directory; %d' % res
562 'non-0 exit code updating working directory; %d' % res
563 )
563 )
564
564
565
565
566 def run_tests(ssh_client, python_version, test_flags=None):
566 def run_tests(ssh_client, python_version, test_flags=None):
567 """Run tests on a remote Linux machine via an SSH client."""
567 """Run tests on a remote Linux machine via an SSH client."""
568 test_flags = test_flags or []
568 test_flags = test_flags or []
569
569
570 print('running tests')
570 print('running tests')
571
571
572 if python_version == 'system2':
572 if python_version == 'system2':
573 python = '/usr/bin/python2'
573 python = '/usr/bin/python2'
574 elif python_version == 'system3':
574 elif python_version == 'system3':
575 python = '/usr/bin/python3'
575 python = '/usr/bin/python3'
576 elif python_version.startswith('pypy'):
576 elif python_version.startswith('pypy'):
577 python = '/hgdev/pyenv/shims/%s' % python_version
577 python = '/hgdev/pyenv/shims/%s' % python_version
578 else:
578 else:
579 python = '/hgdev/pyenv/shims/python%s' % python_version
579 python = '/hgdev/pyenv/shims/python%s' % python_version
580
580
581 test_flags = ' '.join(shlex.quote(a) for a in test_flags)
581 test_flags = ' '.join(shlex.quote(a) for a in test_flags)
582
582
583 command = (
583 command = (
584 '/bin/sh -c "export TMPDIR=/hgwork/tmp; '
584 '/bin/sh -c "export TMPDIR=/hgwork/tmp; '
585 'cd /hgwork/src/tests && %s run-tests.py %s"' % (python, test_flags)
585 'cd /hgwork/src/tests && %s run-tests.py %s"' % (python, test_flags)
586 )
586 )
587
587
588 chan, stdin, stdout = exec_command(ssh_client, command)
588 chan, stdin, stdout = exec_command(ssh_client, command)
589
589
590 stdin.close()
590 stdin.close()
591
591
592 for line in stdout:
592 for line in stdout:
593 print(line, end='')
593 print(line, end='')
594
594
595 return chan.recv_exit_status()
595 return chan.recv_exit_status()
General Comments 0
You need to be logged in to leave comments. Login now