[ARVADOS] created: 2.1.0-817-g0532a70b5
Git user
git at public.arvados.org
Wed May 26 14:33:19 UTC 2021
at 0532a70b56f2e8d93bcf0db7e0f63f7104507c1c (commit)
commit 0532a70b56f2e8d93bcf0db7e0f63f7104507c1c
Author: Javier Bértoli <jbertoli at curii.com>
Date: Wed May 26 11:32:26 2021 -0300
17603: Webshell is just a nginx vhost
Arvados-DCO-1.1-Signed-off-by: Javier Bértoli <jbertoli at curii.com>
diff --git a/tools/salt-install/provision.sh b/tools/salt-install/provision.sh
index 32fcb4513..89616528d 100755
--- a/tools/salt-install/provision.sh
+++ b/tools/salt-install/provision.sh
@@ -430,7 +430,10 @@ else
if [ "x${USE_LETSENCRYPT}" = "xyes" ]; then
grep -q "letsencrypt" ${S_DIR}/top.sls || echo " - letsencrypt" >> ${S_DIR}/top.sls
fi
- grep -q "arvados.${R}" ${S_DIR}/top.sls || echo " - arvados.${R}" >> ${S_DIR}/top.sls
+ # webshell role is just a nginx vhost, so it has no state
+ if [ "${R}" != "webshell" ]; then
+ grep -q "arvados.${R}" ${S_DIR}/top.sls || echo " - arvados.${R}" >> ${S_DIR}/top.sls
+ fi
# Pillars
grep -q "nginx_passenger" ${P_DIR}/top.sls || echo " - nginx_passenger" >> ${P_DIR}/top.sls
grep -q "nginx_${R}_configuration" ${P_DIR}/top.sls || echo " - nginx_${R}_configuration" >> ${P_DIR}/top.sls
commit a766778e0f838f9713ec9f4eb048b80acf5cd97f
Author: Javier Bértoli <jbertoli at curii.com>
Date: Tue May 25 14:21:39 2021 -0300
17603: Separate shell's IP from webshell's
Arvados-DCO-1.1-Signed-off-by: Javier Bértoli <jbertoli at curii.com>
diff --git a/tools/salt-install/config_examples/multi_host/aws/states/host_entries.sls b/tools/salt-install/config_examples/multi_host/aws/states/host_entries.sls
index 82fb6f4ec..6e0deb49c 100644
--- a/tools/salt-install/config_examples/multi_host/aws/states/host_entries.sls
+++ b/tools/salt-install/config_examples/multi_host/aws/states/host_entries.sls
@@ -52,9 +52,15 @@ extra_extra_hosts_entries_etc_hosts_keepweb_host_present:
- download.{{ arvados.cluster.name }}.{{ arvados.cluster.domain }}
- collections.{{ arvados.cluster.name }}.{{ arvados.cluster.domain }}
-extra_extra_hosts_entries_etc_hosts_shell_host_present:
+extra_extra_hosts_entries_etc_hosts_webshell_host_present:
host.present:
- ip: __WEBSHELL_INT_IP__
+ - names:
+ - webshell.{{ arvados.cluster.name }}.{{ arvados.cluster.domain }}
+
+extra_extra_hosts_entries_etc_hosts_shell_host_present:
+ host.present:
+ - ip: __SHELL_INT_IP__
- names:
- shell.{{ arvados.cluster.name }}.{{ arvados.cluster.domain }}
diff --git a/tools/salt-install/local.params.example.multiple_hosts b/tools/salt-install/local.params.example.multiple_hosts
index 5dcc1d7c9..af2de2e72 100644
--- a/tools/salt-install/local.params.example.multiple_hosts
+++ b/tools/salt-install/local.params.example.multiple_hosts
@@ -40,6 +40,7 @@ WORKBENCH1_INT_IP=10.0.0.5
WORKBENCH2_INT_IP=10.0.0.5
WEBSHELL_INT_IP=10.0.0.5
DATABASE_INT_IP=10.0.0.6
+SHELL_INT_IP=10.0.0.7
INITIAL_USER="admin"
INITIAL_USER_PASSWORD="password"
diff --git a/tools/salt-install/provision.sh b/tools/salt-install/provision.sh
index 0ecb7217d..32fcb4513 100755
--- a/tools/salt-install/provision.sh
+++ b/tools/salt-install/provision.sh
@@ -273,6 +273,7 @@ for f in $(ls "${SOURCE_PILLARS_DIR}"/*); do
s#__KEEPSTORE1_INT_IP__#${KEEPSTORE1_INT_IP}#g;
s#__KEEPWEB_INT_IP__#${KEEPWEB_INT_IP}#g;
s#__WEBSHELL_INT_IP__#${WEBSHELL_INT_IP}#g;
+ s#__SHELL_INT_IP__#${SHELL_INT_IP}#g;
s#__WORKBENCH1_INT_IP__#${WORKBENCH1_INT_IP}#g;
s#__WORKBENCH2_INT_IP__#${WORKBENCH2_INT_IP}#g;
s#__DATABASE_INT_IP__#${DATABASE_INT_IP}#g;
commit ff5f0cdbc9f6861cde5836cdc53b662adda27271
Author: Javier Bértoli <jbertoli at curii.com>
Date: Thu May 20 16:37:01 2021 -0300
17603: Reorder example IPs so they match the documentation
Arvados-DCO-1.1-Signed-off-by: Javier Bértoli <jbertoli at curii.com>
diff --git a/tools/salt-install/local.params.example.multiple_hosts b/tools/salt-install/local.params.example.multiple_hosts
index 6a5bc175c..5dcc1d7c9 100644
--- a/tools/salt-install/local.params.example.multiple_hosts
+++ b/tools/salt-install/local.params.example.multiple_hosts
@@ -26,17 +26,20 @@ WORKBENCH2_EXT_SSL_PORT=443
# Internal IPs for the configuration
CLUSTER_INT_CIDR=10.0.0.0/16
+
+# Note the IPs in this example are shared between roles, as suggested in
+# https://doc.arvados.org/main/install/salt-multi-host.html
CONTROLLER_INT_IP=10.0.0.1
WEBSOCKET_INT_IP=10.0.0.1
KEEP_INT_IP=10.0.0.2
+# Both for collections and downloads
+KEEPWEB_INT_IP=10.0.0.2
KEEPSTORE0_INT_IP=10.0.0.3
KEEPSTORE1_INT_IP=10.0.0.4
-# Both for collections and downloads
-KEEPWEB_INT_IP=10.0.0.5
-WEBSHELL_INT_IP=10.0.0.6
-WORKBENCH1_INT_IP=10.0.0.7
-WORKBENCH2_INT_IP=10.0.0.7
-DATABASE_INT_IP=10.0.0.8
+WORKBENCH1_INT_IP=10.0.0.5
+WORKBENCH2_INT_IP=10.0.0.5
+WEBSHELL_INT_IP=10.0.0.5
+DATABASE_INT_IP=10.0.0.6
INITIAL_USER="admin"
INITIAL_USER_PASSWORD="password"
commit 35bf3afb8b312dcc30fe3b61042779d33bf1db10
Author: Javier Bértoli <jbertoli at curii.com>
Date: Tue May 18 13:04:56 2021 -0300
17603: Enable webshell role pillars in provision script
refs #17603
Arvados-DCO-1.1-Signed-off-by: Javier Bértoli <jbertoli at curii.com>
diff --git a/doc/install/salt-multi-host.html.textile.liquid b/doc/install/salt-multi-host.html.textile.liquid
index 827d65db2..04ef9e868 100644
--- a/doc/install/salt-multi-host.html.textile.liquid
+++ b/doc/install/salt-multi-host.html.textile.liquid
@@ -39,6 +39,7 @@ We suggest distributing the Arvados components in the following way, creating at
# WORKBENCH node:
## arvados workbench
## arvados workbench2
+## arvados webshell
# KEEPPROXY node:
## arvados keepproxy
## arvados keepweb
@@ -145,7 +146,7 @@ ssh user at host sudo ./provision.sh --config local.params --roles keepstore
#. Workbench
<notextile>
<pre><code>scp -r provision.sh local* user at host:
-ssh user at host sudo ./provision.sh --config local.params --roles workbench,workbench2
+ssh user at host sudo ./provision.sh --config local.params --roles workbench,workbench2,webshell
</code></pre>
</notextile>
diff --git a/tools/salt-install/provision.sh b/tools/salt-install/provision.sh
index 5cef4d0fc..0ecb7217d 100755
--- a/tools/salt-install/provision.sh
+++ b/tools/salt-install/provision.sh
@@ -28,14 +28,15 @@ usage() {
echo >&2 " Possible values are:"
echo >&2 " api"
echo >&2 " controller"
+ echo >&2 " dispatcher"
+ echo >&2 " keepproxy"
echo >&2 " keepstore"
- echo >&2 " websocket"
echo >&2 " keepweb"
- echo >&2 " workbench2"
- echo >&2 " keepproxy"
echo >&2 " shell"
+ echo >&2 " webshell"
+ echo >&2 " websocket"
echo >&2 " workbench"
- echo >&2 " dispatcher"
+ echo >&2 " workbench2"
echo >&2 " Defaults to applying them all"
echo >&2 " -h, --help Display this help and exit"
echo >&2 " -v, --vagrant Run in vagrant and use the /vagrant shared dir"
@@ -70,7 +71,7 @@ arguments() {
for i in ${2//,/ }
do
# Verify the role exists
- if [[ ! "database,api,controller,keepstore,websocket,keepweb,workbench2,keepproxy,shell,workbench,dispatcher" == *"$i"* ]]; then
+ if [[ ! "database,api,controller,keepstore,websocket,keepweb,workbench2,webshell,keepproxy,shell,workbench,dispatcher" == *"$i"* ]]; then
echo "The role '${i}' is not a valid role"
usage
exit 1
@@ -422,7 +423,7 @@ else
grep -q "nginx_passenger" ${P_DIR}/top.sls || echo " - nginx_passenger" >> ${P_DIR}/top.sls
grep -q "nginx_${R}_configuration" ${P_DIR}/top.sls || echo " - nginx_${R}_configuration" >> ${P_DIR}/top.sls
;;
- "controller" | "websocket" | "workbench" | "workbench2" | "keepweb" | "keepproxy")
+ "controller" | "websocket" | "workbench" | "workbench2" | "webshell" | "keepweb" | "keepproxy")
# States
grep -q "nginx.passenger" ${S_DIR}/top.sls || echo " - nginx.passenger" >> ${S_DIR}/top.sls
if [ "x${USE_LETSENCRYPT}" = "xyes" ]; then
@@ -443,7 +444,6 @@ else
grep -q "arvados.${R}" ${S_DIR}/top.sls || echo " - arvados.${R}" >> ${S_DIR}/top.sls
# Pillars
grep -q "" ${P_DIR}/top.sls || echo " - docker" >> ${P_DIR}/top.sls
- grep -q "nginx_webshell_configuration" ${P_DIR}/top.sls || echo " - nginx_webshell_configuration" >> ${P_DIR}/top.sls
;;
"dispatcher")
# States
-----------------------------------------------------------------------
hooks/post-receive
--
More information about the arvados-commits
mailing list