lbry-docker/contrib/k8s-lbry/run.sh

657 lines
22 KiB
Bash
Executable file

#!/bin/bash
BASEDIR=$(cd $(dirname "$0"); pwd)
BASENAME=$(basename "$0")
## RELEASE - The name for the helm release.
RELEASE=${RELEASE:-k8s-lbry}
## NAMESPACE - The Kubernetes namespace for the release
NAMESPACE=${NAMESPACE:-k8s-lbry}
## CHART - The helm chart location (local path, or from repository)
### Use the stable chart from the HELM_REPO repository:
CHART=${CHART:-lbry/k8s-lbry}
### Use the chart from the current directory:
#CHART=${CHART:-$BASEDIR}
## VALUES - The path to your configured helm values.yaml:
VALUES=${VALUES:-$BASEDIR/values-dev.yaml}
DEFAULT_VALUES_URL=${DEFAULT_VALUES_URL:-https://raw.githubusercontent.com/lbryio/lbry-docker/master/contrib/k8s-lbry/values.yaml}
## HELM_REPO - The stable helm chart repository for this chart:
HELM_REPO=${HELM_REPO:-https://k8s-lbry.sfo2.digitaloceanspaces.com}
## TILLER_HOST - The host that runs tiller
TILLER_HOST=${TILLER_HOST:-localhost}
## CERTMANAGER_NAMESPACE - The namespace cert-manager runs in
CERTMANAGER_NAMESPACE=${CERTMANAGER_NAMESPACE:-cert-manager}
CERTMANAGER_VERSION=v0.7.1
## NGINX_ENABLE_HTTPS - This needs to be set false until DNS is setup.
NGINX_ENABLE_HTTPS=${NGINX_ENABLE_HTTPS:-false}
## Mysql Database snapshot download url:
CHAINQUERY_SNAPSHOT_URL=${CHAINQUERY_SNAPSHOT_URL:-https://lbry-chainquery-mysql-dump.sfo2.digitaloceanspaces.com/chainquery_height_560900.mysql-backup.tar.gz}
## lbrycrd snapshot download url:
LBRYCRD_SNAPSHOT_URL=${LBRYCRD_SNAPSHOT_URL:-https://lbry-chainquery-mysql-dump.sfo2.digitaloceanspaces.com/lbrycrd-566588-snapshot.tar.gz}
LBRYCRD_RPC_USER=${LBRYCRD_RPC_USER:-lbry}
LBRYCRD_RPC_PASSWORD=${LBRYCRD_RPC_PASSWORD:-lbry}
## Bash alias name for run.sh
## defaults to $NAMESPACE (k8s-lbry)
RUN_ALIAS=${RUN_ALIAS:-$NAMESPACE}
## Package bucket (ignore this, only used by developer of this package)
PACKAGE_BUCKET=${PACKAGE_BUCKET:-"s3://k8s-lbry"}
exe() { ( echo "## $*"; $*; ) }
setup() {
### Check for external dependencies:
if ! which helm > /dev/null; then
echo "Error: You must install helm"
echo "On Ubuntu you can run: sudo snap install --classic helm"
echo "For other platforms, see https://github.com/helm/helm/releases/latest"
exit 1
fi
if ! which kubectl > /dev/null; then
echo "Error: You must install kubectl"
echo "On Ubuntu you can run: sudo snap install --classic kubectl"
echo "For other platforms, see https://kubernetes.io/docs/tasks/tools/install-kubectl/"
exit 1
fi
if ! which git > /dev/null; then
echo "Error: You must install git"
echo "On Ubuntu you can run: sudo apt install -y git"
echo "For other platforms, see https://git-scm.com/downloads"
exit 1
fi
### Initialize helm locally, but do not install tiller to the cluster:
HELM=$(which helm)
if [ ! -f "$HOME"/.helm/repository/repositories.yaml ]; then
exe "$HELM" init --client-only
fi
### Add the stable helm chart repository:
if [ "$CHART" != "$BASEDIR" ]; then
exe "$HELM" repo add lbry "$HELM_REPO"
exe "$HELM" repo update
fi
### Install helm-tiller plugin, so that no tiller needs to be installed to the cluster:
exe "$HELM" plugin install https://github.com/rimusz/helm-tiller || true
### Setup the values.yaml for the chart, using the VALUES environment variable or script default
### If no values file exists, interactively ask if a default config should be created in its place.
if [ ! -f "$VALUES" ]; then
echo ""
echo "Values file does not exist: $VALUES"
read -p "Would you like to create a default config file here? (y/N)" choice
echo ""
case "$choice" in
y|Y ) curl "$DEFAULT_VALUES_URL" -Lo "$VALUES"
echo "Default configuration created: $VALUES"
;;
* ) echo "You must create your own values file: $VALUES (using values.yaml as a template.)"
echo "Or set VALUES=/path/to/values.yaml before subsequent commands."
exit 1
;;
esac
else
echo "Configuration found: $VALUES"
fi
echo "Edit this config file to suit your own environment before install/upgrade"
}
helm() {
## Redefine all helm commands to run through local tiller instance
## https://rimusz.net/tillerless-helm
HELM=$(which helm)
exe "$HELM" tiller run "$NAMESPACE" -- helm "$*"
}
kubectl() {
## kubectl wrapper that defaults to k8s-lbry namespace, so you don't have to
## type as much, but still passes all the provided arguments on to kubectl.
## So you can still specify a different namespace, because the client args
## are applied last.
KUBECTL=$(which kubectl)
exe "$KUBECTL" --namespace "$NAMESPACE" "$*"
}
install-nginx-ingress() {
### Install nginx-ingress from stable helm repository
### See https://github.com/helm/charts/tree/master/stable/nginx-ingress
helm install stable/nginx-ingress --namespace "$NAMESPACE" --name nginx-ingress --set nginx-ingress.controller.service.enableHttps="$NGINX_ENABLE_HTTPS" --set tcp.9246="$NAMESPACE/lbrycrd:9246"
}
upgrade-nginx-ingress() {
### Upgrade nginx-ingress
helm upgrade nginx-ingress stable/nginx-ingress --namespace "$NAMESPACE" --set nginx-ingress.controller.service.enableHttps="$NGINX_ENABLE_HTTPS" --set tcp.9246="$NAMESPACE/lbrycrd:9246"
}
install-cert-manager() {
### Install cert-manager from jetstack helm repository
### See https://docs.cert-manager.io/en/latest/index.html
kubectl apply -f https://raw.githubusercontent.com/jetstack/cert-manager/release-0.7/deploy/manifests/00-crds.yaml
helm repo add jetstack https://charts.jetstack.io
helm repo update
helm install --name cert-manager --namespace "$CERTMANAGER_NAMESPACE" jetstack/cert-manager --version $CERTMANAGER_VERSION
kubectl label namespace "$CERTMANAGER_NAMESPACE" certmanager.k8s.io/disable-validation="true"
}
upgrade-cert-manager() {
### Upgrade cert-manager
helm upgrade cert-manager jetstack/cert-manager --namespace "$CERTMANAGER_NAMESPACE" --version $CERTMANAGER_VERSION
}
install() {
### Install the k8s-lbry helm chart
if [ ! -f "$VALUES" ]; then
echo "Could not find chart values file: $VALUES"
exit 1
fi
helm install "$CHART" --name "$RELEASE" --namespace="$NAMESPACE" -f "$VALUES"
}
upgrade() {
### Upgrade the k8s-lbry helm chart
if [ ! -f "$VALUES" ]; then
echo "Could not find chart values file: $VALUES"
exit 1
fi
helm upgrade "$RELEASE" "$CHART" --namespace="$NAMESPACE" -f "$VALUES"
}
shell() {
### Execute a shell in the running container with helm app name provided by first argument
(
if [ "$#" -eq 1 ]; then
KUBECTL=$(which kubectl)
POD=$($KUBECTL get --namespace "$NAMESPACE" pod -l app="$1" -o name | sed s+pod/++)
exe kubectl exec -it "$POD" /bin/bash
else
echo "Required arg: helm app name of pod to shell into"
fi
)
}
shell-pvc() {
### Start a utility container shell with an attached Persistent Volume Claim.
(
# From https://gist.github.com/yuanying/3aa7d59dcce65470804ab43def646ab6
IMAGE="ubuntu:18.04"
COMMAND="/bin/bash"
SUFFIX=$(date +%s | shasum | base64 | fold -w 10 | head -1 | tr '[:upper:]' '[:lower:]')
usage_exit() {
echo "Usage: $0 [-c command] [-i image] PVC ..." 1>&2
exit 1
}
if [ "$#" -ne 1 ]; then
usage_exit
fi
while getopts i:h OPT
do
case $OPT in
i) IMAGE=$OPTARG
;;
c) COMMAND=$OPTARG
;;
h) usage_exit
;;
\?) usage_exit
;;
esac
done
shift $(($OPTIND - 1))
VOL_MOUNTS=""
VOLS=""
COMMA=""
for i in $@
do
VOL_MOUNTS="${VOL_MOUNTS}${COMMA}{\"name\": \"${i}\",\"mountPath\": \"/pvcs/${i}\"}"
VOLS="${VOLS}${COMMA}{\"name\": \"${i}\",\"persistentVolumeClaim\": {\"claimName\": \"${i}\"}}"
COMMA=","
done
$(which kubectl) --namespace "$NAMESPACE" run -it --rm --restart=Never --image="${IMAGE}" pvc-mounter-"${SUFFIX}" --overrides "
{
\"spec\": {
\"hostNetwork\": true,
\"containers\":[
{
\"args\": [\"${COMMAND}\"],
\"stdin\": true,
\"tty\": true,
\"name\": \"pvc\",
\"image\": \"${IMAGE}\",
\"volumeMounts\": [
${VOL_MOUNTS}
]
}
],
\"volumes\": [
${VOLS}
]
}
}
" -- "${COMMAND}"
)
}
restart() {
### Restart the pod given by a helm app name
(
if [ "$#" -eq 1 ]; then
KUBECTL=$(which kubectl)
POD=$($KUBECTL get --namespace "$NAMESPACE" pod -l app="$1" -o name | sed s+pod/++)
exe kubectl delete pod "$POD"
else
echo "Required arg: helm app name of pod to restart"
fi
)
}
package() {
### Create a packaged helm release and upload to the S3 repository
(
cd $BASEDIR
set -e
if [ "$#" -eq 1 ]; then
if ! grep "version: $1" Chart.yaml; then
echo "Chart.yaml version does not match intended package version ($1)."
exit 1
fi
else
echo "required argument: package version"
exit 1
fi
PACKAGE="k8s-lbry-$1.tgz"
# Package release and rebuild repository
exe helm dependency update
exe helm package .
exe helm repo index --merge index.yaml .
## Replace URLs in index to point to github releases:
sed -i 's|- k8s-lbry-\(.*\).tgz$|- https://github.com/lbryio/lbry-docker/releases/download/k8s-lbry-\1/k8s-lbry-\1.tgz|g' index.yaml
echo "Created package: $PACKAGE"
echo "Next steps: "
echo " 1) Create new github release (https://github.com/lbryio/lbry-docker/releases) and upload $PACKAGE"
echo " 2) Commit index.yaml and push to github"
)
}
chainquery-mysql-client() {
### Access the mysql shell for chainquery
KUBECTL=$(which kubectl)
POD=$($KUBECTL -n "$NAMESPACE" get pod -l app=chainquery-mysql -o name | sed s+pod/++)
if [ ${#POD} -gt 0 ]; then
kubectl exec -it "$POD" -- mysql -u chainquery -pchainquery
else
echo "chainquery-mysql pod not found"
fi
}
speech-mysql-client() {
### Access the mysql shell for speech
KUBECTL=$(which kubectl)
POD=$($KUBECTL -n "$NAMESPACE" get pod -l app=speech-mysql -o name | sed s+pod/++)
if [ ${#POD} -gt 0 ]; then
kubectl exec -it "$POD" -- mysql -u speech -pspeech
else
echo "speech-mysql pod not found"
fi
}
chainquery-override-snapshot() {
### Delete the existing chainquery database and download a snapshot to restore
read -p "Would you like to DESTROY the existing chainquery database,
and restore from a fresh snapshot? (y/N) " destroy_chainquery
case "$destroy_chainquery" in
y|Y )
kubectl delete deployments chainquery chainquery-mysql || true
echo "Please wait.."
IMAGE="ubuntu:18.04"
SUFFIX=$(date +%s | shasum | base64 | fold -w 10 | head -1 | tr '[:upper:]' '[:lower:]')
VOL_MOUNTS="{\"name\": \"chainquery-mysql\",\"mountPath\": \"/pvcs/chainquery-mysql\"}"
VOLS="{\"name\": \"chainquery-mysql\",\"persistentVolumeClaim\": {\"claimName\": \"chainquery-mysql\"}}"
COMMAND="rm -rf /pvcs/chainquery-mysql/* && apt-get update && apt-get install -y curl && curl -s ${CHAINQUERY_SNAPSHOT_URL} | tar xvz -C /pvcs/chainquery-mysql/"
$(which kubectl) --namespace "$NAMESPACE" run -it --rm --restart=Never --image=${IMAGE} pvc-mounter-"${SUFFIX}" --overrides "
{
\"spec\": {
\"hostNetwork\": true,
\"containers\":[
{
\"args\": [\"bin/bash\", \"-c\", \"${COMMAND}\"],
\"stdin\": true,
\"tty\": true,
\"name\": \"pvc\",
\"image\": \"${IMAGE}\",
\"volumeMounts\": [
${VOL_MOUNTS}
]
}
],
\"volumes\": [
${VOLS}
]
}
}
"
echo "Extraction complete"
;;
* ) echo "Aborted."
;;
esac
}
lbrycrd-override-snapshot() {
### Delete the existing lbrycrd data and download a snapshot to restore
read -p "Would you like to DESTROY the existing lbrycrd data,
and restore from a fresh snapshot? (y/N) " destroy_lbrycrd
case "$destroy_lbrycrd" in
y|Y )
kubectl delete deployments lbrycrd || true
echo "Please wait.."
IMAGE="ubuntu:18.04"
SUFFIX=$(date +%s | shasum | base64 | fold -w 10 | head -1 | tr '[:upper:]' '[:lower:]')
VOL_MOUNTS="{\"name\": \"lbrycrd\",\"mountPath\": \"/pvcs/lbrycrd\"}"
VOLS="{\"name\": \"lbrycrd\",\"persistentVolumeClaim\": {\"claimName\": \"lbrycrd\"}}"
COMMAND="rm -rf /pvcs/lbrycrd/* && apt-get update && apt-get install -y curl && curl -s ${LBRYCRD_SNAPSHOT_URL} | tar xvz -C /pvcs/lbrycrd/"
$(which kubectl) --namespace "$NAMESPACE" run -it --rm --restart=Never --image=${IMAGE} pvc-mounter-"${SUFFIX}" --overrides "
{
\"spec\": {
\"hostNetwork\": true,
\"containers\":[
{
\"args\": [\"bin/bash\", \"-c\", \"${COMMAND}\"],
\"stdin\": true,
\"tty\": true,
\"name\": \"pvc\",
\"image\": \"${IMAGE}\",
\"volumeMounts\": [
${VOL_MOUNTS}
]
}
],
\"volumes\": [
${VOLS}
]
}
}
"
echo "Extraction complete"
;;
* ) echo "Aborted."
;;
esac
}
logs() {
### Watch the logs of a pod by helm app name
(
set -e
if [ "$#" -eq 1 ]; then
kubectl logs -l app="$1" -f
elif [ "$#" -eq 2 ]; then
KUBECTL=$(which kubectl)
POD=$($KUBECTL get --namespace "$NAMESPACE" pod -l app="$1" -o name | sed s+pod/++)
kubectl logs "$POD" "$2" -f
else
echo "Required arg: app_name"
fi
)
}
lbrynet-copy-wallet() {
### Copy the lbrynet wallet to a local path for backup
(
set -e
if [ "$#" -eq 1 ]; then
WALLET=/home/lbrynet/.local/share/lbry/lbryum/wallets/default_wallet
KUBECTL=$(which kubectl)
POD=$($KUBECTL -n "$NAMESPACE" get pod -l app=lbrynet -o name | sed s+pod/++)
kubectl cp "$POD":$WALLET "$1"
chmod 600 "$1"
echo "lbrynet wallet copied to $1"
else
echo "Required arg: path of backup location for wallet"
fi
)
}
speech-lbrynet-copy-wallet() {
### Copy the speech-lbrynet wallet to a local path for backup
(
set -e
if [ "$#" -eq 1 ]; then
WALLET=/home/lbrynet/.local/share/lbry/lbryum/wallets/default_wallet
KUBECTL=$(which kubectl)
POD=$($KUBECTL -n "$NAMESPACE" get pod -l app=speech -o name | sed s+pod/++)
kubectl cp "$POD":$WALLET "$1" -c speech-lbrynet
chmod 600 "$1"
echo "lbrynet wallet copied to $1"
else
echo "Required arg: path of backup location for wallet"
fi
)
}
lbrycrd-cli() {
## Run lbrycrd-cli client from inside the running pod outputting to your local terminal
KUBECTL=$(which kubectl)
POD=$($KUBECTL -n "$NAMESPACE" get pod -l app=lbrycrd -o name | sed s+pod/++)
if [ ${#POD} -gt 0 ]; then
kubectl exec "$POD" -- lbrycrd-cli -rpcuser="$LBRYCRD_RPC_USER" -rpcpassword="$LBRYCRD_RPC_PASSWORD" "$*"
else
echo "lbrycrd pod not found"
fi
}
lbrynet() {
## Run lbrynet client from inside the running pod outputting to your local terminal
KUBECTL=$(which kubectl)
POD=$($KUBECTL -n "$NAMESPACE" get pod -l app=lbrynet -o name | sed s+pod/++)
if [ ${#POD} -gt 0 ]; then
kubectl exec "$POD" -- lbrynet "$*"
else
echo "lbrynet pod not found"
fi
}
speech-lbrynet() {
## Run lbrynet client from inside the running pod outputting to your local terminal
KUBECTL=$(which kubectl)
POD=$($KUBECTL -n "$NAMESPACE" get pod -l app=speech -o name | sed s+pod/++)
if [ ${#POD} -gt 0 ]; then
kubectl exec "$POD" -c speech-lbrynet -- lbrynet "$*"
else
echo "lbrynet pod not found"
fi
}
SUBCOMMANDS_NO_ARGS=(setup install install-nginx-ingress install-cert-manager upgrade
upgrade-nginx-ingress upgrade-cert-manager chainquery-mysql-client
speech-mysql-client chainquery-override-snapshot lbrycrd-override-snapshot
setup-alias)
SUBCOMMANDS_PASS_ARGS=(helm kubectl shell shell-pvc restart package logs lbrynet-copy-wallet lbrynet speech-lbrynet-copy-wallet speech-lbrynet lbrycrd-cli completion)
completion() {
if [ "$#" -eq 1 ] && [ "$1" == "bash" ]; then
cat <<EOF
__delegate_k8s_lbry() {
alias kubectl="kubectl --namespace=$NAMESPACE"
local cur subs
cur="\${COMP_WORDS[COMP_CWORD]}" # partial word, if any
subs="${SUBCOMMANDS_NO_ARGS[*]} ${SUBCOMMANDS_PASS_ARGS[*]}"
if [[ \${COMP_CWORD} -gt 1 ]]; then
# complete subcommands
_command \${@: -1}
else
# complete with the list of subcommands
COMPREPLY=( \$(compgen -W "\${subs}" -- \${cur}) )
fi
}
if [[ $(type -t compopt) = "builtin" ]]; then
complete -o default -F __delegate_k8s_lbry $RUN_ALIAS
else
complete -o default -o nospace -F __delegate_k8s_lbry $RUN_ALIAS
fi
EOF
else
echo "## I only know how to do completion for the bash shell."
echo "## Try '$0 completion bash' instead."
fi
}
setup-alias() {
if [[ $SHELL != */bash ]]; then
echo "It looks like you are currently running $SHELL";
echo "This tool only supports bash."
echo ""
echo "You will need to setup an alias in your own shell called \"$RUN_ALIAS\" for $BASEDIR/run.sh"
echo ""
read -p "Would you like to setup the alias for bash anyway? (Y/n)" choice
case "$choice" in
y|Y )
echo "Note: You will need to run bash as a subshell before running $RUN_ALIAS"
;;
* ) echo "Aborting" && exit 1
;;
esac
echo ""
fi
$(which kubectl) completion bash > "$BASEDIR"/completion.bash.inc
$(which helm) completion bash >> "$BASEDIR"/completion.bash.inc
completion bash >> "$BASEDIR"/completion.bash.inc
if [[ -z $K8S_LBRY_HOME ]] && ! grep "K8S_LBRY_HOME" "$HOME"/.bashrc > /dev/null; then
echo "K8S_LBRY_HOME not set."
read -p "Would you this script to edit $HOME/.bashrc to add tab completion support? (y/N) " choice
case "$choice" in
y|Y )
cat <<EOF >> "$HOME"/.bashrc
## Enable bash completion
if [ -f /etc/bash_completion ]; then
source /etc/bash_completion
fi
## k8s-lbry alias and tab completion
K8S_LBRY_HOME=$BASEDIR
alias $RUN_ALIAS=\$K8S_LBRY_HOME/run.sh
if [ -f \$K8S_LBRY_HOME/completion.bash.inc ]; then
source \$K8S_LBRY_HOME/completion.bash.inc
fi
EOF
echo "Created new alias: $RUN_ALIAS"
echo "To use the new alias, run \"source ~/.bashrc\" or just close your terminal session and restart it."
;;
* ) echo "Aborting" && exit 1;;
esac
else
echo "K8S_LBRY_HOME environment already setup. Nothing left to do."
fi
}
if printf '%s\n' ${SUBCOMMANDS_NO_ARGS[@]} | grep -q -P "^$1$"; then
## Subcommands that take no arguments:
(
set -e
if [ "$#" -eq 1 ]; then
$*
else
echo "$1 does not take any additional arguments"
fi
)
elif printf '%s\n' ${SUBCOMMANDS_PASS_ARGS[@]} | grep -q -P "^$1$"; then
## Subcommands that pass all arguments:
(
set -e
$*
)
else
if [[ $# -gt 0 ]]; then
echo "## Invalid command: $1"
else
echo "## Must specify a command:"
fi
echo ""
echo "## $0 setup"
echo "## - Setup dependencies"
echo ""
echo "## $0 install-nginx-ingress"
echo "## - Deploy nginx-ingress chart"
echo ""
echo "## $0 install-cert-manager"
echo "## - Deploy cert-manager chart"
echo ""
echo "## $0 install"
echo "## - Deploy main k8s-lbry chart"
echo ""
echo "## $0 upgrade"
echo "## - Upgrade an existing release"
echo ""
echo "## $0 shell <app>"
echo "## - execute shell into running helm application pod"
echo ""
echo "## $0 shell-pvc [-c command] [-i image] PVC"
echo "## - run a utility shell with the named PVC mounted in /pvcs"
echo ""
echo "## $0 helm <cmd> [...] "
echo "## - run any helm command (through helm-tiller wrapper)"
echo ""
echo "## $0 kubectl <cmd> [...]"
echo "## - run any kubectl command (defaulting to configured namespace)"
echo ""
echo "## $0 chainquery-mysql-client"
echo "## - run mysql shell for chainquery database"
echo ""
echo "## $0 speech-mysql-client"
echo "## - run mysql shell for speech database"
echo ""
echo "## $0 chainquery-override-snapshot"
echo "## - Restore chainquery database from snapshot"
echo ""
echo "## $0 lbrycrd-override-snapshot"
echo "## - Restore lbrycrd database from snapshot"
echo ""
echo "## $0 logs <app> [container]"
echo "## - Stream the logs for the pod running the helm app name provided"
echo "## (specify which container if the pod has more than one.)"
echo ""
echo "## $0 lbrynet-copy-wallet <local-path>"
echo "## - Backup the lbrynet wallet file to a local path"
echo ""
echo "## $0 lbrynet <args ... >"
echo "## - Run lbrynet client inside running lbrynet pod"
echo ""
echo "## $0 speech-lbrynet-copy-wallet <local-path>"
echo "## - Backup the speech-lbrynet wallet file to a local path"
echo ""
echo "## $0 speech-lbrynet <args ... >"
echo "## - Run speech-lbrynet client inside running speech pod"
echo ""
echo "## $0 setup-alias"
echo "## - Setup bash alias and tab completion for run.sh"
echo ""
exit 1
fi