forked from HariSekhon/DevOps-Bash-tools
-
Notifications
You must be signed in to change notification settings - Fork 0
/
.envrc-gcp
146 lines (124 loc) · 4.85 KB
/
.envrc-gcp
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
#!/usr/bin/env bash
# vim:ts=4:sts=4:sw=4:et
#
# Author: Hari Sekhon
# Date: Mon Feb 22 17:42:01 2021 +0000
#
# https://github.com/HariSekhon/DevOps-Bash-tools
#
# License: see accompanying Hari Sekhon LICENSE file
#
# If you're using my code you're welcome to connect with me on LinkedIn and optionally send me feedback to help steer this or other code I publish
#
# https://www.linkedin.com/in/HariSekhon
#
# ============================================================================ #
# G C P D i r E n v
# ============================================================================ #
# https://direnv.net/man/direnv-stdlib.1.html
# See Also:
#
# .envrc
# .envrc-aws
# .envrc-kubernetes
# direnv stdlib - loads .envrc from parent dir up to /
#
# useful to accumulate parent and child directory .envrc settings eg. adding Kubernetes namespace, ArgoCD app etc.
#
# bypasses security authorization though - use with care
#source_up
#
# source_up must be loaded before set -u otherwise gets this error:
#
# direnv: loading .envrc
# /bin/bash: line 226: $1: unbound variable
set -euo pipefail
[ -n "${DEBUG:-}" ] && set -x
srcdir="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
# https://cloud.google.com/sdk/gcloud/reference/config
# If using other services, infer the environment variables to put below by reading:
#
# gcloud topic configurations
# or
# gcloud config set --help
gcloud_config(){
local config="${1:-}"
if [ -z "$config" ]; then
echo "no config passed to gcloud_config() function" >&2
return 1
fi
if [ -z "${CI:-}" ]; then
return
fi
# don't waste time if not using GCloud SDK, ie. not found in $PATH
if type -P gcloud; then
# protect from setting this if the config does exist as this can cause auth problems by unsetting the core.account
if gcloud config configurations list --format='get(name)' | grep -q "^$config$"; then
export CLOUDSDK_ACTIVE_CONFIG_NAME="$config"
fi
fi
}
#gcloud_config dev
#gcloud_config staging
#gcloud_config production
# XXX: Edit
export CLOUDSDK_CORE_PROJECT=myproject
# XXX: Edit
export REGION=europe-west2
export CLOUDSDK_COMPUTE_REGION="$REGION"
# you should probably not set CLOUDSDK_COMPUTE_ZONE
#
# 'gcloud compute ssh' will auto-determine the zone
#
# setting CLOUDSDK_COMPUTE_ZONE explicitly breaks the above command in 2/3 cases due to a VM being in a different zone:
#
# ERROR: (gcloud.compute.ssh) Could not fetch resource:
# - The resource 'projects/<MY_PROJECT>/zones/<ZONE>/instances/<VM_NAME>' was not found
#
# gcp/gce_ssh.sh script in this repo can work around that if you do set this
#
#export CLOUDSDK_COMPUTE_ZONE="${REGION}-a" # or b or c
export CLOUDSDK_AI_REGION="$REGION"
export CLOUDSDK_AI_PLATFORM_REGION="$REGION"
export CLOUDSDK_DATAPROC_REGION="$REGION"
export CLOUDSDK_DEPLOY_REGION="$REGION"
export CLOUDSDK_FILESTORE_REGION="$REGION"
export CLOUDSDK_FUNCTIONS_REGION="$REGION"
export CLOUDSDK_MEMCACHE_REGION="$REGION"
export CLOUDSDK_REDIS_REGION="$REGION"
export CLOUDSDK_RUN_REGION="$REGION"
export CLOUDSDK_RUN_CLUSTER_LOCATION="$REGION"
export CLOUDSDK_VMWARE_REGION="$REGION"
# XXX: Edit
export CLOUDSDK_RUN_PLATFORM=managed
#export CLOUDSDK_RUN_PLATFORM=gke
#export CLOUDSDK_RUN_PLATFORM=kubernetes
#export CLOUDSDK_RUN_CLUSTER=mycluster
export CLOUDSDK_GCLOUDIGNORE_ENABLED=True
#export CLOUDSDK_BUILDS_USE_KANIKO=True
# XXX: Edit, or remove if only have 1 cluster in project, will auto-determine below
export CLOUDSDK_CONTAINER_CLUSTER=mycluster # GKE cluster name
# If CLOUDSDK_CONTAINNER_CLUSTER isn't set and there is only one EKS cluster in this account and region, then use it - smart, but slower, prefer setting it explicitly for speed
if [ -z "${CLOUDSDK_CONTAINER_CLUSTER:-}" ]; then
gke_clusters=()
while IFS='' read -r line; do
gke_clusters+=("$line")
done < <(gcloud container clusters list --format='get(name)')
if [ "${#gke_clusters[@]}" -eq 1 ]; then
export CLOUDSDK_CONTAINER_CLUSTER="${gke_clusters[*]}"
fi
fi
# alternatively call gke_kube_context() function in .envrc-kubernetes which will do this
# and comment out auto-running kube_context() on sourcing .envrc-kubernetes
if [ -n "${CLOUDSDK_CONTAINER_CLUSTER:-}" ]; then
# kubectl context is easily created by running adjacent aws_kube_creds.sh script first
export GKE_CONTEXT="gke_${CLOUDSDK_CORE_PROJECT}_${CLOUDSDK_COMPUTE_REGION}_${CLOUDSDK_CONTAINER_CLUSTER}"
# XXX: safer to inline .envrc-kubernetes if you're worried about changes to it bypassing 'direnv allow' authorization
# shellcheck disable=SC1090,SC1091
. "$srcdir/.envrc-kubernetes" "$GKE_CONTEXT"
fi
# pull the secret using this command whenever you need it:
#
# gcp_secret_get.sh "$JENKINS_ADMIN_PASSWORD_GCP_SECRET" | copy_to_clipboard.sh
#
export JENKINS_ADMIN_PASSWORD_GCP_SECRET="jenkins-admin-password"