Running with gitlab-runner 14.1.0 (8925d9a0)  on gitlab-runners-bigbang-gl-packages-privileged-gitlab-runneb7dvc L_wAsvdS section_start:1630343779:resolve_secrets Resolving secrets section_end:1630343779:resolve_secrets section_start:1630343779:prepare_executor Preparing the "kubernetes" executor Using Kubernetes namespace: gitlab-runners Using Kubernetes executor with image registry.dso.mil/platform-one/big-bang/pipeline-templates/pipeline-templates/k3d-builder:0.0.5 ... Using attach strategy to execute scripts... section_end:1630343779:prepare_executor section_start:1630343779:prepare_script Preparing environment Waiting for pod gitlab-runners/runner-lwasvds-project-7074-concurrent-1s4phl to be running, status is Pending Running on runner-lwasvds-project-7074-concurrent-1s4phl via gitlab-runners-bigbang-gl-packages-privileged-gitlab-runneb7dvc... section_end:1630343786:prepare_script section_start:1630343786:get_sources Getting source from Git repository Fetching changes with git depth set to 50... Initialized empty Git repository in /builds/L_wAsvdS/1/platform-one/big-bang/apps/sandbox/vault/.git/ Created fresh repository. Checking out 4df9c21a as refs/merge-requests/7/head... Skipping Git submodules setup section_end:1630343787:get_sources section_start:1630343787:step_script Executing "step_script" stage of the job script $ echo -e "\e[0Ksection_start:`date +%s`:cluster_setup[collapsed=true]\r\e[0KCluster Setup" section_start:1630343787:cluster_setup[collapsed=true] Cluster Setup $ if [ -z ${PIPELINE_REPO_BRANCH} ]; then # collapsed multi-line command $ git clone -b ${PIPELINE_REPO_BRANCH} ${PIPELINE_REPO} ${PIPELINE_REPO_DESTINATION} Cloning into '../pipeline-repo'... $ source ${WAIT_PATH} $ i=0; while [ "$i" -lt 12 ]; do docker info &>/dev/null && break; sleep 5; i=$(( i + 1 )) ; done $ docker network create ${CI_JOB_ID} --driver=bridge -o "com.docker.network.driver.mtu"="1450" d7269784b6b43d520a6692f0632ac8790b431e0eae45c15a7341b363feeb775f $ k3d cluster create ${CI_JOB_ID} --config ${K3D_CONFIG_PATH} --network ${CI_JOB_ID} INFO[0000] Using config file ../pipeline-repo/jobs/k3d-ci/config.yaml INFO[0000] Prep: Network INFO[0000] Network with name '6081868' already exists with ID 'd7269784b6b43d520a6692f0632ac8790b431e0eae45c15a7341b363feeb775f' INFO[0000] Created volume 'k3d-6081868-images' INFO[0001] Creating node 'k3d-6081868-server-0' INFO[0002] Pulling image 'docker.io/rancher/k3s:v1.20.4-k3s1' INFO[0004] Creating LoadBalancer 'k3d-6081868-serverlb' INFO[0005] Pulling image 'docker.io/rancher/k3d-proxy:v4.3.0' INFO[0007] Starting cluster '6081868' INFO[0007] Starting servers... INFO[0007] Starting Node 'k3d-6081868-server-0' INFO[0013] Starting agents... INFO[0013] Starting helpers... INFO[0013] Starting Node 'k3d-6081868-serverlb' INFO[0013] (Optional) Trying to get IP of the docker host and inject it into the cluster as 'host.k3d.internal' for easy access INFO[0015] Successfully added host record to /etc/hosts in 2/2 nodes and to the CoreDNS ConfigMap INFO[0015] Cluster '6081868' created successfully! INFO[0015] --kubeconfig-update-default=false --> sets --kubeconfig-switch-context=false INFO[0015] You can now use it like this: kubectl config use-context k3d-6081868 kubectl cluster-info $ until kubectl get deployment coredns -n kube-system -o go-template='{{.status.availableReplicas}}' | grep -v -e ''; do sleep 1s; done 1 $ if [ ! -z ${PROJECT_NAME} ]; then # collapsed multi-line command namespace/vault created secret/private-registry created secret/private-registry-mil created $ echo -e "\e[0Ksection_end:`date +%s`:cluster_setup\r\e[0K" section_end:1630343825:cluster_setup  $ echo "Installing ${CI_PROJECT_NAME} from ${CI_DEFAULT_BRANCH}" Installing vault from master $ echo -e "\e[0Ksection_start:`date +%s`:package_checkout1[collapsed=true]\r\e[0KPackage Checkout" section_start:1630343825:package_checkout1[collapsed=true] Package Checkout $ git fetch && git checkout ${CI_DEFAULT_BRANCH} From https://repo1.dso.mil/platform-one/big-bang/apps/sandbox/vault * [new branch] deploy-vault -> origin/deploy-vault * [new branch] deploy-vault1 -> origin/deploy-vault1 * [new branch] dev-p1cyber-1429 -> origin/dev-p1cyber-1429 * [new branch] master -> origin/master Previous HEAD position was 4df9c21 add values.yaml Switched to a new branch 'master' Branch 'master' set up to track remote branch 'master' from 'origin'. $ echo -e "\e[0Ksection_end:`date +%s`:package_checkout1\r\e[0K" section_end:1630343826:package_checkout1  $ echo -e "\e[0Ksection_start:`date +%s`:dependency_up1[collapsed=true]\r\e[0KDependency Install and Wait" section_start:1630343826:dependency_up1[collapsed=true] Dependency Install and Wait $ if [ -f "tests/dependencies.yaml" ]; then # collapsed multi-line command $ if [ -f "tests/dependencies.yaml" ]; then # collapsed multi-line command $ echo -e "\e[0Ksection_end:`date +%s`:dependency_up1\r\e[0K" section_end:1630343826:dependency_up1  $ if [ ! -z ${PROJECT_NAME} ]; then # collapsed multi-line command Helm installing vault/chart into vault namespace using default values Release "vault" does not exist. Installing it now. Error: template: vault/templates/bigbang/virtualservice.yaml:1:40: executing "vault/templates/bigbang/virtualservice.yaml" at <.Values.istio.vault.enabled>: nil pointer evaluating interface {}.enabled section_end:1630343826:step_script section_start:1630343826:after_script Running after_script Running after script... $ if [ -e success ]; then # collapsed multi-line command Job Failed, cluster status: NAMESPACE NAME READY STATUS RESTARTS AGE kube-system pod/metrics-server-86cbb8457f-rnb8w 1/1 Running 0 9s kube-system pod/local-path-provisioner-5ff76fc89d-49jmf 1/1 Running 0 9s kube-system pod/coredns-854c77959c-4n9kb 1/1 Running 0 9s NAMESPACE NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE default service/kubernetes ClusterIP 10.43.0.1 443/TCP 26s kube-system service/kube-dns ClusterIP 10.43.0.10 53/UDP,53/TCP,9153/TCP 24s kube-system service/metrics-server ClusterIP 10.43.56.130 443/TCP 24s NAMESPACE NAME READY UP-TO-DATE AVAILABLE AGE kube-system deployment.apps/metrics-server 1/1 1 1 24s kube-system deployment.apps/local-path-provisioner 1/1 1 1 24s kube-system deployment.apps/coredns 1/1 1 1 24s NAMESPACE NAME DESIRED CURRENT READY AGE kube-system replicaset.apps/metrics-server-86cbb8457f 1 1 1 9s kube-system replicaset.apps/local-path-provisioner-5ff76fc89d 1 1 1 9s kube-system replicaset.apps/coredns-854c77959c 1 1 1 9s $ echo -e "\e[0Ksection_start:`date +%s`:cluster_clean[collapsed=true]\r\e[0KCluster Cleanup" section_start:1630343826:cluster_clean[collapsed=true] Cluster Cleanup $ k3d cluster delete ${CI_JOB_ID} INFO[0000] Deleting cluster '6081868' INFO[0000] Deleted k3d-6081868-serverlb INFO[0001] Deleted k3d-6081868-server-0 INFO[0001] Deleting image volume 'k3d-6081868-images' INFO[0001] Removing cluster details from default kubeconfig... INFO[0001] Removing standalone kubeconfig file (if there is one)... INFO[0001] Successfully deleted cluster 6081868! $ docker network rm ${CI_JOB_ID} 6081868 $ echo -e "\e[0Ksection_end:`date +%s`:cluster_clean\r\e[0K" section_end:1630343828:cluster_clean  section_end:1630343828:after_script section_start:1630343828:upload_artifacts_on_failure Uploading artifacts for failed job Uploading artifacts... WARNING: images.txt: no matching files  WARNING: tests/cypress/screenshots: no matching files WARNING: tests/cypress/videos: no matching files  WARNING: cypress-artifacts: no matching files  ERROR: No files to upload  section_end:1630343828:upload_artifacts_on_failure section_start:1630343828:cleanup_file_variables Cleaning up file based variables section_end:1630343829:cleanup_file_variables ERROR: Job failed: command terminated with exit code 1