diff --git a/.travis.yml b/.travis.yml index ea6f79db0..3377272cc 100644 --- a/.travis.yml +++ b/.travis.yml @@ -82,7 +82,7 @@ jobs: - ssh -o StrictHostKeyChecking=no root@$CLUSTER_IP "kubectl get nodes" - ssh -o StrictHostKeyChecking=no root@$CLUSTER_IP "mkdir /data;kubectl create -f /root/pravega-operator/test/e2e/resources/local-storage.yaml" - ssh -o StrictHostKeyChecking=no root@$CLUSTER_IP "curl -L https://raw.githubusercontent.com/helm/helm/master/scripts/get-helm-3 | bash" - - ssh -o StrictHostKeyChecking=no root@$CLUSTER_IP "helm repo add stable https://kubernetes-charts.storage.googleapis.com;helm install stable/nfs-server-provisioner --generate-name;kubectl -n default create -f /root/pravega-operator/test/e2e/resources/tier2.yaml" + - ssh -o StrictHostKeyChecking=no root@$CLUSTER_IP "helm repo add stable https://charts.helm.sh/stable;helm install stable/nfs-server-provisioner --generate-name;kubectl -n default create -f /root/pravega-operator/test/e2e/resources/tier2.yaml" - ssh -o StrictHostKeyChecking=no root@$CLUSTER_IP "kubectl create -f /root/pravega-operator/test/e2e/resources/zookeeper.yaml" - ssh -o StrictHostKeyChecking=no root@$CLUSTER_IP "kubectl -n default create -f /root/pravega-operator/test/e2e/resources/bookkeeper.yaml" diff --git a/README.md b/README.md index ded6eb88f..d87319836 100644 --- a/README.md +++ b/README.md @@ -74,7 +74,7 @@ Check out the available [options for long term storage](doc/longtermstorage.md) For demo purposes, you can quickly install a toy NFS server. ``` -$ helm repo add stable https://kubernetes-charts.storage.googleapis.com +$ helm repo add stable https://charts.helm.sh/stable $ helm repo update $ helm install stable/nfs-server-provisioner --generate-name ``` diff --git a/pkg/test/e2e/e2eutil/pravegacluster_util.go b/pkg/test/e2e/e2eutil/pravegacluster_util.go index 53db66d45..fb8f381fb 100644 --- a/pkg/test/e2e/e2eutil/pravegacluster_util.go +++ b/pkg/test/e2e/e2eutil/pravegacluster_util.go @@ -614,9 +614,13 @@ func RestartTier2(t *testing.T, f *framework.Framework, ctx *framework.TestCtx, t.Log("restarting tier2 storage") tier2 := NewTier2(namespace) - err := f.Client.Delete(goctx.TODO(), tier2) - if err != nil { - return fmt.Errorf("failed to delete tier2: %v", err) + _, err := f.KubeClient.CoreV1().PersistentVolumeClaims(namespace).Get(tier2.Name, metav1.GetOptions{}) + + if err == nil { + err := f.Client.Delete(goctx.TODO(), tier2) + if err != nil { + return fmt.Errorf("failed to delete tier2: %v", err) + } } err = wait.Poll(RetryInterval, 3*time.Minute, func() (done bool, err error) {