Does anyone have any advice for stork pods that keep restarting
Does anyone have any advice for stork pods that keep restarting with: ``` Observed a panic: "invalid memory address or nil pointer dereference" (runtime error: invalid memory address or nil pointer dereference) panic: runtime error: invalid memory address or nil pointer dereference [signal SIGSEGV: segmentation violation code=0x1 addr=0x28 pc=0x29aad38]``` I'm running storkctl Version: 2.7.0-2e5098a and k8s version 1.22.5Solved51Views0likes14CommentsLooking for a tool / script that will take a vmware snapshot
Looking for help. I have looked around but cant find it. Looking for a tool / script that will take a vmware snapshot, then a pure snap shot and then delete the vmware snapshot. I started writing a script but thought I would ask the question first. Thanks26Views0likes4CommentsHey I observed the following issue with pso which renders it unusable in 25 of our customer clusters
We can still not upgrade to portworx because of missing features, ```E0105 1755.820180 1 runtime.go:69] Observed a panic: "invalid memory address or nil pointer dereference" (runtime error: invalid memory address or nil pointer dereference) /go/src/pso.purestorage.com/cockroach-operator/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:76 /go/src/pso.purestorage.com/cockroach-operator/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:65 /go/src/pso.purestorage.com/cockroach-operator/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:51 /usr/local/go/src/runtime/panic.go:969 /usr/local/go/src/runtime/panic.go:212 /usr/local/go/src/runtime/signal_unix.go:742 /go/src/pso.purestorage.com/cockroach-operator/pkg/apis/pso/v1alpha1/intrusion_types.go:232 /go/src/pso.purestorage.com/cockroach-operator/pkg/apis/pso/v1alpha1/intrusion_types.go:209 /go/src/pso.purestorage.com/cockroach-operator/pkg/controller/cockroach/intrusion/intrusion_controller.go:444 /go/src/pso.purestorage.com/cockroach-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:215 /go/src/pso.purestorage.com/cockroach-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:158 /go/src/pso.purestorage.com/cockroach-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:133 /go/src/pso.purestorage.com/cockroach-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:134 /go/src/pso.purestorage.com/cockroach-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88 /usr/local/go/src/runtime/asm_amd64.s:1374 panic: runtime error: invalid memory address or nil pointer dereference [recovered] panic: runtime error: invalid memory address or nil pointer dereference [signal SIGSEGV: segmentation violation code=0x1 addr=0x0 pc=0xf6b5a7] goroutine 417 [running]: http://k8s.io/apimachinery/pkg/util/runtime.HandleCrash(0x0|k8s.io/apimachinery/pkg/util/runtime.HandleCrash(0x0, 0x0, 0x0) /go/src/pso.purestorage.com/cockroach-operator/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:58 +0x10c panic(0x161fd00, 0x22731b0) /usr/local/go/src/runtime/panic.go:969 +0x1b9 http://pso.purestorage.com/cockroach-operator/pkg/apis/pso/v1alpha1.GetNodeIdentifier|pso.purestorage.com/cockroach-operator/pkg/apis/pso/v1alpha1.GetNodeIdentifier(0xc0006d9564, 0xa, 0x0, 0xc00079d360, 0x11, 0x0, 0x0) /go/src/pso.purestorage.com/cockroach-operator/pkg/apis/pso/v1alpha1/intrusion_types.go:232 +0x67 http://pso.purestorage.com/cockroach-operator/pkg/apis/pso/v1alpha1.ValidateIntrusionSpec(0xc0005f2408|pso.purestorage.com/cockroach-operator/pkg/apis/pso/v1alpha1.ValidateIntrusionSpec(0xc0005f2408, 0xc0002d0140, 0x0) /go/src/pso.purestorage.com/cockroach-operator/pkg/apis/pso/v1alpha1/intrusion_types.go:209 +0x2c5 http://pso.purestorage.com/cockroach-operator/pkg/controller/cockroach/intrusion.(*ReconcileIntrusion).Reconcile(0xc00057b6c0|pso.purestorage.com/cockroach-operator/pkg/controller/cockroach/intrusion.(*ReconcileIntrusion).Reconcile(0xc00057b6c0, 0xc0006d9520, 0xa, 0xc0006d9504, 0x6, 0xc0007ef578, 0xc0007ef578, 0x448f33, 0xc000092a18) /go/src/pso.purestorage.com/cockroach-operator/pkg/controller/cockroach/intrusion/intrusion_controller.go:444 +0x12f2 http://sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem(0xc00025ca00|sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem(0xc00025ca00, 0xc0008eab00) /go/src/pso.purestorage.com/cockroach-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:215 +0x1de http://sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1()|sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1() /go/src/pso.purestorage.com/cockroach-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:158 +0x36 http://k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1(0xc0005994c0)|k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1(0xc0005994c0) /go/src/pso.purestorage.com/cockroach-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:133 +0x5f http://k8s.io/apimachinery/pkg/util/wait.JitterUntil(0xc0005994c0|k8s.io/apimachinery/pkg/util/wait.JitterUntil(0xc0005994c0, 0x3b9aca00, 0x0, 0x1, 0xc0005c3d40) /go/src/pso.purestorage.com/cockroach-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:134 +0x105 http://k8s.io/apimachinery/pkg/util/wait.Until(0xc0005994c0|k8s.io/apimachinery/pkg/util/wait.Until(0xc0005994c0, 0x3b9aca00, 0xc0005c3d40) /go/src/pso.purestorage.com/cockroach-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88 +0x4d created by http://sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start|sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start /go/src/pso.purestorage.com/cockroach-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:157 +0x331```15Views0likes1CommentI'm developing a Pure Storage reporting module for PowerShell
It will be open sourced on GitHub and MIT licensed. I wanted to include the likeness of some of the Purity GUI icons. I say likeness because they would be converted from svg to 24x24 png's and then to a base64 string. The icons I'm referring to are the ones from the Storage tab in the "ARRAYNAME" group i.e. Hosts, Host Groups, and so on. I can add a copyright notice/license information for them as well. Who can I discuss this with to see if I'm able to include the base64 representation of those icons with the module?15Views0likes0CommentsAttention coders! Pure Dev event is in May
Pure Dev event is on May 11th and features an industry panel with Intel, HashiCorp, CTO Advisor and Pure Storage. Topics include building cloud with infrastructure-as-code, self-service storage and data services on Kubernetes. Register here7Views0likes0CommentsI'm trying to redeploy a workload
I'm trying to redeploy a workload. When the pod is trying to spin up i'm getting this error: ``` MountVolume.SetUp failed for volume "pvc-af2b1e0f-2732-4c13-ac9f-a8b8332069a0" : rpc error: code = AlreadyExists desc = Volume is attached on another node ``` I thought PX was supposed to move the volumes to where the nodes are, or to schedule the pods to be run where the volume is? this volume has a repl of 2.9Views0likes0CommentsWe have had a major outage
We have had a major outage because of a combination of network outage and the way we designed our environment. Working with Cisco and Pure support to get to the exact details, but main issue is: Our design has ESXi hosts only connected to Pure array in DC1 and we have pods that replicate SYNC to DC2. During network maintenance there was a planned core switch failover that caused a longer outage then expected, which made the mediator unavailable and the latency for the preferred site too high. This all caused the pods to become unavailable on our preferred site and "switch" over the non-preferred site. We're still looking into the details, but I'm also searching for a possible work around during future core maintenance. For that, the question I have is if it is possible to automate (RestAPI / Powershell) the sync of pod's to pause, to prevent failovers during network maintenance and of course enable it again later on.Solved26Views0likes4Comments