FREE BEER FOR ALL!!! Now That I Have Your Attention, Let's Talk About Purity Updates.
WAIT WAIT WAIT - don't leave yet because of my free beer tomfoolery....hear me out. Listen, we get it. Storage OS updates are historically the LAST thing you ever want to consider for your already impossibly thin maintenance windows. And, we all know NOBODY ever grew up saying, "When I get older, I want to manage enterprise storage for its rock and roll lifestyle." 😀 But - hear me out. Any past pain, suffering, or heavy drinking you may have taken on during previous OS updates with other legacy vendors has been minimized or even flat out eliminated by how we handle updating Purity for FlashArray and FlashBlade. We offer two tracks you can leverage for making them happen by either working directly with support for a white glove update experience where they do all the work remotely, or you can complete them via the Self Support Update (SSU) feature built into Pure1. We encourage regular Purity updates for two reasons: Performance. stability and security improvements...obviously New feature adoption. Want Fusion 2.0? Want the ability to deliver NFS/SMB shares on your FlashArray? These are bundled into your Purity updates and require no additional licensing costs to adopt if you want them. Think of them as over the air feature updates that are all the rage for EVs... For now, take a quick look at the Purity version you are running. If you haven't updated it in a year or two (which many of you probably haven't), you're missing out on being able to squeeze extra value out of your storage. I will be posting some supporting demos and other materials to help you visualize the process in the coming month or so. I would LOVE any feedback from the community, good or bad, on current or past experiences with our updating experience...through it all we can get more boats to rise with the tide! Stay tuned! DP17Views3likes1CommentAnnouncing the General Availability of Purity//FA 6.5.11 LLR
We are happy to announce the general availability of 6.5.11, the twelfth release in the 6.5 Long-Life Release (LLR) line! This release line is based on the feature set introduced in 6.4, providing long-term consistency in capabilities, user experience, and interoperability, with the latest fixes and security updates. This 6.5 LLR line has demonstrated sufficient accumulated runtime data to be recommended for critical customer workloads and has been declared Enterprise Ready (ER) effective with the 6.5.3 release. For more detailed information about bug fixes and security updates included in each release, see the release notes. UPGRADE RECOMMENDATIONS AND EOL SCHEDULE We recommend customers already running 6.5 who are looking for the latest fixes and updates to upgrade to this long-life release. When possible, customers should make use of Self-Service Upgrades (SSU) to ease the process of planning and executing non-disruptive Purity upgrades for their fleet. The 6.5 LLR line is planned for development through September 2026. HARDWARE SUPPORT This release is supported on the following FlashArray Platforms: FA//M (R2), FA//X, FA//C, and FA//XL. Note, DFS software version 2.2.3 is recommended with this release. ACKNOWLEDGEMENTS We would like to thank everyone within the engineering, support, technical program management, product management, product marketing, finance and technical product specialist teams who contributed to this release. LINKS AND REFERENCES Purity//FA 6.5 Release Notes Self-Service Upgrades Purity//FA Release and End-of-Life Schedule FlashArray Hardware and End-of-Support DirectFlash Shelf Software Compatibility Matrix FlashArray Capacity and Feature Limits6Views0likes0CommentsNew Reference Architecture: SQL Server on Azure VM's with Pure Cloud Block Store
This is a brand new, weeks-old reference architecture — and I’m really excited about this one. During development, one of the most surprising discoveries was just how much Azure VM performance is limited by the IOPS cap tied to managed disks. It caught me off guard how much planning it takes just to size storage and compute together when you go the native route. With CBS, I was able to bypass those constraints. It felt more like working with enterprise storage (which is what its meant to do!) , I could pull from a pool, scale performance independently of VM size, and provision storage volumes in a clean and easy way. This new RA covers: SQL Server architecture on Azure VMs with Pure Cloud Block Store Snapshot-based backup and restore DR patterns using ActiveDR™ and HA using ActiveCluster™ Dev/test database cloning with volume snapshots Performance benchmarking vs. Azure Premium SSD v2 It prooved: ~40% more transactional throughput (TPROC-C) ~93% better analytical query performance (TPROC-H) (using queries per minute normalization) 3–5x data reduction vs. raw data Download the full reference architecture here Would love to hear your thoughts on this architecture and how we could improve the expirience!6Views2likes0CommentsAnnouncing the General Availability of Purity//FA 6.7.4 LLR
We are happy to announce the general availability of 6.7.4, the fifth release in the 6.7 Long-Life Release (LLR) line! This release line is based on the feature set introduced in 6.6, providing long-term consistency in capabilities, user experience, and interoperability, with the latest fixes and security updates. When the 6.7 LLR line demonstrates sufficient accumulated runtime data to be recommended for critical customer workloads, it will be declared Enterprise Ready (ER). Until then, Purity//FA 6.5 is the latest ER-designated LLR line. For more detailed information about bug fixes and security updates included in each release, see the release notes. UPGRADE RECOMMENDATIONS AND EOL SCHEDULE We recommend customers with compatible hardware who are looking for the latest feature set offered for long-term maintenance upgrade to this long-life release. When possible, customers should make use of Self-Service Upgrades (SSU) to ease the process of planning and executing non-disruptive Purity upgrades for their fleet. The 6.7 LLR line is planned for development through October 2027. HARDWARE SUPPORT This release is supported on the following FlashArray Platforms: Cloud Block Store for Azure and AWS, FA//X (R2, R3, R4), FA//C, FA//XL, and FA//E. Note, DFS software version 2.2.3 is recommended with this release. ACKNOWLEDGEMENTS We would like to thank everyone within the engineering, support, technical program management, product management, product marketing, finance and technical product specialist teams who contributed to this release. LINKS AND REFERENCES Purity//FA 6.7 Release Notes Purity//FA 6.6/6.7 Feature Content Self-Service Upgrades Purity//FA Release and End-of-Life Schedule FlashArray Hardware and End-of-Support DirectFlash Shelf Software Compatibility Matrix FlashArray Capacity and Feature Limits FlashArray Feature Interoperability Matrix19Views2likes0CommentsBackup and Restore FA Configuration
Hi All, Purestorage has how to save or backup and restore configuration when an issue occurs? For example, the software crashed and could not be used anymore. Once the problem was fixed, we restored the existing configuration file from backup.Solved137Views2likes6CommentsAnnouncing the General Availability of Purity//FA 6.8.6
We are happy to announce the general availability of 6.8.6, the seventh release in the 6.8 Feature Release line, including the SMB Continuous Availability feature, which guarantees zero downtime for customers' businesses during controller disruptions and upgrades, ensuring uninterrupted access to shared files. Some of the improvements to Purity contained in this release include: SMB Continuous Availability preserves file handles to ensure uninterrupted SMB access during controller failovers and upgrades. Target Pods for Pgroup Replication allows customers to target a specific pod for protection groups, avoiding the clutter of snapshots replicating to an array’s root pod. CBS for AWS Write Optimization for Amazon S3 improves how its data is committed and managed on Amazon S3 and can significantly drop the AWS infrastructure operating cost, providing customers with write or replication heavy workloads with cost reductions of up to 50%. Allow NVMe Read-Only Volumes for ActiveDR eliminates restriction on promotion/demotion of pods containing NVMe-connected volumes, saving customers from unexpected command failures and time-consuming workarounds. For more detailed information about features, bug fixes, and security updates included in each release, see the release notes. UPGRADE RECOMMENDATIONS AND EOL SCHEDULE We recommend customers with compatible hardware who are looking for the fastest access to the latest features upgrade to this new feature release. When possible, customers should make use of Self-Service Upgrades (SSU) to ease the process of planning and executing non-disruptive Purity upgrades for their fleet. This 6.8 release line is planned for feature development through May 2025 with additional fixes coming in June 2025 to end the release line. HARDWARE SUPPORT This release is supported on the following FlashArray Platforms: Cloud Block Store for Azure and AWS, FA//X (R3, R4), FA//C (R3, R4), FA//XL, FA//E, and FA//RC (starting with 6.8.5). Note, DFS firmware version 2.2.3 is recommended with this release. ACKNOWLEDGEMENTS We would like to thank everyone within the engineering, support, technical program management, product management, product marketing, finance and technical product specialist teams who contributed to this release. LINKS AND REFERENCES Purity//FA 6.8 Release Notes Self-Service Upgrades Purity//FA Release and End-of-Life Schedule FlashArray Hardware and End-of-Support DirectFlash Shelf Software Compatibility Matrix FlashArray Capacity and Feature Limits48Views1like0CommentsSnapshots and growth
I have a question about snapshot growth and retention. Last week we had 14 days worth of snapshots and due to some storage growth, we changed this to 7 days worth of snapshots. Before the change was made snapshots were taking up about 21 TB of space, after the change that number is around 10 TB. This reduction of space was more than expected. We expected around a 5 TB reduction. We just added up days 8-14 to get the 5 number. The other 6 TB reduction came from the most recent snapshot which at the time was 11 TB in size and now its down to around 5 TB in size. Does anybody know why the most current snapshot also had a large reduction after making this change? We are trying to figure out future growth including snapshot growth.160Views0likes4CommentsActiveCluster Asynchronous
Hello Having question about ActiveCluster Asynchronous is it possible to configure between 2 existing clusters Poland to Germany . Idea is to have ability to perform DR tests in Datacenters located in different country The customer is currently running a configuration based on X20R2/R4. They have two systems in the Poland and two in the Germany. Local replication between arrays is configured using ActiveCluster Synchronous. ################## Reviewed documentation is ActiveCluster over Fibre Channel and ActiveDR or Async is supported on the same system starting with Purity 6.1.3+. ActiveDR must be configured with separate volumes (and pods) from ActiveCluster volumes. ActiveCluster asynchronous works in such way that that both arrays Metro Arrays replicate data to 3d array protection group snapshot only Leveraging asynchronous replication is easy to do, it's a simple matter of defining a Target array in a Protection Group after connecting the array. Once defined in a Protection Group, the Protection Group itself can be moved into an ActiveCluster (our synchronous replication, RPO0 replication service) Pod, where the Protection Group is owned by two arrays. The defined Target can replicate regularly scheduled snapshots to a third array. This Active-Active Asynchronous Replication is shared by the ActiveCluster arrays and in the event that either array is offline, the alternate array will assume ownership of continual snapshot replication to the third array. In summary, you can replicate snapshots as desired between any number of arrays to any other number of arrays, requiring a defined array connection and Protection Group Target. These Protection Groups can also be moved into a pod for sharing between ActiveCluster arrays for disaster recovery purposes as well. The sequence of steps for enabling asynchronous replication: Connect arrays so the source and target arrays are aware of each other Create a Protection Group with desired snapshot policies Add any array to replicate snapshots to the Target field If using in an ActiveCluster pair, move the Protection Group into the ActiveCluster podSolved125Views1like3CommentsPurity FA 6.6.10* Introduces Better Data Security
Purity FA 6.6.10 introduces better data security by auditing file access. This means for all SMB and NFS shares, all access events can be captured and recorded to a local file or sent to a remote CIS log server. Check out the Pure Storage //Launch Round-Up for August! Have more questions? We’re all ears!33Views1like0Comments