When Data Becomes the Mission
Why state and local government, cities, and research universities are reorganizing infrastructure around data itself If you remember one thing from this article: infrastructure used to organize around applications. Increasingly, now it organizes around data. If you spend enough time around enterprise infrastructure, you start to notice something about how conversations begin. Someone asks about storage. Not in a philosophical way. In a practical way. How much capacity do we have left? What’s the refresh cycle? Is this staying on premises or moving to cloud? What’s the backup strategy? For years, that framing made perfect sense. Infrastructure was the foundation, and the job of infrastructure teams was to keep the lights on and the foundation solid. But lately, in conversations with customers across state and local government, municipalities, cities, and universities, something feels different. Because eventually someone says something like this: “We have this data… but we can’t actually use it.” And that is when the real conversation begins. Why the public sector reveals the truth about data There’s a perspective I heard recently that stuck with me. The public sector isn’t a niche market. It’s a microcosm of the entire enterprise technology world. At first that sounds counterintuitive. The stereotype is that government IT has been quietly living under a rock since the previous century, next to a beige server and a stack of COBOL manuals. But if you look closely, the opposite is true. State agencies, cities, and research institutions operate in environments that combine nearly every architectural challenge the private sector faces — all at once. Massive datasets Highly distributed users Strict security requirements Long retention policies Global collaboration And an absolute requirement that systems remain available when people need them most. In other words, the public sector experiences the full spectrum of data challenges simultaneously. If you want to stress-test a data architecture, put it inside government. Think about it. A state government may run thousands of systems across dozens of agencies, each serving different missions but increasingly sharing the same underlying data. A city manages infrastructure at the physical edge of society — traffic, water, SCADA, emergency services — where real-time decisions depend on accurate information. Universities generate some of the largest research datasets on earth while collaborating across institutions and countries. Each of these environments demands something slightly different from infrastructure. But they all demand the same thing from data: Security. Integrity. Mobility. Context. Availability. And when those requirements collide in one environment, something interesting happens. The solutions that work there tend to work everywhere. A laboratory for the modern data enterprise This is why many technology leaders quietly view the public sector as something more than a vertical market. It’s a laboratory for enterprise-scale data architecture. If a platform can operate in a world where: sensitive personal data must remain protected • systems span thousands of locations • regulatory oversight is constant • and uptime has real public consequences …then that architecture will almost certainly succeed in commercial environments. Banks, manufacturers, healthcare providers, and global enterprises face the same challenges. Just rarely all at once. Government simply compresses those problems into a single environment. Solve the data problem for government, and you solve it for the enterprise. That’s one reason the shift toward data-centric platforms is becoming so important. When organizations treat infrastructure as a place to store files, they solve only a small part of the problem. But when they treat data as the central operational asset — something that must be understood, governed, protected, and made usable across environments — the architecture begins to look very different. And the public sector, with all its complexity, becomes the place where those architectures are tested first. Which brings us back to the shift we’re seeing across the industry. Because once you start looking at infrastructure through the lens of data itself, something else becomes obvious. The center of gravity has moved. When multiple systems depend on the same dataset, the data becomes part of the operating foundation. And once that happens, moving it — or even restructuring it — becomes dramatically harder. Which brings us to the concept that explains a lot of what is happening right now. The quiet physics of data gravity The first time I heard the term “data gravity” wasn’t in a conference keynote or a vendor presentation. It was in 2015, when a recruiter from a startup called DataGravity (now Anomalo) reached out and asked if I would be interested in interviewing. At the time, the idea sounded fascinating — and slightly theoretical. The company was built around the premise that data itself was becoming the most valuable asset in the data center, and that infrastructure needed to understand the content, context, and behavior of data, not just store it. The name alone hinted at something deeper: the idea that as datasets grow, they start exerting a kind of gravitational pull on the systems around them. Back then, it felt like an interesting concept. Today it feels like a description of reality. The term “data gravity” itself was introduced by Dave McCrory back in 2010, and it turns out to be a remarkably accurate way to describe modern infrastructure. Dave McCrory Blog The idea is simple. As datasets grow, they become harder to move. More applications depend on them. More workflows connect to them. More policies govern them. Eventually, the architecture starts organizing around the data itself. Not because someone designed it that way. Because the physics of large systems leave you very little choice. Imagine trying to relocate a state Medicaid dataset that has been integrated with multiple benefit programs, identity verification systems, and fraud detection tools. Technically possible? Sure. Operationally trivial? Not even close. The larger and more interconnected the dataset becomes, the stronger its gravitational pull. Compute moves closer to the data. Applications move closer to the data. Infrastructure reorganizes around the data. This is why organizations that once talked primarily about storage capacity are now talking about data platforms. The center of gravity moved. When data stops being passive The moment data becomes operational, everything changes. For years, most organizations treated data as something that accumulated quietly inside systems. Applications produced it. Storage kept it safe. Backups made sure it could be restored. But that model starts to break down when the data itself becomes part of real-time decision making. You can see this most clearly in environments that generate enormous volumes of information. Cities now run infrastructure that continuously streams telemetry — traffic sensors, utility meters, environmental monitors, emergency response platforms. A water meter that once reported usage once a month might now generate thousands of readings per year. A traffic system that once relied on static timing can adapt dynamically to real-time conditions. Each improvement creates more data. More importantly, it creates operational dependence on that data. Universities experience the same phenomenon in a different form. Research environments produce extraordinary datasets across genomics, climate science, and artificial intelligence. Sequencing a single human genome generates roughly 100 gigabytes of raw data, and large research programs may create terabytes or petabytes of new information every week. In those environments the challenge isn’t just storing data. It’s feeding it fast enough to the systems that depend on it. Modern research clusters and GPU environments can process enormous volumes of information, but only if the underlying data pipeline keeps up. When storage cannot deliver data fast enough, expensive compute resources sit idle and discovery slows down. And that reveals an important truth about modern infrastructure. When systems depend on data in real time, the question stops being where the infrastructure lives. The question becomes whether the data is available, trustworthy, and recoverable. That distinction also explains why ransomware has become so disruptive to public institutions. Attackers understand that the real leverage is not the servers or the network. It’s the data. When access to data disappears, the services built on top of it disappear as well. Which brings us back to the deeper shift happening across the industry. If data has become this central to operations, services, and discovery, then managing it as a passive byproduct of infrastructure is no longer enough. Infrastructure alone is no longer the strategic layer. The strategic layer is the data itself. Organizations still need performance, availability, and resilience. Those fundamentals have not changed. What has changed is the expectation that infrastructure should also help organizations understand, govern, protect, and use their data more effectively. That is a very different problem than simply storing it. And it is the reason the conversation is evolving from storage management to data management platforms. The real punch line Public sector organizations didn’t set out to become data enterprises. Over time the data accumulated. Then the dependencies formed. And eventually everything started orbiting the datasets that mattered most. Data has gravity. Data has risk. Data has power. Infrastructure still matters. But increasingly, the real mission is something else entirely. The mission is the data. Appreciate you reading. Dmitry Gorbatov © 2025 Dmitry Gorbatov | #dmitrywashere39Views0likes0CommentsBig news with Nokia and Pure
Nokia has selected Pure Storage to power the high‑performance, all‑flash data layer for its telco cloud on Red Hat OpenShift, enabling secure, scalable CNFs from edge to core. This is very big news as Pure continues to grow its global footprint in the telecom industry, helping telcos across multiple use cases: RAN modernization (5G/6G), AI, telco clouds, autonomous networks, OSS/BSS and lots more. Read all about the Nokia partnership in the Pure Storage blog: https://blog.purestorage.com/news-events/nokia-pure-storage-telco-red-hat/69Views1like0CommentsHealthcare AI: Why the "Build Reflex" is Killing Your ROI
In this article, For Healthcare Leaders, Build vs. Buy Determines ROI on Enterprise AI, featuring Matthew Crowson, MD, of Wolters Kluwer, Matthew argues that healthcare organizations must abandon their traditional "build reflex" for AI solutions, citing a high 95% failure rate. This traditional habit in the healthcare system is in stark contrast with tight margins and the competitive AI talent market. Crowson advocates for a shift to a hybrid partnership model where the organization "buys" a vendor's customizable platform. This model is crucial because it addresses trust issues by ensuring that sensitive patient data (PHI) remains secure behind the facility's firewall. He stresses to first focus on problem diagnosis, be realistic about their in-house talent, and ensure their data foundation is clean before engaging any vendors. This pragmatic approach is essential for achieving a positive ROI on enterprise AI. Community Question: What do you think? Is your organization currently struggling with the build vs. buy decision? Let's discuss! Click through to read the entire article above and let us know your thoughts around it in the comments below!53Views0likes0CommentsVideo Storage: Less Cost, More Reliability
Learn how a leading MSO switched from cloud-based video storage to Pure Storage FlashBlade//E and saved 87% on cloud costs. And all while getting a more reliable service and a better customer experience via reduced latency. Read all about it.75Views0likes0CommentsAI is changing everything in Telco
Like every industry, AI is having a massive impact on telecom. But what do your peers think about it? We partnered with NVIDIA to sponser a research report about AI in the telecom space. It includes expert opinions plus thoughts from AI leaders at MetTel, Telus and Verizon. There's also a very informative webinar featuring Pure Storage Telecom Field CTO Patrick Lopez and Chris Penrose, VP, Head of Telecoms Business Development at NVIDIA. All the links are available in this blog post.73Views0likes0CommentsCanadian Telecom Summit, 2025
Pure Storage will be speaking at the Canadian Telecom Summit on a panel session, "Technology Megatrends – Private Networks and Edge, Cloud, AI, and Quantum" (details here). The Summit takes place June 3-4, 2025, at the Metro Toronto Convention Center. We hope to see you there!93Views0likes0CommentsPure making a difference in Healthcare
I look forward to particpating in this group and sharing the experiences of our team who come from the Healthcare Industry. Should you have any challenges, please feel free to reach out to our team - don't pay for consulting leverage Pure Storages Healthcare IT expertise!102Views0likes0CommentsFinding predictions in the unpredictable
At Pure we are looking to make meaningful changes and bring value to the markets we serve. While our technology is second-to-none, we know that business problems in healthcare aren't always solved by throwing technology at it alone. If you are able to join us at our customer event: Accelerate you will notice we have a dedicated Healthcare day. We will be delivering high-impact, reasonable length sessions, detailing out some exciting new solutions and announcements for our healthcare customers. One of our solutions is EvergreenOne// Medical imaging (PACS/VNA). This solution is truly unique in the market in the way it solves for data management while also delivering predictable pricing. Billed "Per Study" we lock in a price for you, for up to 10-years and take on the risk of variability in study size growth, performance, privacy, and reliability. Organizations seemingly fall into two camps. Either they have moved to a "cloud-like" billing model(monthly cost per TB consumed), or continue capital purchase where you guess your next 5 years capacity needs and performance needs for your PACS/VNA data. Historically there are 3 data needs to address for your organizations PACS/VNA deployment: Burning through OpEx budget since every month your PACS/VNA data grows larger Running out of storage sooner (under provisioning) or overpaying ( over provisioning) with CapEx purchase Performance requirements change ( e.g. running 10 AI algorithms while servicing the human clinicians off the same data platform, simultaneously) With our EvergreenOne//Medical Imaging solution, we believe we have to answer to solve for all 3 above. If youll be able to join us at Accelerate in June, you will be the first to hear, otherwise stay tuned for our public announcement happening just after Accelerate ends.85Views1like0CommentsHealthcare Payers
Did you know Pure has a dedicated Healthcare Payer vertical? I work with all our customers and prospects to create solutions to Payer specific enviroments such as Mainframe Backups, Epic Payer Platform and Clean Rooms. To learn more, visit https://www.purestorage.com/solutions/industries/healthcare/payers.html And as always, feel free to reach out if there's anything we can do for you or anything you'd like to share! Looking forward to hearing from you all. Priscilla Sandberg - Senior Manager Global Payer Alliances psandberg@purestorage.com108Views2likes0Comments