Modern Data Architectures Provide a Foundation for Innovation

Sanjeev Mohan
4 min readAug 25, 2023

--

In May 2023, I had an opportunity to participate at the Precisely data integrity summit, Trust ’23. As I have done for many of the conferences I have spoken at this summer, I have summarized the discussion on modern data architectures.

The full video of the panel discussion can be found here.

Chief Operating Officer Eric Yau moderated an expert panel discussion on modern data architectures with featured panelists including Atif Salam, CxO Advisor & Enterprise Technologist at AWS; Precisely Chief Technology Officer, Tendü Yogurtçu, Ph.D and myself.

We kicked off the session by exchanging ideas about what it means to have a modern data architecture. Atif Salam noted that as recently as a year ago, the primary focus in many organizations was on ingesting data and building data lakes. Today, they’re doing everything they can to prevent their data lakes from turning into data swamps. That calls for federating data products into distinct data domains so that businesses can get timely insights.

Building on that thought, I noted that the building blocks of modern data architectures remain more or less the same, but today’s businesses are focused on optimization. They’re concerned with overall cost in the face of significant economic headwinds, and they want to accelerate their time to value so they can drive faster and better decisions.

Tendü Yogurtçu offered that organizations today are viewing data as a product, and they are working to streamline access to both data and metadata for faster insights that help business users deliver tangible value. In Yogurtçu’s perspective this conversation about data products requires an increase in the speed of access to data at scale, a focus on developing trust in data products, and better collaboration between business users and the IT department.

Challenges and Benefits of Modern Data Architecture

To be effective, Salam stated that a modern data architecture must make the right data available to the right users at the right time — and at the right price point. Conventional architectures simply cannot deliver that because they’re too rigid. They don’t lend themselves to scale, or they don’t support innovation at the speed and scale that today’s businesses require.

Cloud adoption is key to creating a modern data architecture environment because it offers cost efficiencies, rapid deployment, and agility. Salam noted that organizations are offloading computational horsepower and data from on-premises infrastructure to the cloud. This provides developers, engineers, data scientists and leaders with the opportunity to more easily experiment with new data practices such as zero-ETL or technologies like AI/ML.

In my experience, modern data architectures are increasingly segmented into very specific specialized tools, prompting businesses to focus on determining the optimal approach for integration and interoperability. Metadata, taking on an equal position alongside other data as a first-class citizen in the modern data stack, is the solution. Today, every product is gathering its own metadata without overarching standards, which puts organizations at risk of creating islands of metadata. Unification of metadata into a common plane is required to enable business use cases that deliver faster time to value, greater agility, and lower costs.

As the scale and scope of data continue to increase, that creates new challenges with respect to compliance, governance, and data quality. To create more value from data, organizations must take a very proactive approach to data integrity. That means finding and resolving data quality issues before they turn into actual problems in advanced analytics, C-level dashboards, or AI/ML models.

Data Observability and the Holistic Approach to Data Integrity

One exciting new application of AI for data management is data observability. This exciting new capability monitors data to proactively identify trends and patterns that could indicate potential issues before they impact downstream applications and analytics. Data observability also helps users identify the root cause of problems in the data. Yogurtçu spoke to the combined power of data observability, data quality, and data governance working together through shared metadata, which has become more critical than ever.

I called out this powerful use of AI and machine learning as an increasingly important factor in defining a scalable approach to data integrity. Today’s enterprises are processing more data sources, at a higher volume and velocity than ever before. To get trustworthy results, organizations must find new ways to inject intelligence and automation into their data integrity processes.

Words of Advice

The panelists in the Trust ’23 session offered some words of advice for organizations struggling to keep up with the rapid pace of change. I recommended frequent and ongoing experimentation. “Get your hands dirty”. “With very little experimentation, you can start deriving new intelligence and new insights from your corporate data. There is no other option but to get your hands dirty and start experimenting.”

Atif Salam offered his own words of wisdom: “First, establish an enterprise data strategy and a chief data office to execute that strategy. Identify the right owners for your data — and they should be in business — and get top-down support for your enterprise strategy.” That includes a long-term view funded by a multi-year investment, he says.

Tendü Yogurtçu’s advice is to “always, always start with the business case and desired outcomes, and make data integrity a priority, not an afterthought.” She also emphasizes the need to automate as much as possible. Without that, it can be virtually impossible to develop and maintain data integrity at scale.

--

--

Sanjeev Mohan
Sanjeev Mohan

Written by Sanjeev Mohan

Sanjeev researches the space of data and analytics. Most recently he was a research vice president at Gartner. He is now a principal with SanjMo.

No responses yet