talk-data.com talk-data.com

K

Speaker

Kevin Petrie

16

talks

Vice President of Research, Eckerson Group Eckerson Group

Frequent Collaborators

Filter by Event / Source

Talks & appearances

16 activities · Newest first

Search activities →

Kevin Petrie, the Vice President of Research at Eckerson Group, and Dan O’Brien, research analyst, discussed large language models (LLMs), which are neural networks that analyze text to predict the next word or phrase. These models use training data, often from the internet, to understand word relationships and provide accurate answers to natural language questions.

It’s hard to find a data discipline today that is under more pressure than data governance. One on side, the supply of data is exploding. As enterprises transform their business to compete in the 2020s, they digitize myriad events and interactions, which creates mountains of data that they need to control. On the other side, demand for data is exploding. Business owners at all levels of the enterprise need to inform their decisions and drive their operations with data.

Under these pressures, data governance teams must ensure business owners access and consume the right, high-quality data. This requires master data management—the reconciliation of disparate data records into a golden record and source of truth—which assists data governance at many modern enterprises.

In this episode, our host Kevin Petrie, VP of Research at Eckerson Group talks with our guests Felicia Perez, Managing Director, Information as a Product Program at National Student Clearinghouse, and Patrick O'Halloran, enterprise data scientist as they define what data quality and MDM are, why you need them, and how best to achieve effective data quality and MDM.

In the physical world, you can see a bridge rusting or a building facade crumbling and know you have to intervene to prevent the infrastructure from collapsing. But when all you have is bits and bytes - digital stuff, like software and data ---how can you tell if your customer-facing digital interactions or data-driven analytics and models are about to go up in smoke?

Observability is a new term that describes what we used to call IT monitoring. The new moniker is fitting given all the technology changes that have happened in the past decade. The cloud, big data, microservices, containers, cloud applications, machine learning, and artificial intelligence have created a dramatically complex IT and data environment that is harder than ever to manage. And the stakes are higher as organizations move their operations online to compete with digital natives. Today, you can't run digital or data operations without observability tools.

Kevin Petrie is one of the industry's foremost experts on observability. He is vice president of research at Eckerson Group where he leads a team of distinguished analysts. He recently wrote an article titled "The Five Shades of Observability" that describes five types of observability tools. In this podcast, we discuss what observability is, why you need it, and the types of available tools. We also speculate on the future of this technology and recommend how to select an appropriate observability product.

podcast_episode
with Sean Hewitt (Eckerson Group) , Joe Hilleary (Eckerson Group) , Dave Wells (Eckerson Group) , Kevin Petrie (Eckerson Group) , Andrew Sohn (Crawford & Company)

Every December, Eckerson Group fulfills its industry obligation to summon its collective knowledge and insights about data and analytics and speculate about what might happen in the coming year. The diversity of predictions from our research analysts and consultants exemplifies the breadth of their research and consulting experiences and the depth of their thinking. Predictions from Kevin Petrie, Joe Hilleary, Dave Wells, Andrew Sohn, and Sean Hewitt range from data and privacy governance to artificial intelligence with stops along the way for DataOps, data observability, data ethics, cloud platforms, and intelligent robotic automation.

The COVID shock forces enterprises in every market to accelerate and reshape their data analytics strategies. This trend is likely to continue. “Data Elite” enterprises survived this year through a mix of agility, efficiency, and intelligence. They met these requirements of survival as they accelerated their digital transformations, adopted cloud data platforms and embraced advanced analytics. As these data leaders continue their momentum in 2021, the data laggards will strive to catch up.

In this episode, Kevin Petrie, VP of Research at Eckerson Group, interviews Sumeet Agrawal, VP of Product Management at Informatica, to discuss the impact of COVID on enterprises. Sumeet talks about the trends of adoption during the onslaught of COVID and how enterprises are navigating in the post-pandemic era.

Continuous Intelligence (CI) integrates historical and real-time analytics to automatically monitor and update various types of systems, including supply chains, telecommunications networks and e-commerce sites. CI encompasses data ingestion, transformation and analytics, as well as operational “triggers” that recommend or initiate specific real-time actions.

CI casts a wider net than traditional analytics because it includes contextual data, for example related to market behavior, weather patterns or social media trends, that help enterprises operate the core systems more intelligently.

In this episode, our VP of Research Kevin Petrie interviews Simon Crosby, CTO at Swim.ai, a continuous intelligence software vendor that focuses on edge-based learning for fast-data. He co-founded security vendor Bromium in 2010, later sold to HP Inc in 2019.

This is an audio blog on BI on the Cloud Data Lake and how to improve the productivity of data engineers. We'll dive deeper into the question; what’s the best measure of success for data pipeline efficiency? This is part 2 of a two part blog.

Originally published at: https://www.eckerson.com/articles/business-intelligence-on-the-cloud-data-lake-part-2-improving-the-productivity-of-data-engineers

This is an audio blog about the perplexities of the Data Lakehouse and if it is, indeed, the "paradigm of the decade". To hear more of Eckerson Group perspectives on the data lakehouse be sure to check out the blogs from colleagues, Wayne Eckerson and Kevin Petrie, and the recording of our recent Shop Talk discussion.

Originally published at: https://www.eckerson.com/articles/an-architect-s-view-of-the-data-lakehouse-perplexity-and-perspective

This audio blog discusses the Data Lakehouse, a marketing concept that evokes clean PowerPoint imagery, and why and how the New Cloud Data Lake will play a very real role in modern enterprise environments.

Originally published at: https://www.eckerson.com/articles/data-lakehouses-hold-water-thanks-to-the-cloud-data-lake

Streaming Change Data Capture

There are many benefits to becoming a data-driven organization, including the ability to accelerate and improve business decision accuracy through the real-time processing of transactions, social media streams, and IoT data. But those benefits require significant changes to your infrastructure. You need flexible architectures that can copy data to analytics platforms at near-zero latency while maintaining 100% production uptime. Fortunately, a solution already exists. This ebook demonstrates how change data capture (CDC) can meet the scalability, efficiency, real-time, and zero-impact requirements of modern data architectures. Kevin Petrie, Itamar Ankorion, and Dan Potter—technology marketing leaders at Attunity—explain how CDC enables faster and more accurate decisions based on current data and reduces or eliminates full reloads that disrupt production and efficiency. The book examines: How CDC evolved from a niche feature of database replication software to a critical data architecture building block Architectures where data workflow and analysis take place, and their integration points with CDC How CDC identifies and captures source data updates to assist high-speed replication to one or more targets Case studies on cloud-based streaming and streaming to a data lake and related architectures Guiding principles for effectively implementing CDC in cloud, data lake, and streaming environments The Attunity Replicate platform for efficiently loading data across all major database, data warehouse, cloud, streaming, and Hadoop platforms