Talk by ECB and IBM Consulting on a versatile platform enabling advanced network analytics for banking supervision.
talk-data.com
Topic
Analytics
4552
tagged
Activity Trend
Top Events
Summary
The insurance industry is notoriously opaque and hard to navigate. Max Cho found that fact frustrating enough that he decided to build a business of making policy selection more navigable. In this episode he shares his journey of data collection and analysis and the challenges of automating an intentionally manual industry.
Announcements
Hello and welcome to the Data Engineering Podcast, the show about modern data management Introducing RudderStack Profiles. RudderStack Profiles takes the SaaS guesswork and SQL grunt work out of building complete customer profiles so you can quickly ship actionable, enriched data to every downstream team. You specify the customer traits, then Profiles runs the joins and computations for you to create complete customer profiles. Get all of the details and try the new product today at dataengineeringpodcast.com/rudderstack This episode is brought to you by Datafold – a testing automation platform for data engineers that finds data quality issues before the code and data are deployed to production. Datafold leverages data-diffing to compare production and development environments and column-level lineage to show you the exact impact of every code change on data, metrics, and BI tools, keeping your team productive and stakeholders happy. Datafold integrates with dbt, the modern data stack, and seamlessly plugs in your data CI for team-wide and automated testing. If you are migrating to a modern data stack, Datafold can also help you automate data and code validation to speed up the migration. Learn more about Datafold by visiting dataengineeringpodcast.com/datafold As more people start using AI for projects, two things are clear: It’s a rapidly advancing field, but it’s tough to navigate. How can you get the best results for your use case? Instead of being subjected to a bunch of buzzword bingo, hear directly from pioneers in the developer and data science space on how they use graph tech to build AI-powered apps. . Attend the dev and ML talks at NODES 2023, a free online conference on October 26 featuring some of the brightest minds in tech. Check out the agenda and register today at Neo4j.com/NODES. You shouldn't have to throw away the database to build with fast-changing data. You should be able to keep the familiarity of SQL and the proven architecture of cloud warehouses, but swap the decades-old batch computation model for an efficient incremental engine to get complex queries that are always up-to-date. With Materialize, you can! It’s the only true SQL streaming database built from the ground up to meet the needs of modern data products. Whether it’s real-time dashboarding and analytics, personalization and segmentation or automation and alerting, Materialize gives you the ability to work with fresh, correct, and scalable results — all in a familiar SQL interface. Go to dataengineeringpodcast.com/materialize today to get 2 weeks free! Your host is Tobias Macey and today I'm interviewing Max Cho about the wild world of insurance companies and the challenges of collecting quality data for this opaque industry
Interview
Introduction How did you get involved in the area of data management? Can you describe what CoverageCat is and the story behind it? What are the different sources of data that you work with?
What are the most challenging aspects of collecting that data? Can you describe the formats and characteristics (3 Vs) of that data?
What are some of the ways that the operational model of insurance companies have contributed to its opacity as an industry from a data perspective? Can you describe how you have architected your data platform?
How have the design and goals changed since you first started working on it? What are you optimizing for in your selection and implementation process?
What are the sharp edges/weak points that you worry about in your existing data flows?
How do you guard against those flaws in your day-to-day operations?
What are the
Dante joins Cris and Mark to digest the September jobs report. The outsized job gain during the month was surprising, but after Dante’s masterful dissection of the data, the group agrees there is a lot to like in the report. After the stats game, they discussion turns to the recent surge in long-term interest rates and how big a threat it poses to the economy. Follow Mark Zandi @MarkZandi, Cris deRitis @MiddleWayEcon, and Marisa DiNatale on LinkedIn for additional insight.
Questions or Comments, please email us at [email protected]. We would love to hear from you. To stay informed and follow the insights of Moody's Analytics economists, visit Economic View.
It turns out data plays a big role in getting cereal manufactured and delivered so you can enjoy your Cheerios reliably for breakfast. We talk with Arjun Narayan, CEO of Materialize, a company building an operational warehouse, and Nathan Bean, a data leader at General Mills responsible for all of the company's manufacturing analytics and insights. We discuss Materialize's founding story, how streaming technology has matured, and how exactly companies are leveraging their warehouse to operationalize their business—in this case, at one of the largest consumer product companies in the United States. For full show notes and to read 6+ years of back issues of the podcast's companion newsletter, head to https://roundup.getdbt.com. The Analytics Engineering Podcast is sponsored by dbt Labs.
Can artificial intelligence change the future of life insurance? Join Host Jason Foster as he sits down with Glenn Hofmann, the Chief Analytics Officer at New York Life, to explore the transformative power of AI in the insurance industry. In this episode, discover how AI is reshaping risk assessment, revolutionising customer retention strategies, and optimising office spaces at one of the largest financial companies in the United States. Glenn shares invaluable insights on responsible AI implementation, governance, and the ethical considerations essential for success in the insurance sector.
Join Avery for a captivating episode with Data Consulting expert Leon Gordon, to delve into the world of data consulting and the skills needed to excel in this field.
Don't miss out on this fascinating conversation filled with valuable insights and advice for data professionals looking to make a difference as consultants.
Tune in now to better understand the data consulting industry and how you can navigate your way to a successful career!
Connect with Leon Gordon:
🤝 Connect on Linkedin
🎒 Learn About Onyx Data
🤝 Ace your data analyst interview with the interview simulator
📩 Get my weekly email with helpful data career tips
📊 Come to my next free “How to Land Your First Data Job” training
🏫 Check out my 10-week data analytics bootcamp
Timestamps:
(5:56) - Data Consulting
(9:45) - Power BI
(25:00) - Data Projects
Connect with Avery:
📺 Subscribe on YouTube
🎙Listen to My Podcast
👔 Connect with me on LinkedIn
🎵 TikTok
Mentioned in this episode: Join the last cohort of 2025! The LAST cohort of The Data Analytics Accelerator for 2025 kicks off on Monday, December 8th and enrollment is officially open!
To celebrate the end of the year, we’re running a special End-of-Year Sale, where you’ll get: ✅ A discount on your enrollment 🎁 6 bonus gifts, including job listings, interview prep, AI tools + more
If your goal is to land a data job in 2026, this is your chance to get ahead of the competition and start strong.
👉 Join the December Cohort & Claim Your Bonuses: https://DataCareerJumpstart.com/daa https://www.datacareerjumpstart.com/daa
Send us a text Sam Torres is the Chief Digital Officer and co-founder of The Gray Dot Company, Gray Dot Company is a consulting firm that specializes in search engine optimization. Sam outlines expertise in complex digital analytics and consumer insights data. 03:40 Meet Sam Torres05:57 Marketing Platforms07:49 Digital Consumer Intelligence14:55 Defining Success17:55 AIs Impact on Google22:07 Should I Trust Sponsored Adds?23:58 GenAI Positives LinkedIn: linkedin.com/in/samantha-torres-seo Website: https://thegray.company, https://legendarypodcasts.com/sam-torres/
Want to be featured as a guest on Making Data Simple? Reach out to us [email protected] and tell us why you should be next. The MakingData Simple Podcast is hosted by Al Martin, WW VP Technical Sales, IBM, wherewe explore trending technologies, business innovation, and leadership ... whilekeeping it simple & fun. Want to be featured as a guest on Making Data Simple? Reach out to us at [email protected] and tell us why you should be next. The Making Data Simple Podcast is hosted by Al Martin, WW VP Technical Sales, IBM, where we explore trending technologies, business innovation, and leadership ... while keeping it simple & fun.
Most of the time, we think of analytics as taking historical data for a business, munging it in various ways, and then using the results of that munging to make decisions. But, what if the business has no (or very little) historical data… because it's a startup? That's the situation venture capitalists — especially those focused on early stage startups — face constantly. We were curious as to how and where data and analytics play a role in such a world, and Sam Wong, a partner at Blackbird Ventures, joined Michael, Val, and Tim to explore the subject. Hypotheses and KPIs came up a lot, so our hypothesis that there was a relevant tie-in to the traditional focus of this show was validated, and, as a result, the valuation of the podcast itself tripled and we are accepting term sheets. For complete show notes, including links to items mentioned in this episode and a transcript of the show, visit the show page.
Amazon Redshift powers analytic cloud data warehouses worldwide, from startups to some of the largest enterprise data warehouses available today. This practical guide thoroughly examines this managed service and demonstrates how you can use it to extract value from your data immediately, rather than go through the heavy lifting required to run a typical data warehouse. Analytic specialists Rajesh Francis, Rajiv Gupta, and Milind Oke detail Amazon Redshift's underlying mechanisms and options to help you explore out-of-the box automation. Whether you're a data engineer who wants to learn the art of the possible or a DBA looking to take advantage of machine learning-based auto-tuning, this book helps you get the most value from Amazon Redshift. By understanding Amazon Redshift features, you'll achieve excellent analytic performance at the best price, with the least effort. This book helps you: Build a cloud data strategy around Amazon Redshift as foundational data warehouse Get started with Amazon Redshift with simple-to-use data models and design best practices Understand how and when to use Redshift Serverless and Redshift provisioned clusters Take advantage of auto-tuning options inherent in Amazon Redshift and understand manual tuning options Transform your data platform for predictive analytics using Redshift ML and break silos using data sharing Learn best practices for security, monitoring, resilience, and disaster recovery Leverage Amazon Redshift integration with other AWS services to unlock additional value
Summary
Artificial intelligence applications require substantial high quality data, which is provided through ETL pipelines. Now that AI has reached the level of sophistication seen in the various generative models it is being used to build new ETL workflows. In this episode Jay Mishra shares his experiences and insights building ETL pipelines with the help of generative AI.
Announcements
Hello and welcome to the Data Engineering Podcast, the show about modern data management Introducing RudderStack Profiles. RudderStack Profiles takes the SaaS guesswork and SQL grunt work out of building complete customer profiles so you can quickly ship actionable, enriched data to every downstream team. You specify the customer traits, then Profiles runs the joins and computations for you to create complete customer profiles. Get all of the details and try the new product today at dataengineeringpodcast.com/rudderstack This episode is brought to you by Datafold – a testing automation platform for data engineers that finds data quality issues before the code and data are deployed to production. Datafold leverages data-diffing to compare production and development environments and column-level lineage to show you the exact impact of every code change on data, metrics, and BI tools, keeping your team productive and stakeholders happy. Datafold integrates with dbt, the modern data stack, and seamlessly plugs in your data CI for team-wide and automated testing. If you are migrating to a modern data stack, Datafold can also help you automate data and code validation to speed up the migration. Learn more about Datafold by visiting dataengineeringpodcast.com/datafold You shouldn't have to throw away the database to build with fast-changing data. You should be able to keep the familiarity of SQL and the proven architecture of cloud warehouses, but swap the decades-old batch computation model for an efficient incremental engine to get complex queries that are always up-to-date. With Materialize, you can! It’s the only true SQL streaming database built from the ground up to meet the needs of modern data products. Whether it’s real-time dashboarding and analytics, personalization and segmentation or automation and alerting, Materialize gives you the ability to work with fresh, correct, and scalable results — all in a familiar SQL interface. Go to dataengineeringpodcast.com/materialize today to get 2 weeks free! As more people start using AI for projects, two things are clear: It’s a rapidly advancing field, but it’s tough to navigate. How can you get the best results for your use case? Instead of being subjected to a bunch of buzzword bingo, hear directly from pioneers in the developer and data science space on how they use graph tech to build AI-powered apps. . Attend the dev and ML talks at NODES 2023, a free online conference on October 26 featuring some of the brightest minds in tech. Check out the agenda and register at Neo4j.com/NODES. Your host is Tobias Macey and today I'm interviewing Jay Mishra about the applications for generative AI in the ETL process
Interview
Introduction How did you get involved in the area of data management? What are the different aspects/types of ETL that you are seeing generative AI applied to?
What kind of impact are you seeing in terms of time spent/quality of output/etc.?
What kinds of projects are most likely to benefit from the application of generative AI? Can you describe what a typical workflow of using AI to build ETL workflows looks like?
What are some of the types of errors that you are likely to experience from the AI? Once the pipeline is defined, what does the ongoing maintenance look like? Is the AI required to operate within the pipeline in perpetuity?
For individuals/teams/organizations who are experimenting with AI in their data engineering workflows, what are the concerns/questions that they are trying to address? What are the most interesting, innovative, or unexpected w
The fast approaching federal government shutdown is top of mind on this week’s podcast. Mark and Cris consider how the shutdown may play out and the economic consequences. Two other colleagues, Steve Cochrane and Stefan Angrick, then join the conversation to assess the all important Chinese and Japanese economies. Are the economic fortunes of these two massive economies flip-flopping? For China and Japan: Facing History, a book outlining the strained historical relationship between these two countries, click here Follow Mark Zandi @MarkZandi, Cris deRitis @MiddleWayEcon, and Marisa DiNatale on LinkedIn for additional insight.
Questions or Comments, please email us at [email protected]. We would love to hear from you. To stay informed and follow the insights of Moody's Analytics economists, visit Economic View.
Join me with Tableau expert Andy Kriebel about data visualization tips! 📊
Andy shares his expertise to help you level up your data skills, from creating purposeful dashboards to the power of dot maps.
Tune in now and take your data visualizations to the next level! 🎧
Connect with Andy Kriebel:
🤝 Connect on Linkedin
▶️ Subscribe to Youtube Channel
📔 Buy #MakeoverMonday Book
🎒 Learn About The Data School
🤝 Ace your data analyst interview with the interview simulator
📩 Get my weekly email with helpful data career tips
📊 Come to my next free “How to Land Your First Data Job” training
🏫 Check out my 10-week data analytics bootcamp
Timestamps:
(11:15) - When dealing with time series, your best friend is the versatile line chart 📈
(13:34) - Dive into dot maps for super-detailed visualizations 🌍
(17:30) Want to compare rankings over time? Bump chart!📊📈
(21:10) Keep your graphs simple and spiced up with context📉
(26:45) Dashboards should tell a story; ensure they have a purpose and context to keep folks engaged. 📋💡
(34:27) Prepping for an interview? Be chatty, be prepared with your interviewers. 🗣️🤝
Connect with Avery:
📺 Subscribe on YouTube
🎙Listen to My Podcast
👔 Connect with me on LinkedIn
🎵 TikTok
Mentioned in this episode: Join the last cohort of 2025! The LAST cohort of The Data Analytics Accelerator for 2025 kicks off on Monday, December 8th and enrollment is officially open!
To celebrate the end of the year, we’re running a special End-of-Year Sale, where you’ll get: ✅ A discount on your enrollment 🎁 6 bonus gifts, including job listings, interview prep, AI tools + more
If your goal is to land a data job in 2026, this is your chance to get ahead of the competition and start strong.
👉 Join the December Cohort & Claim Your Bonuses: https://DataCareerJumpstart.com/daa https://www.datacareerjumpstart.com/daa
Summary
The rapid growth of machine learning, especially large language models, have led to a commensurate growth in the need to store and compare vectors. In this episode Louis Brandy discusses the applications for vector search capabilities both in and outside of AI, as well as the challenges of maintaining real-time indexes of vector data.
Announcements
Hello and welcome to the Data Engineering Podcast, the show about modern data management Introducing RudderStack Profiles. RudderStack Profiles takes the SaaS guesswork and SQL grunt work out of building complete customer profiles so you can quickly ship actionable, enriched data to every downstream team. You specify the customer traits, then Profiles runs the joins and computations for you to create complete customer profiles. Get all of the details and try the new product today at dataengineeringpodcast.com/rudderstack This episode is brought to you by Datafold – a testing automation platform for data engineers that finds data quality issues before the code and data are deployed to production. Datafold leverages data-diffing to compare production and development environments and column-level lineage to show you the exact impact of every code change on data, metrics, and BI tools, keeping your team productive and stakeholders happy. Datafold integrates with dbt, the modern data stack, and seamlessly plugs in your data CI for team-wide and automated testing. If you are migrating to a modern data stack, Datafold can also help you automate data and code validation to speed up the migration. Learn more about Datafold by visiting dataengineeringpodcast.com/datafold You shouldn't have to throw away the database to build with fast-changing data. You should be able to keep the familiarity of SQL and the proven architecture of cloud warehouses, but swap the decades-old batch computation model for an efficient incremental engine to get complex queries that are always up-to-date. With Materialize, you can! It’s the only true SQL streaming database built from the ground up to meet the needs of modern data products. Whether it’s real-time dashboarding and analytics, personalization and segmentation or automation and alerting, Materialize gives you the ability to work with fresh, correct, and scalable results — all in a familiar SQL interface. Go to dataengineeringpodcast.com/materialize today to get 2 weeks free! If you’re a data person, you probably have to jump between different tools to run queries, build visualizations, write Python, and send around a lot of spreadsheets and CSV files. Hex brings everything together. Its powerful notebook UI lets you analyze data in SQL, Python, or no-code, in any combination, and work together with live multiplayer and version control. And now, Hex’s magical AI tools can generate queries and code, create visualizations, and even kickstart a whole analysis for you – all from natural language prompts. It’s like having an analytics co-pilot built right into where you’re already doing your work. Then, when you’re ready to share, you can use Hex’s drag-and-drop app builder to configure beautiful reports or dashboards that anyone can use. Join the hundreds of data teams like Notion, AllTrails, Loom, Mixpanel and Algolia using Hex every day to make their work more impactful. Sign up today at dataengineeringpodcast.com/hex to get a 30-day free trial of the Hex Team plan! Your host is Tobias Macey and today I'm interviewing Louis Brandy about building vector indexes in real-time for analytics and AI applications
Interview
Introduction How did you get involved in the area of data management? Can you describe what vector search is and how it differs from other search technologies?
What are the technical challenges related to providing vector search? What are the applications for vector search that merit the added complexity?
Vector databases have been gaining a lot of attention recently with the proliferation of LLM applicati
The amount of data generated from various processes and platforms has increased exponentially in the past decade, and the challenges of filtering useful data out of streams of raw data has become even greater. Meanwhile, the essence of making useful insights from that data has become even more important. In this incisive report, Federico Castanedo examines the challenges companies face when acting on data at rest as well as the benefits you unlock when acting on data as it's generated. Data engineers, enterprise architects, CTOs, and CIOs will explore the tools, processes, and mindset your company needs to process streaming data in real time. Learn how to make quick data-driven decisions to gain an edge on competitors. This report helps you: Explore gaps in today's real-time data architectures, including the limitations of real-time analytics to act on data immediately Examine use cases that can't be served efficiently with real-time analytics Understand how stream processing engines work with real-time data Learn how distributed data processing architectures, stream processing, streaming analytics, and event-based architectures relate to real-time data Understand how to transition from traditional batch processing environments to stream processing Federico Castanedo is an academic director and adjunct professor at IE University in Spain. A data science and AI leader, he has extensive experience in academia, industry, and startups.
Inside Economics welcomes Rob Fauber, President and CEO of Moody’s Corporation to the podcast. He discusses his concept of exponential risk and the opportunities and challenges of using AI. Rob shares what it's like to lead a global firm and answers a few get-to-know-you questions from Mark, Cris and Marisa. But before the conversation with Rob, the team discusses how worried they are about various economic threats and their potential impact on the macroeconomic economy. Guest: Rob Fauber, President & CEO of Moody’s Corporation For more insight into the era of exponential risk, click here. Follow Mark Zandi @MarkZandi, Cris deRitis @MiddleWayEcon, and Marisa DiNatale on LinkedIn for additional insight.
Questions or Comments, please email us at [email protected]. We would love to hear from you. To stay informed and follow the insights of Moody's Analytics economists, visit Economic View.
Mergulhamos no emocionante mundo da análise de dados na maior cervejaria do Brasil. Prepare-se para uma jornada fascinante, onde a paixão pela cerveja encontra a ciência dos dados. Em um papo, muito divertido, exploramos como a análise de dados desempenha um papel crucial desde o plantio, até a distribuição e qualidade consistente dos produtos icônicos da Ambev.
Neste episódio do Data Hackers — a maior comunidade de AI e Data Science do Brasil-, conheçam esse time de especialistas da Ambev Tech — núcleo de tecnologia da Ambev : o Daniel Cassiano — Diretor de Data & Analytics na América do Sul; a Gabriela Madia — Gerente de Governança de Dados & Analytics; o Daniel Henrique - Gerente de Engenharia de Plataforma de Dados, e o Felipe Contratres — Líder do Centro de Excelência de Analytics.
Lembrando que você pode encontrar todos os podcasts da comunidade Data Hackers no Spotify, iTunes, Google Podcast, Castbox e muitas outras plataformas. Caso queira, você também pode ouvir o episódio aqui no post mesmo!
Link do Medium:
Falamos no episódio
Conheça nosso convidado:
Daniel Cassiano — Diretor de Data & Analytics para América do Sul Gabriela Madia — Gerente de Governança de Dados & Analytics Daniel Henrique — Gerente de Engenharia de Plataforma de Dados Felipe Contratres — Líder do Centro de Excelência de Analytics
Bancada Data Hackers:
Paulo Vasconcellos Monique Femme Gabriel Lages
Links de referências:
Projeto que monitora o lado da rua que o caminhão estaciona, (post explicativo): https://www.instagram.com/reel/CxatwlqrGuV/?igshid=MzRlODBiNWFlZA== Projeto Diesel Analytics ( Projeto no Meetup Tech & Cheers em São Paulo): https://www.instagram.com/reel/CuHfmL8rQGI/?igshid=MzRlODBiNWFlZA== Link de Vagas (Ambev Tech) — https://ambevtech.gupy.io/ Ambev Tech (@ambevtech) — Fotos e vídeos do Instagram — https://www.instagram.com/ambevtech/ Linkedin: https://www.linkedin.com/company/ambevtech/ Ambev Tech Talk (Podcast com episódio mensal sobre Dados — Papo de Dados) : https://open.spotify.com/show/07cPNODgBHWh2JMkHbZxXG?si=2432135d6daa4a18
In this episode of the Data Career Podcast, I sit down with Luke Barousse, a data analyst and data star on YouTube, to discuss the impact of AI on the role of data analysts.
We delve into the question of whether AI will replace data analysts or enhance their work, sharing insights and perspectives on the topic.
Connect with Luke Barousse:
🤝 Connect on Linkedin
▶️ Subscribe on Youtube
📊 Datanerd.tech
📩 Get my weekly email with helpful data career tips
📊 Come to my next free “How to Land Your First Data Job” training
🏫 Check out my 10-week data analytics bootcamp
Timestamps:
(05:22) - ChatGPT Code Interpreter
(09:09) - Will AI Steal Your Job?
Connect with Avery:
📺 Subscribe on YouTube
🎙Listen to My Podcast
👔 Connect with me on LinkedIn
🎵 TikTok
Mentioned in this episode: Join the last cohort of 2025! The LAST cohort of The Data Analytics Accelerator for 2025 kicks off on Monday, December 8th and enrollment is officially open!
To celebrate the end of the year, we’re running a special End-of-Year Sale, where you’ll get: ✅ A discount on your enrollment 🎁 6 bonus gifts, including job listings, interview prep, AI tools + more
If your goal is to land a data job in 2026, this is your chance to get ahead of the competition and start strong.
👉 Join the December Cohort & Claim Your Bonuses: https://DataCareerJumpstart.com/daa https://www.datacareerjumpstart.com/daa
Today I’m joined by Anthony Deighton, General Manager of Data Products at Tamr. Throughout our conversation, Anthony unpacks his definition of a data product and we discuss whether or not he feels that Tamr itself is actually a data product. Anthony shares his views on why it’s so critical to focus on solving for customer needs and not simply the newest and shiniest technology. We also discuss the challenges that come with building a product that’s designed to facilitate the creation of better internal data products, as well as where we are in this new wave of data product management, and the evolution of the role.
Highlights/ Skip to:
I introduce Anthony, General Manager of Data Products at Tamr, and the topics we’ll be discussing today (00:37) Anthony shares his observations on how BI analytics are an inch deep and a mile wide due to the data that’s being input (02:31) Tamr’s focus on data products and how that reflects in Anthony’s recent job change from Chief Product Officer to General Manager of Data Products (04:35) Anthony’s definition of a data product (07:42) Anthony and I explore whether he feels that decision support is necessary for a data product (13:48) Whether or not Anthony feels that Tamr qualifies as a data product (17:08) Anthony speaks to the importance of focusing on outcomes and benefits as opposed to endlessly knitting together features and products (19:42) The challenges Anthony sees with metrics like Propensity to Churn (21:56) How Anthony thinks about design in a product like Tamr (30:43) Anthony shares how data science at Tamr is a tool in his toolkit and not viewed as a “fourth” leg of the product triad/stool (36:01) Anthony’s views on where we are in the evolution of the DPM role (41:25) What Anthony would do differently if he could start over at Tamr knowing what he knows now (43:43)
Links Tamr: https://www.tamr.com/ Innovating: https://www.amazon.com/Innovating-short-guide-making-things/dp/B0C8R79PVB The Mom Test: https://www.amazon.com/The-Mom-Test-Rob-Fitzpatrick-audiobook/dp/B07RJZKZ7F LinkedIn: https://www.linkedin.com/in/anthonydeighton/
Summary
A significant amount of time in data engineering is dedicated to building connections and semantic meaning around pieces of information. Linked data technologies provide a means of tightly coupling metadata with raw information. In this episode Brian Platz explains how JSON-LD can be used as a shared representation of linked data for building semantic data products.
Announcements
Hello and welcome to the Data Engineering Podcast, the show about modern data management This episode is brought to you by Datafold – a testing automation platform for data engineers that finds data quality issues before the code and data are deployed to production. Datafold leverages data-diffing to compare production and development environments and column-level lineage to show you the exact impact of every code change on data, metrics, and BI tools, keeping your team productive and stakeholders happy. Datafold integrates with dbt, the modern data stack, and seamlessly plugs in your data CI for team-wide and automated testing. If you are migrating to a modern data stack, Datafold can also help you automate data and code validation to speed up the migration. Learn more about Datafold by visiting dataengineeringpodcast.com/datafold Introducing RudderStack Profiles. RudderStack Profiles takes the SaaS guesswork and SQL grunt work out of building complete customer profiles so you can quickly ship actionable, enriched data to every downstream team. You specify the customer traits, then Profiles runs the joins and computations for you to create complete customer profiles. Get all of the details and try the new product today at dataengineeringpodcast.com/rudderstack You shouldn't have to throw away the database to build with fast-changing data. You should be able to keep the familiarity of SQL and the proven architecture of cloud warehouses, but swap the decades-old batch computation model for an efficient incremental engine to get complex queries that are always up-to-date. With Materialize, you can! It’s the only true SQL streaming database built from the ground up to meet the needs of modern data products. Whether it’s real-time dashboarding and analytics, personalization and segmentation or automation and alerting, Materialize gives you the ability to work with fresh, correct, and scalable results — all in a familiar SQL interface. Go to dataengineeringpodcast.com/materialize today to get 2 weeks free! If you’re a data person, you probably have to jump between different tools to run queries, build visualizations, write Python, and send around a lot of spreadsheets and CSV files. Hex brings everything together. Its powerful notebook UI lets you analyze data in SQL, Python, or no-code, in any combination, and work together with live multiplayer and version control. And now, Hex’s magical AI tools can generate queries and code, create visualizations, and even kickstart a whole analysis for you – all from natural language prompts. It’s like having an analytics co-pilot built right into where you’re already doing your work. Then, when you’re ready to share, you can use Hex’s drag-and-drop app builder to configure beautiful reports or dashboards that anyone can use. Join the hundreds of data teams like Notion, AllTrails, Loom, Mixpanel and Algolia using Hex every day to make their work more impactful. Sign up today at dataengineeringpodcast.com/hex to get a 30-day free trial of the Hex Team plan! Your host is Tobias Macey and today I'm interviewing Brian Platz about using JSON-LD for building linked-data products
Interview
Introduction How did you get involved in the area of data management? Can you describe what the term "linked data product" means and some examples of when you might build one?
What is the overlap between knowledge graphs and "linked data products"?
What is JSON-LD?
What are the domains in which it is typically used? How does it assist in developing linked data products?
what are the characterist
The looming UAW strike is top of mind, and no one better to talk to about how it may play out and what it means for the economy than Jonathan Smoke of COX Automotive and our own vehicle industry expert, Mike Brisson. Bernard Yaros also joins the podcast to talk about the consumer inflation report. Mark and Cris agree that while the current economic numbers look good, there’s plenty to worry about. For more from Jonathan Smoke, click here Follow Mark Zandi @MarkZandi, Cris deRitis @MiddleWayEcon, and Marisa DiNatale on LinkedIn for additional insight.
Questions or Comments, please email us at [email protected]. We would love to hear from you. To stay informed and follow the insights of Moody's Analytics economists, visit Economic View.
Hosted by Simplecast, an AdsWizz company. See pcm.adswizz.com for information about our collection and use of personal data for advertising.