talk-data.com talk-data.com

V

Speaker

Vishal Singh

2

talks

Head of Products Starburst

Filter by Event / Source

Talks & appearances

2 activities · Newest first

Search activities →

Today I’m joined by Vishal Singh, Head of Data Products at Starburst and co-author of the newly published e-book, Data Products for Dummies. Throughout our conversation, Vishal explains how the variations in definitions for a data product actually led to the creation of the e-book, and we discuss the differences between our two definitions. Vishal gives a detailed description of how he believes Data Product Managers should be conducting their discovery and gathering feedback from end users, and how his team evaluates whether their data products are truly successful and user-friendly.

Highlights/ Skip to:

I introduce Vishal, the Head of Data Products at Starburst and contributor of the e-book Data Products for Dummies (00:37) Vishal describes how his customers at Starburst all had a common problem, but differing definitions of a data product, which led to the creation of his e-book (01:15) Vishal shares his one-sentence definition of a data product (02:50) How Vishal’s definition of a data product differs from mine, and we both expand on the possibilities between the two (05:33) The tactics Vishal uses to useful feedback to ensure the data products he develops are valuable for end users (07:48) Why Vishal finds it difficult to get one on one feedback from users during the iteration phase of data product development (11:07) The danger of sunk cost bias in the iteration phase of data product development (13:10) Vishal describes how he views the role of a DPM when it comes to doing effective initial discovery (15:27) How Vishal structures his teams and their interactions with each other and their end users (21:34) Vishal’s thoughts on how design affects both data scientists and end users (24:16) How DPMs at Starburst evaluate if the data product design is user-friendly (28:45) Vishal’s views on where Designers are valuable in the data product development process (35:00) Vishal and I discuss the importance of ensuring your products truly solve your user’s problems (44:44) Where you can learn more about Vishal’s upcoming events and the e-book, Data Products for Dummies (49:48)

Links Starburst: https://www.starburst.io/ Data Products for Dummies: https://www.starburst.io/info/data-products-for-dummies/ “How to Measure the Impact of Data Products with Doug Hubbard”: https://designingforanalytics.com/resources/episodes/080-how-to-measure-the-impact-of-data-productsand-anything-else-with-forecasting-and-measurement-expert-doug-hubbard/ Trino Summit: https://www.starburst.io/info/trinosummit2023/ Galaxy Platform: https://www.starburst.io/platform/starburst-galaxy/ Datanova Summit: https://www.starburst.io/datanova/ LinkedIn: https://www.linkedin.com/in/singhsvishal/ Twitter: https://twitter.com/vishal_singh

Summary

With all of the messaging about treating data as a product it is becoming difficult to know what that even means. Vishal Singh is the head of products at Starburst which means that he has to spend all of his time thinking and talking about the details of product thinking and its application to data. In this episode he shares his thoughts on the strategic and tactical elements of moving your work as a data professional from being task-oriented to being product-oriented and the long term improvements in your productivity that it provides.

Announcements

Hello and welcome to the Data Engineering Podcast, the show about modern data management When you're ready to build your next pipeline, or want to test out the projects you hear about on the show, you'll need somewhere to deploy it, so check out our friends at Linode. With their new managed database service you can launch a production ready MySQL, Postgres, or MongoDB cluster in minutes, with automated backups, 40 Gbps connections from your application hosts, and high throughput SSDs. Go to dataengineeringpodcast.com/linode today and get a $100 credit to launch a database, create a Kubernetes cluster, or take advantage of all of their other services. And don't forget to thank them for their continued support of this show! Modern data teams are dealing with a lot of complexity in their data pipelines and analytical code. Monitoring data quality, tracing incidents, and testing changes can be daunting and often takes hours to days or even weeks. By the time errors have made their way into production, it’s often too late and damage is done. Datafold built automated regression testing to help data and analytics engineers deal with data quality in their pull requests. Datafold shows how a change in SQL code affects your data, both on a statistical level and down to individual rows and values before it gets merged to production. No more shipping and praying, you can now know exactly what will change in your database! Datafold integrates with all major data warehouses as well as frameworks such as Airflow & dbt and seamlessly plugs into CI workflows. Visit dataengineeringpodcast.com/datafold today to book a demo with Datafold. RudderStack helps you build a customer data platform on your warehouse or data lake. Instead of trapping data in a black box, they enable you to easily collect customer data from the entire stack and build an identity graph on your warehouse, giving you full visibility and control. Their SDKs make event streaming from any app or website easy, and their extensive library of integrations enable you to automatically send data to hundreds of downstream tools. Sign up free at dataengineeringpodcast.com/rudder Build Data Pipelines. Not DAGs. That’s the spirit behind Upsolver SQLake, a new self-service data pipeline platform that lets you build batch and streaming pipelines without falling into the black hole of DAG-based orchestration. All you do is write a query in SQL to declare your transformation, and SQLake will turn it into a continuous pipeline that scales to petabytes and delivers up to the minute fresh data. SQLake supports a broad set of transformations, including high-cardinality joins, aggregations, upserts and window operations. Output data can be streamed into a data lake for query engines like Presto, Trino or Spark SQL, a data warehouse like Snowflake or Redshift., or any other destination you choose. Pricing for SQLake is simple. You pay $99 per terabyte ingested into your data lake using SQLake, and run unlimited transformation pipelines for free. That way data engineers and data users can process to their heart’s content without worrying about their cloud bill. For data engineering podcast listeners, we’re offering a 30 day trial with unlimited data, so go to dataengineeringpodcast.com/upsolver today and see for yourself how to avoid DAG hell. Your host is Tobias Macey and today I'm interviewing Vishal Singh about his experience