talk-data.com talk-data.com

Topic

AI/ML

Artificial Intelligence/Machine Learning

data_science algorithms predictive_analytics

9014

tagged

Activity Trend

1532 peak/qtr
2020-Q1 2026-Q1

Activities

9014 activities · Newest first

As AI continually changes how businesses operate, new questions emerge around ethics and privacy. Nowadays, algorithms can set prices and personalize offers, but how do companies ensure they’re doing this responsibly? What does it mean to be transparent with customers about data use, and how can businesses avoid unintended bias? Balancing innovation with trust is key, but achieving this balance isn’t always straightforward. Dr. Jose Mendoza is Academic Director and Clinical Associate Professor in Integrated Marketing at NYU, and was formerly an Associate Professor of Practice at The University of Arizona in Tucson, Arizona. His focus is on consumer pricing, digital retailing, intelligent retail stores, neuromarketing, big data, artificial intelligence, and machine learning. Previously, he taught marketing courses at Sacred Heart University and Western Michigan University. He is also an experienced senior global marketing executive with over 18 years of experience in global marketing alone and a career as an Engineer in Information Sciences. Dr. Mendoza is also a Doctoral Researcher in Strategic and Global pricing, Consumer Behavior, and Pricing Research methodologies. He had international roles in Latin America, Europe, and the USA with scope in over 50 countries.  In the episode, Richie and Jose explore AI-driven pricing, consumer perceptions and ethical pricing, the complexity of dynamic pricing models, explainable AI, data privacy and customer trust, legal and ethical guardrails, innovations in dynamic pricing and much more.  Links Mentioned in the Show: NYUConnect with JoseAmazon Dynamic Pricing Strategy in 2024Course: AI EthicsRelated Episode: The Future of Marketing Analytics with Cory Munchbach, CEO at BlueConicSign up to RADAR: Forward Edition New to DataCamp? Learn on the go using the DataCamp mobile appEmpower your business with world-class data and AI skills with DataCamp for business

I speak at a lot of conferences, and I've lost track of how many questions I've answered. Since conferences are top of mind for me right now, here are some tips for asking good (and bad) questions of speakers.

My works:

📕Fundamentals of Data Engineering: https://www.oreilly.com/library/view/fundamentals-of-data/9781098108298/

🎥 Deeplearning.ai Data Engineering Certificate: https://www.coursera.org/professional-certificates/data-engineering

🔥Practical Data Modeling: https://practicaldatamodeling.substack.com/

🤓 My SubStack: https://joereis.substack.com/

Apache Spark for Machine Learning

Dive into the power of Apache Spark as a tool for handling and processing big data required for machine learning. With this book, you will explore how to configure, execute, and deploy machine learning algorithms using Spark's scalable architecture and learn best practices for implementing real-world big data solutions. What this Book will help me do Understand the integration of Apache Spark with large-scale infrastructures for machine learning applications. Employ data processing techniques for preprocessing and feature engineering efficiently with Spark. Master the implementation of advanced supervised and unsupervised learning algorithms using Spark. Learn to deploy machine learning models within Spark ecosystems for optimized performance. Discover methods for analyzing big data trends and machine learning model tuning for improved accuracy. Author(s) The author, Deepak Gowda, is an experienced data scientist with over ten years of expertise in machine learning and big data. His career spans industries such as supply chain, cybersecurity, and more where he has utilized Apache Spark extensively. Deepak's teaching style is marked by clarity and practicality, making complex concepts approachable. Who is it for? Apache Spark for Machine Learning is tailored for data engineers, machine learning practitioners, and computer science students looking to advance their ability to process, analyze, and model using large datasets. If you're already familiar with basic machine learning and want to scale your solutions using Spark, this book is ideal for your studies and professional growth.

In this episode of The Deep Dive, we explore Retrieval-Augmented Generation, or RAG, and its revolutionary impact on AI. We break down five game-changing applications of RAG, each transforming how AI interacts with real-time data and complex information. Discover how RAG is enhancing everything from customer service to academic research, by tackling challenges like outdated information and static AI models. Key Highlights: Real-time Q&A Systems: How RAG ensures that AI provides the most up-to-date answers, making customer support smarter and more reliable.Dynamic Content Creation: No more stale reports—learn how RAG allows for content that updates in real-time.Multi-Source Summarization: Summarizing complex, often conflicting information from multiple sources for balanced insights.Intelligent Chatbots: Discover how RAG-driven chatbots bring up-to-the-minute responses, improving user experience in real-time.Specialized Knowledge Integration: From medical diagnoses to legal precedents, see how RAG is revolutionizing fields requiring precise, specialized knowledge.Tune in to see how RAG is shaping the future of AI, making it more adaptable, intelligent, and responsive to our world’s ever-changing landscape! Resources: Article: "5 Game-Changing Techniques to Boost Your NLP Projects with Retrieval Augmented Generation"Explore hands-on with RAG at Hugging FaceResearch and community forums for deeper learning and discussions on RAG

Send us a text Welcome to Datatopics Unplugged, where the tech world’s buzz meets laid-back banter. In each episode, we dive into the latest in AI, data science, and technology—perfect for your inner geek or curious mind. Pull up a seat, tune in, and join us for insights, laughs, and the occasional hot take on the digital world.

In this episode, we are joined by Vitale to discuss:

Meta’s video generation breakthrough: Explore Meta’s new “MovieGen” model family that generates hyper-realistic, 16-second video clips with reflections, consistent spatial details, and multi-frame coherence. Also discussed: Sora, a sneak peek at Meta’s open-source possibilities. For a look back, check out this classic AI-generated video of Will Smith eating spaghetti. Anthropic’s Claude 3.5 updates: Meet Claude 3.5 and its “computer use” feature, letting it navigate your screen for you. Easily fine-tune & train LLMs, faster with Unsloth: Discover tools that simplify model fine-tuning and deployment, making it easier for small-scale developers to harness AI’s power. Don’t miss Gerganov’s GitHub contributions in this space, too. Deno 2.0 release hype: With a splashy promo video, Deno’s JavaScript runtime enters the scene as a streamlined, secure alternative to Node.js.

In this special Halloween episode, Jason Foster talks to James Lupton, CTO at Cynozure, and Keith Goldthorpe, Director of Solution Architecture at Cynozure. Reflecting on their conversation from last year, the group revisit some of their predictions and discuss the most significant developments in the realms of data and AI. They also engage in a game of 'AI Spell or Human Skill,' where they debate whether humans or AI tools would outperform one another in a series of spellbinding scenarios.  *****    Cynozure is a leading data, analytics and AI company that helps organisations to reach their data potential. It works with clients on data and AI strategy, data management, data architecture and engineering, analytics and AI, data culture and literacy, and data leadership. The company was named one of The Sunday Times' fastest-growing private companies in both 2022 and 2023, and recognised as The Best Place to Work in Data by DataIQ in 2023 and 2024. 

Artificial Intelligence for Cybersecurity

Explore how artificial intelligence can transform your cybersecurity strategies with "Artificial Intelligence for Cybersecurity". This book provides practical insights into applying AI methods to a variety of cybersecurity problems, from malware analysis to threat detection. By understanding these concepts, you'll gain the knowledge needed to protect your organization's data and networks effectively. What this Book will help me do Understand how AI methods can address cybersecurity concerns effectively. Develop practical skills using AI tools to combat cyber threats. Design AI-powered solutions for malware identification and anomaly detection. Navigate real-world applications of AI in cybersecurity scenarios. Recognize and mitigate common pitfalls while implementing AI methods in cybersecurity. Author(s) The authors, Bojan Kolosnjaji, Huang Xiao, Peng Xu, and Apostolis Zarras, are experts in machine learning and cybersecurity. With extensive backgrounds in both academia and industry, they bring a wealth of knowledge to the book. Their practical and educational approach makes complex AI and cybersecurity concepts accessible, empowering readers to apply these methods to real-world problems. Who is it for? This book is ideal for professionals in cybersecurity who are keen to integrate AI techniques into their frameworks and workflows. It's also suitable for machine learning enthusiasts who want to delve into the realm of cybersecurity. If you possess a basic understanding of Python programming and machine learning fundamentals, this book will guide you through to advanced concepts. Whether you are a student or an industry veteran, this book offers valuable insights for enhancing your cybersecurity strategies with AI.

Building Modern Data Applications Using Databricks Lakehouse

This book, "Building Modern Data Applications Using Databricks Lakehouse," provides a comprehensive guide for data professionals to master the Databricks platform. You'll learn to effectively build, deploy, and monitor robust data pipelines with Databricks' Delta Live Tables, empowering you to manage and optimize cloud-based data operations effortlessly. What this Book will help me do Understand the foundations and concepts of Delta Live Tables and its role in data pipeline development. Learn workflows to process and transform real-time and batch data efficiently using the Databricks lakehouse architecture. Master the implementation of Unity Catalog for governance and secure data access in modern data applications. Deploy and automate data pipeline changes using CI/CD, leveraging tools like Terraform and Databricks Asset Bundles. Gain advanced insights in monitoring data quality and performance, optimizing cloud costs, and managing DataOps tasks effectively. Author(s) Will Girten, the author, is a seasoned Solutions Architect at Databricks with over a decade of experience in data and AI systems. With a deep expertise in modern data architectures, Will is adept at simplifying complex topics and translating them into actionable knowledge. His books emphasize real-time application and offer clear, hands-on examples, making learning engaging and impactful. Who is it for? This book is geared towards data engineers, analysts, and DataOps professionals seeking efficient strategies to implement and maintain robust data pipelines. If you have a basic understanding of Python and Apache Spark and wish to delve deeper into the Databricks platform for streamlining workflows, this book is tailored for you.

Microsoft Power Apps Cookbook - Third Edition

Microsoft Power Apps Cookbook is a comprehensive guide to harnessing the full potential of Microsoft Power Apps, a powerful low-code platform for building business applications. Packed with practical recipes, this book details how to develop scalable, efficient apps, automate workflows with RPA, and utilize new capabilities like AI-powered Microsoft Copilot and the Power Apps Component Framework. What this Book will help me do Create and deploy scalable canvas and model-driven apps using Microsoft Power Apps. Utilize AI-powered features like Copilot to speed up app creation and development. Implement robust data management strategies with Microsoft Dataverse. Extend app functionalities using the Power Apps Component Framework for custom components. Design and build secure external-facing websites with Microsoft Power Pages. Author(s) Eickhel Mendoza is an experienced Microsoft Power Platform developer and educator who has helped numerous organizations enhance their capabilities through low-code app development. Authoring from extensive hands-on experience, their teaching style bridges technical theory and practical application. Eickhel is passionate about empowering users to achieve more with modern app development tools. Who is it for? This book is ideal for information workers and developers looking to streamline their application development processes with Microsoft's low-code solutions. It is particularly targeted toward users with a foundational understanding of the Power Platform looking to deepen their knowledge. Readers will benefit most if they are eager to learn how to create innovative solutions efficiently. Traditional developers aiming to explore a new paradigm of rapid application development will also find it highly beneficial.

Está no ar, o Data Hackers News !! Os assuntos mais quentes da semana, com as principais notícias da área de Dados, IA e Tecnologia, que você também encontra na nossa Newsletter semanal, agora no Podcast do Data Hackers !!

Aperte o play e ouça agora, o Data Hackers News dessa semana !

Para saber tudo sobre o que está acontecendo na área de dados, se inscreva na Newsletter semanal:

⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠https://www.datahackers.news/⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠

Conheça nossos comentaristas do Data Hackers News:

⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠Monique Femme⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠

Paulo Vasconcellos

⁠Matérias/assuntos comentados:

CrewAI levanta mais de 12 milhões de dólares;

Inscreva-se no Meetup BEES DAY/TA sobre Personalização de Dados na Prática;

Estagiário sabota projeto de IA da dona do TikTok;

Responda a pesquisa do State of Data Brazil.

Demais canais do Data Hackers:

⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠Site⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠

⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠Linkedin⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠

⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠Instagram⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠

⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠Tik Tok⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠

⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠You Tube⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠

Send us a text Peter Wang, Chief AI and Innovation Officer and Co-founder of Anaconda, is back. Peter discusses Anaconda’s AI incubator. "It turns out Python is still a thing". Python and much, much more. 01:27 Meet Peter Wang05:10 Python or R05:51 Anaconda's Differentiation07:08 Why the Name Anaconda08:24 The AI Incubator11:40 GenAI14:39 Enter Python16:08 Anaconda Commercial Services18:40 Security20:57 Common Points of Failure22:53 Branding24:50 watsonx Partnership28:40 AI Risks34:13 Getting Philosophical36:13 China44:52 Leadership StyleLinkedin: linkedin.com/in/pzwang Website: https://www.linkedin.com/company/anacondainc/, https://www.anaconda.com/ Want to be featured as a guest on Making Data Simple? Reach out to us at [email protected] and tell us why you should be next. The Making Data Simple Podcast is hosted by Al Martin, WW VP Technical Sales, IBM, where we explore trending technologies, business innovation, and leadership ... while keeping it simple & fun. Want to be featured as a guest on Making Data Simple? Reach out to us at [email protected] and tell us why you should be next. The Making Data Simple Podcast is hosted by Al Martin, WW VP Technical Sales, IBM, where we explore trending technologies, business innovation, and leadership ... while keeping it simple & fun.

Episode SummaryIn this episode, we dive into the power of AI to tackle the often-overwhelming world of PDFs and technical documents. We explore an innovative tool that makes PDFs more accessible and actionable, from summarizing key insights to generating audio and even preparing Q&A. If you work in data science, AI, or any field that requires you to stay up-to-date with extensive documentation and research, this tool could be your new best friend. Topics Covered: * The PDF Dilemma * How data professionals face information overload from research papers, reports, and white papers. * Why keeping up with technical documents can feel like a “black hole” for your time. * AI-Powered PDF Assistance * Overview of an AI tool that leverages PyPDF2 and HuggingFace for seamless PDF extraction and summarization. * Using Google Text-to-Speech to convert summaries into audio for learning on the go. * Interactive Content Generation * How the tool creates a more interactive PDF experience by generating questions and answers. * Scenarios where this could be useful: preparing for presentations, understanding dense research, and managing technical documentation. * Real-World Scenarios and Use Cases * Examples of how a data scientist, data analyst, or any professional could save time and improve understanding. * AI as a “study buddy” for deeper learning and faster, more efficient information processing. * Balancing AI with Critical Thinking * The importance of using AI as a tool rather than a replacement for human expertise. * How AI challenges us to become more thoughtful consumers of information and better thinkers overall. Key Takeaways: * Save Time and Boost UnderstandingEmbrace AI to extract core insights from complex documents, potentially freeing up hours each week to focus on high-impact tasks. * Learn on the GoTurn PDF content into audio to make commuting, exercising, or downtime more productive. * Engage with Information InteractivelyUse the tool’s Q&A generation feature to explore documents in a more interactive way, perfect for preparing presentations or deep-diving into research. Final Thought:Imagine applying this technology not only to PDFs but also to other information sources like websites, articles, and even books. As AI continues to evolve, how might it transform the way we learn, work, and think? Call to ActionIf this resonates with you, let us know! Share what types of PDFs or documents you’d tackle with this AI-powered tool and how you think it could change your workflow. Don’t forget to subscribe for more insights on the latest AI tools and how they’re shaping the future of work and learning. Link: Blog Post Link This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit mukundansankar.substack.com

Help us become the #1 Data Podcast by leaving a rating & review! We are 67 reviews away! Big changes are happening in the data world, and it’s not just about AI! It’s a mix of challenges and new chances in the data field. Let’s dig into what’s happening and why now’s the time to rethink your next career move. 💌 Join 30k+ aspiring data analysts & get my tips in your inbox weekly 👉 https://www.datacareerjumpstart.com/newsletter 🆘 Feeling stuck in your data journey? Come to my next free "How to Land Your First Data Job" training 👉 https://www.datacareerjumpstart.com/training 👩‍💻 Want to land a data job in less than 90 days? 👉 https://www.datacareerjumpstart.com/daa 👔 Ace The Interview with Confidence 👉 https://www.datacareerjumpstart.com//interviewsimulator 🔗 LIVE DATA TECHNOLOGIES: https://www.livedatatechnologies.com/ ⌚ TIMESTAMPS 01:10 - Data-Driven Insights on the Job Market 02:18 - The Rise of Data Engineering 03:49 - AI's Impact on Data Roles 04:44 - Data Analyst Jobs Are Still Growing 06:27 - Job Hopping in Data Roles 🔗 CONNECT WITH AVERY 🎥 YouTube Channel 🤝 LinkedIn 📸 Instagram 🎵 TikTok 💻 Website Mentioned in this episode: Join the last cohort of 2025! The LAST cohort of The Data Analytics Accelerator for 2025 kicks off on Monday, December 8th and enrollment is officially open!

To celebrate the end of the year, we’re running a special End-of-Year Sale, where you’ll get: ✅ A discount on your enrollment 🎁 6 bonus gifts, including job listings, interview prep, AI tools + more

If your goal is to land a data job in 2026, this is your chance to get ahead of the competition and start strong.

👉 Join the December Cohort & Claim Your Bonuses: https://DataCareerJumpstart.com/daa https://www.datacareerjumpstart.com/daa

The relationship between AI and ethics is both developing and delicate. On one hand, the GenAI advancements to date are impressive. On the other, extreme care needs to be taken as this tech continues to quickly become more commonplace in our lives. In today’s episode, Ovetta Sampson and I examine the crossroads ahead for designing AI and GenAI user experiences.

While professionals and the general public are eager to embrace new products, recent breakthroughs, etc.; we still need to have some guard rails in place. If we don’t, data can easily get mishandled, and people could get hurt. Ovetta possesses firsthand experience working on these issues as they sprout up. We look at who should be on a team designing an AI UX, exploring the risks associated with GenAI, ethics, and need to be thinking about going forward.

Highlights/ Skip to: (1:48) Ovetta's background and what she brings to Google’s Core ML group (6:03) How Ovetta and her team work with data scientists and engineers deep in the stack (9:09)  How AI is changing the front-end of applications (12:46) The type of people you should seek out to design your AI and LLM UXs (16:15) Explaining why we’re only at the very start of major GenAI breakthroughs (22:34) How GenAI tools will alter the roles and responsibilities of designers, developers, and product teams (31:11) The potential harms of carelessly deploying GenAI technology (42:09) Defining acceptable levels of risk when using GenAI in real-world applications (53:16) Closing thoughts from Ovetta and where you can find her

Quotes from Today’s Episode “If artificial intelligence is just another technology, why would we build entire policies and frameworks around it? The reason why we do that is because we realize there are some real thorny ethical issues [surrounding AI]. Who owns that data? Where does it come from? Data is created by people, and all people create data. That’s why companies have strong legal, compliance, and regulatory policies around [AI], how it’s built, and how it engages with people. Think about having a toddler and then training the toddler on everything in the Library of Congress and on the internet. Do you release that toddler into the world without guardrails? Probably not.” - Ovetta Sampson (10:03) “[When building a team] you should look for a diverse thinker who focuses on the limitations of this technology- not its capability. You need someone who understands that the end destination of that technology is an engagement with a human being.  You need somebody who understands how they engage with machines and digital products. You need that person to be passionate about testing various ways that relationships can evolve. When we go from execution on code to machine learning, we make a shift from [human] agency to a shared-agency relationship. The user and machine both have decision-making power. That’s the paradigm shift that [designers] need to understand. You want somebody who can keep that duality in their head as they’re testing product design.” - Ovetta Sampson (13:45) “We’re in for a huge taxonomy change. There are words that mean very specific definitions today. Software engineer. Designer. Technically skilled. Digital. Art. Craft. AI is changing all that. It’s changing what it means to be a software engineer. Machine learning used to be the purview of data scientists only, but with GenAI, all of that is baked in to Gemini. So, now you start at a checkpoint, and you’re like, all right, let’s go make an API, right? So, the skills, the understanding, the knowledge, the taxonomy even, how we talk about these things, how do we talk about the machine who speaks to us talks to us, who could create a podcast out of just voice memos?” - Ovetta Sampson (24:16) “We have to be very intentional [when building AI tools], and that’s the kind of folks you want on teams. [Designers] have to go and play scary scenarios. We have to do that. No designer wants to be “Negative Nancy,” but this technology has huge potential to harm. It has harmed. If we don’t have the skill sets to recognize, document, and minimize harm, that needs to be part of our skill set.  If we’re not looking out for the humans, then who actually is?” - Ovetta Sampson (32:10) “[Research shows] things happen to our brain when we’re exposed to artificial intelligence… there are real human engagement risks that are an opportunity for design.  When you’re designing a self-driving car, you can’t just let the person go to sleep unless the car is fully [automated] and every other car on the road is self-driving. If there are humans behind the wheel, you need to have a feedback loop system—something that’s going to happen [in case] the algorithm is wrong. If you don’t have that designed, there’s going to be a large human engagement risk that a car is going to run over somebody who’s [for example] pushing a bike up a hill[...] Why? The car could not calculate the right speed and pace of a person pushing their bike. It had the speed and pace of a person walking, the speed and pace of a person on a bike, but not the two together. Algorithms will be wrong, right?” - Ovetta Sampson (39:42) “Model goodness used to be the purview of companies and the data scientists. Think about the first search engines. Their model goodness was [about] 77%. That’s good, right? And then people started seeing photos of apes when [they] typed in ‘black people.’ Companies have to get used to going to their customers in a wide spectrum and asking them when they’re [models or apps are] right and wrong.  They can’t take on that burden themselves anymore. Having ethically sourced data input and variables is hard work. If you’re going to use this technology, you need to put into place the governance that needs to be there.” - Ovetta Sampson (44:08)

podcast_episode
by Val Kroll , Martin Broadhurst , Julie Hoyer , Tim Wilson (Analytics Power Hour - Columbus (OH) , Moe Kiss (Canva) , Michael Helbling (Search Discovery)

udging by the number of inbound pitches we get from PR firms, AI is absolutely going to replace most of the work of the analyst some time in the next few weeks. It's just a matter of time until some startup gets enough market traction to make that happen (business tip: niche podcasts are likely not a productive path to market dominance, no matter what Claude from Marketing says). We're skeptical. But that doesn't mean we don't think there are a lot of useful applications of generative AI for the analyst. We do! As Moe posited in this episode, one useful analogy is that thinking of using generative AI effectively is like getting a marketer effectively using MMM when they've been living in an MTA world (it's more nuanced and complicated). Our guest (NOT from a PR firm solicitation!), Martin Broadhurst, agreed: it's dicey to fully embrace generative AI without some understanding of what it's actually doing. Things got a little spicy, but no humans or AI were harmed in the making of the episode. For complete show notes, including links to items mentioned in this episode and a transcript of the show, visit the show page.

Kirk is joined in the studio by Ken Moreano, CEO & President of Scott Data Center, and Bill Kleyman, CEO of Apolo, to chat about the evolving role of AI in data centers, the importance of education in adopting new technologies, real-world examples of AI applications, and the financial aspects of AI services.

For more about us: https://linktr.ee/overwatchmissioncritical

With the recent rapid advancements in AI comes the challenge of navigating an ever-changing field of play, while ensuring the tech we use serves real-world needs. As AI becomes more ingrained in business and everyday life, how do we balance cutting-edge development with practicality and ethical responsibility? What steps are necessary to ensure AI’s growth benefits society, aligns with human values, and avoids potential risks? What similarities can we draw between the way we think, and the way AI thinks for us? Terry Sejnowski is one of the most influential figures in computational neuroscience. At the Salk Institute for Biological Studies, he runs the Computational Neurobiology Laboratory, and hold the Francis Crick Chair. At the University of California, San Diego, he is a Distinguished Professor and runs a neurobiology lab. Terry is also the President of the Neural Information Processing (NIPS) Foundation, and an organizer of the NeurIPS AI conference. Alongside Geoff Hinton, Terry co-invented the Boltzmann machine technique for machine learning. He is the author of over 500 journal articles on neuroscience and AI, and the book "ChatGPT and the Future of AI". In the episode, Richie and Terry explore the current state of AI, historical developments in AI, the NeurIPS conference, collaboration between AI and neuroscience, AI’s shift from academia to industry, large vs small LLMs, creativity in AI, AI ethics, autonomous AI, AI agents, superintelligence, and much more.  Links Mentioned in the Show: NeurIPS ConferenceTerry’s Book—ChatGPT and the Future of AI: The Deep Language RevolutionConnect with TerryTerry on SubstackCourse: Data Communication ConceptsRelated Episode: Guardrails for the Future of AI with Viktor Mayer-Schönberger, Professor of Internet Governance and Regulation at the University of OxfordSign up to RADAR: Forward Edition New to DataCamp? Learn on the go using the DataCamp mobile appEmpower your business with world-class data and AI skills with DataCamp for business

Summary Gleb Mezhanskiy, CEO and co-founder of DataFold, joins Tobias Macey to discuss the challenges and innovations in data migrations. Gleb shares his experiences building and scaling data platforms at companies like Autodesk and Lyft, and how these experiences inspired the creation of DataFold to address data quality issues across teams. He outlines the complexities of data migrations, including common pitfalls such as technical debt and the importance of achieving parity between old and new systems. Gleb also discusses DataFold's innovative use of AI and large language models (LLMs) to automate translation and reconciliation processes in data migrations, reducing time and effort required for migrations. Announcements Hello and welcome to the Data Engineering Podcast, the show about modern data managementImagine catching data issues before they snowball into bigger problems. That’s what Datafold’s new Monitors do. With automatic monitoring for cross-database data diffs, schema changes, key metrics, and custom data tests, you can catch discrepancies and anomalies in real time, right at the source. Whether it’s maintaining data integrity or preventing costly mistakes, Datafold Monitors give you the visibility and control you need to keep your entire data stack running smoothly. Want to stop issues before they hit production? Learn more at dataengineeringpodcast.com/datafold today!Your host is Tobias Macey and today I'm welcoming back Gleb Mezhanskiy to talk about Datafold's experience bringing AI to bear on the problem of migrating your data stackInterview IntroductionHow did you get involved in the area of data management?Can you describe what the Data Migration Agent is and the story behind it?What is the core problem that you are targeting with the agent?What are the biggest time sinks in the process of database and tooling migration that teams run into?Can you describe the architecture of your agent?What was your selection and evaluation process for the LLM that you are using?What were some of the main unknowns that you had to discover going into the project?What are some of the evolutions in the ecosystem that occurred either during the development process or since your initial launch that have caused you to second-guess elements of the design?In terms of SQL translation there are libraries such as SQLGlot and the work being done with SDF that aim to address that through AST parsing and subsequent dialect generation. What are the ways that approach is insufficient in the context of a platform migration?How does the approach you are taking with the combination of data-diffing and automated translation help build confidence in the migration target?What are the most interesting, innovative, or unexpected ways that you have seen the Data Migration Agent used?What are the most interesting, unexpected, or challenging lessons that you have learned while working on building an AI powered migration assistant?When is the data migration agent the wrong choice?What do you have planned for the future of applications of AI at Datafold?Contact Info LinkedInParting Question From your perspective, what is the biggest gap in the tooling or technology for data management today?Closing Announcements Thank you for listening! Don't forget to check out our other shows. Podcast.init covers the Python language, its community, and the innovative ways it is being used. The AI Engineering Podcast is your guide to the fast-moving world of building AI systems.Visit the site to subscribe to the show, sign up for the mailing list, and read the show notes.If you've learned something or tried out a project from the show then tell us about it! Email [email protected] with your story.Links DatafoldDatafold Migration AgentDatafold data-diffDatafold Reconciliation Podcast EpisodeSQLGlotLark parserClaude 3.5 SonnetLookerPodcast EpisodeThe intro and outro music is from The Hug by The Freak Fandango Orchestra / CC BY-SA

The JPMorgan conference delivered a consensus around a Goldilocks macro scenario even as participants say US election outcomes look like a coin-toss. The risks of a blue versus red wave, and versus divided government lead to varied outcomes on growth, inflation, and the Fed. All of these are discussed.

Speakers:

Bruce Kasman

Joseph Lupton

This podcast was recorded on 25 October 2024.

This communication is provided for information purposes only. Institutional clients please visit www.jpmm.com/research/disclosures for important disclosures. © 2024 JPMorgan Chase & Co. All rights reserved. This material or any portion hereof may not be reprinted, sold or redistributed without the written consent of J.P. Morgan. It is strictly prohibited to use or share without prior written consent from J.P. Morgan any research material received from J.P. Morgan or an authorized third-party (“J.P. Morgan Data”) in any third-party artificial intelligence (“AI”) systems or models when such J.P. Morgan Data is accessible by a third-party. It is permissible to use J.P. Morgan Data for internal business purposes only in an AI system or model that protects the confidentiality of J.P. Morgan Data so as to prevent any and all access to or use of such J.P. Morgan Data by any third-party.

We talked about:

00:00 DataTalks.Club intro

00:00 DataTalks.Club anniversary "Ask Me Anything" event with Alexey Grigorev

02:29 The founding of DataTalks .Club

03:52 Alexey's transition from Java work to DataTalks.Club

04:58 Growth and success of DataTalks.Club courses

12:04 Motivation behind creating a free-to-learn community

24:03 Staying updated in data science through pet projects

26 :37 Hosting a second podcast and maintaining programming skills

28:56 Skepticism about LLMs and their relevance

31:53 Transitioning to DataTalks.Club and personal reflections

33:32 Memorable moments and the first event's success

36:19 Community building during the pandemic

38:31 AI's impact on data analysts and future roles

42:24 Discussion on AI in healthcare

44:37 Age and reflections on personal milestones

47:54 Building communities and personal connections

49:34 Future goals for the community and courses

51:18 Community involvement and engagement strategies

53:46 Ideas for competitions and hackathons

54:20 Inviting guests to the podcast

55:29 Course updates and future workshops

56:27 Podcast preparation and research process

58:30 Career opportunities in data science and transitioning fields

1:01 :10 Book recommendations and personal reading experiences

About the speaker:

Alexey Grigorev is the founder of DataTalks.Club.

Join our slack: https://datatalks.club/slack.html