talk-data.com
People (8 results)
See all 8 →Activities & events
| Title & Speakers | Event |
|---|---|
|
Kyle Stratis
2026-01-21 · 19:00
Kyle Stratis
– Founder
@ Stratis Data Labs
Speaker: Kyle Stratis, Founder at Stratis Data Labs |
|
|
Dr. Ali Arsanjani
2026-01-21 · 19:00
Dr. Ali Arsanjani
– Director of Applied AI Engineering; Head of AI Center of Excellence
@ Google Cloud
Speaker: Dr. Ali Arsanjani, Director of Applied AI Engineering; Head of AI Center of Excellence at Google Cloud |
|
|
Sanyam Bhutani
2026-01-21 · 19:00
Sanyam Bhutani
– Partner Engineer, Generative AI Engineer
@ Meta
Speaker: Sanyam Bhutani, Partner Engineer, Generative AI Engineer at Meta |
|
|
Cameron Royce Turner
2026-01-21 · 19:00
Cameron Royce Turner
– Founder and CEO
@ TRUIFY.AI
Speaker: Cameron Royce Turner, Founder and CEO at TRUIFY.AI |
|
|
Claire Longo
2026-01-21 · 19:00
Claire Longo
– AI Researcher
@ Comet
Speaker: Claire Longo, AI Researcher at Comet |
|
|
Sara Zanzottera
2026-01-21 · 19:00
Sara Zanzottera
– Senior Developer
@ BNP Paribas
Speaker: Sara Zanzottera, Senior Developer at BNP Paribas |
|
|
Harpreet Sahota
2026-01-21 · 19:00
Harpreet Sahota
– data science leader
@ Voxel51
Speaker: Harpreet Sahota, Hacker-in-Residence at Voxel51 |
|
|
Holt Skinner
2026-01-21 · 19:00
Holt Skinner
– Developer Advocate
@ Google Cloud AI
Speaker: Holt Skinner, Developer Advocate at Google Cloud AI |
|
|
Interactive Workshops
2026-01-21 · 19:00
Michael Albada
– Principal Applied Scientist
@ Microsoft
,
Manoj Saxena
– Founder & CEO
@ Trustwise
,
Holt Skinner
– Developer Advocate
@ Google Cloud AI
,
Andrea Kropp
– Applied AI Engineer
@ LandingAI
,
Kyle Stratis
– Founder
@ Stratis Data Labs
,
Harpreet Sahota
– data science leader
@ Voxel51
,
Dr. Ali Arsanjani
– Director of Applied AI Engineering; Head of AI Center of Excellence
@ Google Cloud
,
Sanyam Bhutani
– Partner Engineer, Generative AI Engineer
@ Meta
,
Ivan Lee
– CEO
@ Datasaur
,
Claire Longo
– AI Researcher
@ Comet
,
Cameron Royce Turner
– Founder and CEO
@ TRUIFY.AI
,
Sara Zanzottera
– Senior Developer
@ BNP Paribas
,
Sinan Ozdemir
– AI & LLM Expert; Author; Founder & CTO
@ LoopGenius
,
Zain Hasan, PhD
– Staff AI/ML Engineer - DevRel
@ Together AI
Live, expert-led sessions where you’ll build real agentic systems step-by-step. |
Agentic AI Summit | Virtual
|
|
Manoj Saxena
2026-01-21 · 19:00
Manoj Saxena
– Founder & CEO
@ Trustwise
Speaker: Manoj Saxena, Founder & CEO at Trustwise |
|
|
Zain Hasan, PhD
2026-01-21 · 19:00
Zain Hasan, PhD
– Staff AI/ML Engineer - DevRel
@ Together AI
Speaker: Zain Hasan, PhD, Staff AI/ML Engineer - DevRel at Together AI |
|
|
Sinan Ozdemir
2026-01-21 · 19:00
Sinan Ozdemir
– AI & LLM Expert; Author; Founder & CTO
@ LoopGenius
Speaker: Sinan Ozdemir, AI Author and Educator |
|
|
Andrea Kropp
2026-01-21 · 19:00
Andrea Kropp
– Applied AI Engineer
@ LandingAI
Speaker: Andrea Kropp, Applied AI Engineer at LandingAI |
|
|
Michael Albada
2026-01-21 · 19:00
Michael Albada
– Principal Applied Scientist
@ Microsoft
Speaker: Michael Albada, Principal Applied Scientist at Microsoft |
|
|
Nov 6 - Visual Document AI: Because a Pixel is Worth a Thousand Tokens
2025-11-06 · 17:00
Join us for a virtual event to hear talks from experts on the latest developments in Visual Document AI. Date and Location Nov 6, 2025 9-11 AM Pacific Online. Register for the Zoom! Document AI: A Review of the Latest Models, Tasks and Tools In this talk, go through everything document AI: trends, models, tasks, tools! By the end of this talk you will be able to get to building apps based on document models About the Speaker Merve Noyan works on multimodal AI and computer vision at Hugging Face, and she's the author of the book Vision Language Models on O'Reilly. Run Document VLMs in Voxel51 with the VLM Run Plugin — PDF to JSON in Seconds The new VLM Run Plugin for Voxel51 enables seamless execution of document vision-language models directly within the Voxel51 environment. This integration transforms complex document workflows — from PDFs and scanned forms to reports — into structured JSON outputs in seconds. By treating documents as images, our approach remains general, scalable, and compatible with any visual model architecture. The plugin connects visual data curation with model inference, empowering teams to run, visualize, and evaluate document understanding models effortlessly. Document AI is now faster, reproducible, and natively integrated into your Voxel51 workflows. About the Speaker Dinesh Reddy is a founding team member of VLM Run, where he is helping nurture the platform from a sapling into a robust ecosystem for running and evaluating vision-language models across modalities. Previously, he was a scientist at Amazon AWS AI, working on large-scale machine learning systems for intelligent document understanding and visual AI. He completed his Ph.D. at the Robotics Institute, Carnegie Mellon University, focusing on combining learning-based methods with 3D computer vision for in-the-wild data. His research has been recognized with the Best Paper Award at IEEE IVS 2021 and fellowships from Amazon Go and Qualcomm. CommonForms: Automatically Making PDFs Fillable Converting static PDFs into fillable forms remains a surprisingly difficult task, even with the best commercial tools available today. We show that with careful dataset curation and model tuning, it is possible to train high-quality form field detectors for under $500. As part of this effort, we introduce CommonForms, a large-scale dataset of nearly half a million curated form images. We also release a family of highly accurate form field detectors, FFDNet-S and FFDNet-L. About the Speaker Joe Barrow is a researcher at Pattern Data, specializing in document AI and information extraction. He previously worked at the Adobe Document Intelligence Lab after receiving his PhD from the University of Maryland in 2022. Visual Document Retrieval: How to Cluster, Search and Uncover Biases in Document Image Datasets Using Embeddings In this talk you'll learn about the task of visual document retrieval, the models which are widely used by the community, and see them in action through the open source FiftyOne App where you'll learn how to use these models to identify groups and clusters of documents, find unique documents, uncover biases in your visual document dataset, and search over your document corpus using natural language. About the Speaker Harpreet Sahota is a hacker-in-residence and machine learning engineer with a passion for deep learning and generative AI. He’s got a deep interest in VLMs, Visual Agents, Document AI, and Physical AI. |
Nov 6 - Visual Document AI: Because a Pixel is Worth a Thousand Tokens
|
|
Nov 6 - Visual Document AI: Because a Pixel is Worth a Thousand Tokens
2025-11-06 · 17:00
Join us for a virtual event to hear talks from experts on the latest developments in Visual Document AI. Date and Location Nov 6, 2025 9-11 AM Pacific Online. Register for the Zoom! Document AI: A Review of the Latest Models, Tasks and Tools In this talk, go through everything document AI: trends, models, tasks, tools! By the end of this talk you will be able to get to building apps based on document models About the Speaker Merve Noyan works on multimodal AI and computer vision at Hugging Face, and she's the author of the book Vision Language Models on O'Reilly. Run Document VLMs in Voxel51 with the VLM Run Plugin — PDF to JSON in Seconds The new VLM Run Plugin for Voxel51 enables seamless execution of document vision-language models directly within the Voxel51 environment. This integration transforms complex document workflows — from PDFs and scanned forms to reports — into structured JSON outputs in seconds. By treating documents as images, our approach remains general, scalable, and compatible with any visual model architecture. The plugin connects visual data curation with model inference, empowering teams to run, visualize, and evaluate document understanding models effortlessly. Document AI is now faster, reproducible, and natively integrated into your Voxel51 workflows. About the Speaker Dinesh Reddy is a founding team member of VLM Run, where he is helping nurture the platform from a sapling into a robust ecosystem for running and evaluating vision-language models across modalities. Previously, he was a scientist at Amazon AWS AI, working on large-scale machine learning systems for intelligent document understanding and visual AI. He completed his Ph.D. at the Robotics Institute, Carnegie Mellon University, focusing on combining learning-based methods with 3D computer vision for in-the-wild data. His research has been recognized with the Best Paper Award at IEEE IVS 2021 and fellowships from Amazon Go and Qualcomm. CommonForms: Automatically Making PDFs Fillable Converting static PDFs into fillable forms remains a surprisingly difficult task, even with the best commercial tools available today. We show that with careful dataset curation and model tuning, it is possible to train high-quality form field detectors for under $500. As part of this effort, we introduce CommonForms, a large-scale dataset of nearly half a million curated form images. We also release a family of highly accurate form field detectors, FFDNet-S and FFDNet-L. About the Speaker Joe Barrow is a researcher at Pattern Data, specializing in document AI and information extraction. He previously worked at the Adobe Document Intelligence Lab after receiving his PhD from the University of Maryland in 2022. Visual Document Retrieval: How to Cluster, Search and Uncover Biases in Document Image Datasets Using Embeddings In this talk you'll learn about the task of visual document retrieval, the models which are widely used by the community, and see them in action through the open source FiftyOne App where you'll learn how to use these models to identify groups and clusters of documents, find unique documents, uncover biases in your visual document dataset, and search over your document corpus using natural language. About the Speaker Harpreet Sahota is a hacker-in-residence and machine learning engineer with a passion for deep learning and generative AI. He’s got a deep interest in VLMs, Visual Agents, Document AI, and Physical AI. |
Nov 6 - Visual Document AI: Because a Pixel is Worth a Thousand Tokens
|
|
Nov 6 - Visual Document AI: Because a Pixel is Worth a Thousand Tokens
2025-11-06 · 17:00
Join us for a virtual event to hear talks from experts on the latest developments in Visual Document AI. Date and Location Nov 6, 2025 9-11 AM Pacific Online. Register for the Zoom! Document AI: A Review of the Latest Models, Tasks and Tools In this talk, go through everything document AI: trends, models, tasks, tools! By the end of this talk you will be able to get to building apps based on document models About the Speaker Merve Noyan works on multimodal AI and computer vision at Hugging Face, and she's the author of the book Vision Language Models on O'Reilly. Run Document VLMs in Voxel51 with the VLM Run Plugin — PDF to JSON in Seconds The new VLM Run Plugin for Voxel51 enables seamless execution of document vision-language models directly within the Voxel51 environment. This integration transforms complex document workflows — from PDFs and scanned forms to reports — into structured JSON outputs in seconds. By treating documents as images, our approach remains general, scalable, and compatible with any visual model architecture. The plugin connects visual data curation with model inference, empowering teams to run, visualize, and evaluate document understanding models effortlessly. Document AI is now faster, reproducible, and natively integrated into your Voxel51 workflows. About the Speaker Dinesh Reddy is a founding team member of VLM Run, where he is helping nurture the platform from a sapling into a robust ecosystem for running and evaluating vision-language models across modalities. Previously, he was a scientist at Amazon AWS AI, working on large-scale machine learning systems for intelligent document understanding and visual AI. He completed his Ph.D. at the Robotics Institute, Carnegie Mellon University, focusing on combining learning-based methods with 3D computer vision for in-the-wild data. His research has been recognized with the Best Paper Award at IEEE IVS 2021 and fellowships from Amazon Go and Qualcomm. CommonForms: Automatically Making PDFs Fillable Converting static PDFs into fillable forms remains a surprisingly difficult task, even with the best commercial tools available today. We show that with careful dataset curation and model tuning, it is possible to train high-quality form field detectors for under $500. As part of this effort, we introduce CommonForms, a large-scale dataset of nearly half a million curated form images. We also release a family of highly accurate form field detectors, FFDNet-S and FFDNet-L. About the Speaker Joe Barrow is a researcher at Pattern Data, specializing in document AI and information extraction. He previously worked at the Adobe Document Intelligence Lab after receiving his PhD from the University of Maryland in 2022. Visual Document Retrieval: How to Cluster, Search and Uncover Biases in Document Image Datasets Using Embeddings In this talk you'll learn about the task of visual document retrieval, the models which are widely used by the community, and see them in action through the open source FiftyOne App where you'll learn how to use these models to identify groups and clusters of documents, find unique documents, uncover biases in your visual document dataset, and search over your document corpus using natural language. About the Speaker Harpreet Sahota is a hacker-in-residence and machine learning engineer with a passion for deep learning and generative AI. He’s got a deep interest in VLMs, Visual Agents, Document AI, and Physical AI. |
Nov 6 - Visual Document AI: Because a Pixel is Worth a Thousand Tokens
|
|
Nov 6 - Visual Document AI: Because a Pixel is Worth a Thousand Tokens
2025-11-06 · 17:00
Join us for a virtual event to hear talks from experts on the latest developments in Visual Document AI. Date and Location Nov 6, 2025 9-11 AM Pacific Online. Register for the Zoom! Document AI: A Review of the Latest Models, Tasks and Tools In this talk, go through everything document AI: trends, models, tasks, tools! By the end of this talk you will be able to get to building apps based on document models About the Speaker Merve Noyan works on multimodal AI and computer vision at Hugging Face, and she's the author of the book Vision Language Models on O'Reilly. Run Document VLMs in Voxel51 with the VLM Run Plugin — PDF to JSON in Seconds The new VLM Run Plugin for Voxel51 enables seamless execution of document vision-language models directly within the Voxel51 environment. This integration transforms complex document workflows — from PDFs and scanned forms to reports — into structured JSON outputs in seconds. By treating documents as images, our approach remains general, scalable, and compatible with any visual model architecture. The plugin connects visual data curation with model inference, empowering teams to run, visualize, and evaluate document understanding models effortlessly. Document AI is now faster, reproducible, and natively integrated into your Voxel51 workflows. About the Speaker Dinesh Reddy is a founding team member of VLM Run, where he is helping nurture the platform from a sapling into a robust ecosystem for running and evaluating vision-language models across modalities. Previously, he was a scientist at Amazon AWS AI, working on large-scale machine learning systems for intelligent document understanding and visual AI. He completed his Ph.D. at the Robotics Institute, Carnegie Mellon University, focusing on combining learning-based methods with 3D computer vision for in-the-wild data. His research has been recognized with the Best Paper Award at IEEE IVS 2021 and fellowships from Amazon Go and Qualcomm. CommonForms: Automatically Making PDFs Fillable Converting static PDFs into fillable forms remains a surprisingly difficult task, even with the best commercial tools available today. We show that with careful dataset curation and model tuning, it is possible to train high-quality form field detectors for under $500. As part of this effort, we introduce CommonForms, a large-scale dataset of nearly half a million curated form images. We also release a family of highly accurate form field detectors, FFDNet-S and FFDNet-L. About the Speaker Joe Barrow is a researcher at Pattern Data, specializing in document AI and information extraction. He previously worked at the Adobe Document Intelligence Lab after receiving his PhD from the University of Maryland in 2022. Visual Document Retrieval: How to Cluster, Search and Uncover Biases in Document Image Datasets Using Embeddings In this talk you'll learn about the task of visual document retrieval, the models which are widely used by the community, and see them in action through the open source FiftyOne App where you'll learn how to use these models to identify groups and clusters of documents, find unique documents, uncover biases in your visual document dataset, and search over your document corpus using natural language. About the Speaker Harpreet Sahota is a hacker-in-residence and machine learning engineer with a passion for deep learning and generative AI. He’s got a deep interest in VLMs, Visual Agents, Document AI, and Physical AI. |
Nov 6 - Visual Document AI: Because a Pixel is Worth a Thousand Tokens
|
|
Nov 6 - Visual Document AI: Because a Pixel is Worth a Thousand Tokens
2025-11-06 · 17:00
Join us for a virtual event to hear talks from experts on the latest developments in Visual Document AI. Date and Location Nov 6, 2025 9-11 AM Pacific Online. Register for the Zoom! Document AI: A Review of the Latest Models, Tasks and Tools In this talk, go through everything document AI: trends, models, tasks, tools! By the end of this talk you will be able to get to building apps based on document models About the Speaker Merve Noyan works on multimodal AI and computer vision at Hugging Face, and she's the author of the book Vision Language Models on O'Reilly. Run Document VLMs in Voxel51 with the VLM Run Plugin — PDF to JSON in Seconds The new VLM Run Plugin for Voxel51 enables seamless execution of document vision-language models directly within the Voxel51 environment. This integration transforms complex document workflows — from PDFs and scanned forms to reports — into structured JSON outputs in seconds. By treating documents as images, our approach remains general, scalable, and compatible with any visual model architecture. The plugin connects visual data curation with model inference, empowering teams to run, visualize, and evaluate document understanding models effortlessly. Document AI is now faster, reproducible, and natively integrated into your Voxel51 workflows. About the Speaker Dinesh Reddy is a founding team member of VLM Run, where he is helping nurture the platform from a sapling into a robust ecosystem for running and evaluating vision-language models across modalities. Previously, he was a scientist at Amazon AWS AI, working on large-scale machine learning systems for intelligent document understanding and visual AI. He completed his Ph.D. at the Robotics Institute, Carnegie Mellon University, focusing on combining learning-based methods with 3D computer vision for in-the-wild data. His research has been recognized with the Best Paper Award at IEEE IVS 2021 and fellowships from Amazon Go and Qualcomm. CommonForms: Automatically Making PDFs Fillable Converting static PDFs into fillable forms remains a surprisingly difficult task, even with the best commercial tools available today. We show that with careful dataset curation and model tuning, it is possible to train high-quality form field detectors for under $500. As part of this effort, we introduce CommonForms, a large-scale dataset of nearly half a million curated form images. We also release a family of highly accurate form field detectors, FFDNet-S and FFDNet-L. About the Speaker Joe Barrow is a researcher at Pattern Data, specializing in document AI and information extraction. He previously worked at the Adobe Document Intelligence Lab after receiving his PhD from the University of Maryland in 2022. Visual Document Retrieval: How to Cluster, Search and Uncover Biases in Document Image Datasets Using Embeddings In this talk you'll learn about the task of visual document retrieval, the models which are widely used by the community, and see them in action through the open source FiftyOne App where you'll learn how to use these models to identify groups and clusters of documents, find unique documents, uncover biases in your visual document dataset, and search over your document corpus using natural language. About the Speaker Harpreet Sahota is a hacker-in-residence and machine learning engineer with a passion for deep learning and generative AI. He’s got a deep interest in VLMs, Visual Agents, Document AI, and Physical AI. |
Nov 6 - Visual Document AI: Because a Pixel is Worth a Thousand Tokens
|