talk-data.com
People (124 results)
See all 124 →Companies (11 results)
See all 11 →Activities & events
| Title & Speakers | Event |
|---|---|
|
Virtual Open Coffee Club
2026-03-15 · 18:00
Hosted by Santa Monica New Tech / Los Angeles NEW: Click here for your membership and a downloadable resource pack for entrepreneurs! This is a virtual meeting on Zoom. Watch for the meeting ID and passcode after you RSVP. VIDEO CONFERENCE DETAILS: London Time: 18:00-19:00 Introductions on Zoom. 19:00-20:00 Informal Networking using Zoom breakout rooms or other channels. You can download the Zoom client ahead of time at https://zoom.us/download#client_4meeting. I recommend you familiarize yourself with the Zoom app if you've never used it before. Here are some helpful tutorials: https://bit.ly/3alN4NV. A few things to prepare: • You should check in about 10 minutes early to make sure you can connect and your audio and video are working right. • To handle introductions, you will be called on when it's your turn to introduce yourself. A microphone will be required for us to hear you. If you choose computer audio please make sure yours is set up and working with Zoom — there is a test function within the app. • Having your own camera and having it on is required. • Admissions close 10 minutes past the hour to avoid disrupting the meeting in progress. • By default, you will be muted at the start. You can unmute yourself when it’s your turn to speak. Please mute yourself again afterward to avoid disruptions to the meeting. • Once we are done with the introductions we'll have some themed breakout rooms (tech, sales/marketing, legal, finance, other) for networking. • IMPORTANT: If you'd like to connect with other participants after the session by using the messaging feature in Meetup make sure to change your Privacy Settings to allow being contacted by other members. TRADITIONAL COFFEE SHOP MEETING DETAILS: 10AM-11AM Introductions. 11AM-12PM Informal Networking. Watch a short video here: https://youtu.be/f42n5a92zy0 The OpenCoffee Club is a concept that originated in London with the goal to encourage entrepreneurs, developers, and investors to organize real-world informal meetups to chat, network, and grow. There are now hundreds (thousands?) of such groups around the world. Come join the movement! Learn more about the OCC concept here: https://bit.ly/2MuyURs |
Virtual Open Coffee Club
|
|
Berliner IT-Stammtisch für Freelancer vom DBITS e.V.
2026-03-13 · 18:00
Hallo, es ist wieder so weit, der zweite Freitag im September ist schon nächste Woche! Deshalb habe ich im Neffes draußen im Innenhof für uns reserviert. Wenn das Wetter schlecht werden sollte, haben wir die Ecke vor dem - ausgeschalteten - Fernseher für uns :-). Ich freue mich schon, euch wieder persönlich zu sehen! Und hier der übliche Ankündigungstext: Unser zweimonatliches Treffen steht wieder an! An jedem zweiten Freitag eines ungeraden Monats treffen wir Selbständige, Unternehmer und Freiberufler aus der IT uns wieder um 19:00 Uhr im Neffes. Bei einem gemütlichen Zusammensein werden wir uns über fast alle Themen unterhalten. Der Schwerpunkt liegt dabei auf den beruflichen Themen. Z. B. wie gehen andere mit den Kunden und deren Mitarbeitern um, welches sind die neuesten Insider Infos von den verschiedene Unternehmen, wie sieht es bei den anderen mit den Einsatzorten und Arbeitsbedingungen aus und vieles mehr. Es wird wie immer ein interessanter Abend mit einem umfangreichen Erfahrungsaustausch werden, der einen auch beruflich weiterbringt. Mit dem DBITS e.V. wollen wir die Anerkennung von IT-Freiberuflern in der Gesellschaft und Politik vorantreiben, dort gibt es leider noch einen sehr großen Aufhol- und Aufklärungsbedarf und damit viel zu besprechen und zu erfahren. Des Weiteren bietet uns der Verein ein breites Netzwerk an Leuten mit den gleichen Zielen und Problemen. Aber wie gesagt, das Ziel ist ein schöner, informativer Abend, es gibt kein Mitgliederzwang. Nimm teil, wir freuen uns auf dein Kommen! Viele Grüße und ich freue mich schon auf ein Treffen am nächsten Freitag! |
Berliner IT-Stammtisch für Freelancer vom DBITS e.V.
|
|
Présentation du SNDS et ses modalités d'accès
2026-03-12 · 13:00
Le HDH vous convie à une session de formation sur les données de santé du SNDS (Système National des Données de Santé) et ses modalités d'accès. L’objectif de cette formation est de fournir un premier niveau d’information sur les données de la base principale du SNDS, les possibilités qu’elles offrent et leurs modalités d’accès. Elle s’adresse à tout public et notamment des potentiels porteurs de projets visant à utiliser ces données. La formation aura lieu en visioconférence via l’outil Gotomeeting. Seules les personnes inscrites pourront accéder au lien de visioconférence. Vous pouvez également consulter ces sites pour une première approche, poser vos questions sur le SNDS, ou débuter votre demande d'accès :
Participez à la réunion depuis votre ordinateur, tablette ou smartphone : https://meet.goto.com/813848037 |
Présentation du SNDS et ses modalités d'accès
|
|
OpenWoo.app Community Meeting
2026-03-11 · 10:00
Met meerdere overheden en leveranciers werken we samen aan https://openwoo.app/, een oplossing voor overheden om hun reeds bestaande informatie automatisch te ontsluiten naar KOOP en een eigen publicatieplatform. Uniek aan deze oplossing is dat we gebruikmaken van het federaal datastelsel en een federale zoekvraag hebben ontwikkeld. Hierdoor is een landelijke index niet nodig en kunnen overheden ook in elkaars openbare informatie zoeken. Elke tweede woensdag van de maand komen we samen met overheden en leveranciers bij elkaar om nieuwe functionaliteiten te bespreken en de deze maand gerealiseerde functionaliteiten door te nemen. Interesse in openwoo.app? Meld je dan ook vooral aan voor het LinkedIn-event op https://www.linkedin.com/events/openwoo-appcommunitymeetingmaart7164180127860383744/comments/ |
OpenWoo.app Community Meeting
|
|
Virtual Open Coffee Club
2026-03-08 · 18:00
Hosted by Santa Monica New Tech / Los Angeles NEW: Click here for your membership and a downloadable resource pack for entrepreneurs! This is a virtual meeting on Zoom. Watch for the meeting ID and passcode after you RSVP. VIDEO CONFERENCE DETAILS: London Time: 18:00-19:00 Introductions on Zoom. 19:00-20:00 Informal Networking using Zoom breakout rooms or other channels. You can download the Zoom client ahead of time at https://zoom.us/download#client_4meeting. I recommend you familiarize yourself with the Zoom app if you've never used it before. Here are some helpful tutorials: https://bit.ly/3alN4NV. A few things to prepare: • You should check in about 10 minutes early to make sure you can connect and your audio and video are working right. • To handle introductions, you will be called on when it's your turn to introduce yourself. A microphone will be required for us to hear you. If you choose computer audio please make sure yours is set up and working with Zoom — there is a test function within the app. • Having your own camera and having it on is required. • Admissions close 10 minutes past the hour to avoid disrupting the meeting in progress. • By default, you will be muted at the start. You can unmute yourself when it’s your turn to speak. Please mute yourself again afterward to avoid disruptions to the meeting. • Once we are done with the introductions we'll have some themed breakout rooms (tech, sales/marketing, legal, finance, other) for networking. • IMPORTANT: If you'd like to connect with other participants after the session by using the messaging feature in Meetup make sure to change your Privacy Settings to allow being contacted by other members. TRADITIONAL COFFEE SHOP MEETING DETAILS: 10AM-11AM Introductions. 11AM-12PM Informal Networking. Watch a short video here: https://youtu.be/f42n5a92zy0 The OpenCoffee Club is a concept that originated in London with the goal to encourage entrepreneurs, developers, and investors to organize real-world informal meetups to chat, network, and grow. There are now hundreds (thousands?) of such groups around the world. Come join the movement! Learn more about the OCC concept here: https://bit.ly/2MuyURs |
Virtual Open Coffee Club
|
|
AI Builders London Old St. - Mar 5
2026-03-05 · 17:30
🎟️ Get tickets: https://lu.ma/ai-builders 🎟️ ☝️This is a free meetup however a Luma ticket is required! Join our meetup for AI engineers & founders! We share the latest insights about: AI dev tools, Agent frameworks, RAG pipelines, automation hacks, Cursor/Claude code hacks, and new gen-AI models. **:: FOR WHO ::** ✅ Anyone actively building with Generative AI ✅ Devs, Product peeps, Data lovers, ML engineers, Founders ⚠️Technical LLM knowledge required!* :: FORMAT ::
**:: AGENDA ::** 17:30 🤝 Walk-in 18:00 🍕 Pizza (be early!) 18:30 🎤 Pioneer Speaker 19:00 ---- 💬 BREAK TIME ----- 19:30 💻 Demo (TBA) 19.45 💻 Demo (TBA) 20.00 🍻 Drinks 21.00 End :: FAQ :: • What is AI Builders? A self-organizing nonprofit community of 3000+ AI nerds🤓. and yes.. we're building a democratic AI CEO and run on opencollective.com donations. • Can I demo, give a talk, or just help out? Message Arthur (+31636570260) in case you want to shine✨ on stage and grow your network! • *I'm not technical. Can I come? Yes! To fully enjoy the meetup, we recommend chatting with AI to understand these basic LLM concepts: Multimodal, Vector Embeddings, RAG, Chaining, Structured JSON Output, Function Calling, API calls, Knowledge Graphs, Reinforcement Learning, Fine-tuning, AI Agents. • Why hangout at AI Builders?
Location
We look forward to hang out with you:
207 Old St, London EC1V 9NR, United Kingdom
It's close to the Old Street Tube Station
Thanks to our friends at Beyond
|
AI Builders London Old St. - Mar 5
|
|
AI Builders Lisbon - 4 March
2026-03-04 · 16:00
🎟️ Get Tickets on Lu.ma 🎟️ ☝️This is a donation based meet-up (€5 - €10), Luma ticket is required! Join our technical meetup for the biggest AI nerds in Lisbon! We share the latest AI devtools, APIs, RAG & Agent frameworks, techniques and models. **:: FOR WHO ::** ✅ Anyone actively building with Generative AI ✅ Devs, Product peeps, Data lovers, ML engineers, Founders ⚠️ Basic technical LLM / software engineering knowledge is required!* :: FORMAT ::
**:: AGENDA ::** 17:30 🤝 Walk-in 18:00 🍕 Pizza (be early!) 18:30 🎤 Pioneer Talk 19:00 ---- 💬 BREAK ------ 19:30 💻 Demo 19.45 💻 Demo 20.00 🍻 Drinks 21.00 End ::: FAQ :: • What is AI Builders? A self-organizing nonprofit community of 3000+ AI nerds🤓. and yes.. we're building a democratic AI CEO and run on opencollective.com donations. • Can I demo, give a talk, or just help out? Message Arthur (+31636570260) in case you want to shine✨ on stage and grow your network! • I'm not technical. Can I come? Yes! To fully enjoy the meetup, we recommend chatting with AI to understand these basic LLM concepts: Multimodal, Vector Embeddings, RAG, Chaining, Structured JSON Output, Function Calling, API calls, Knowledge Graphs, Reinforcement Learning, Fine-tuning, AI Agents. • Why hangout at AI Builders?
Why donate? We are a non-profit that relies heavily on donations, pay it forward if you enjoy the events. We are in the process of making our finances totally transparent. If you want to become a recurring supporter for perks, go to : Open Collective. 🎟️ Get Tickets on Lu.ma 🎟️ ☝️This is a donation based meet-up (€5 - €10), Luma ticket is required! |
AI Builders Lisbon - 4 March
|
|
Virtual Open Coffee Club
2026-03-01 · 18:00
Hosted by Santa Monica New Tech / Los Angeles NEW: Click here for your membership and a downloadable resource pack for entrepreneurs! This is a virtual meeting on Zoom. Watch for the meeting ID and passcode after you RSVP. VIDEO CONFERENCE DETAILS: London Time: 18:00-19:00 Introductions on Zoom. 19:00-20:00 Informal Networking using Zoom breakout rooms or other channels. You can download the Zoom client ahead of time at https://zoom.us/download#client_4meeting. I recommend you familiarize yourself with the Zoom app if you've never used it before. Here are some helpful tutorials: https://bit.ly/3alN4NV. A few things to prepare: • You should check in about 10 minutes early to make sure you can connect and your audio and video are working right. • To handle introductions, you will be called on when it's your turn to introduce yourself. A microphone will be required for us to hear you. If you choose computer audio please make sure yours is set up and working with Zoom — there is a test function within the app. • Having your own camera and having it on is required. • Admissions close 10 minutes past the hour to avoid disrupting the meeting in progress. • By default, you will be muted at the start. You can unmute yourself when it’s your turn to speak. Please mute yourself again afterward to avoid disruptions to the meeting. • Once we are done with the introductions we'll have some themed breakout rooms (tech, sales/marketing, legal, finance, other) for networking. • IMPORTANT: If you'd like to connect with other participants after the session by using the messaging feature in Meetup make sure to change your Privacy Settings to allow being contacted by other members. TRADITIONAL COFFEE SHOP MEETING DETAILS: 10AM-11AM Introductions. 11AM-12PM Informal Networking. Watch a short video here: https://youtu.be/f42n5a92zy0 The OpenCoffee Club is a concept that originated in London with the goal to encourage entrepreneurs, developers, and investors to organize real-world informal meetups to chat, network, and grow. There are now hundreds (thousands?) of such groups around the world. Come join the movement! Learn more about the OCC concept here: https://bit.ly/2MuyURs |
Virtual Open Coffee Club
|
|
Untitled talk
2026-02-26 · 19:00
|
BioinformaticsLondon - to be announced
|
|
AI Builders Amsterdam :: Pizza, Demos & Networking (paid event)
2026-02-26 · 16:30
🎟️ Get tickets: https://lu.ma/ai-builders 🎟️ ☝️This is a paid meetup (€20 - €10), Luma ticket is required! Join our Monthly AI meetup Practical Demos & Technical Talks about building with LLMs and any Gen-AI model. :: FOR WHO :: ✅ Anyone actively building with Generative AI ✅ Devs, Product peeps, Data lovers, ML engineers, Founders ⚠️Technical LLM knowledge required!* :: FORMAT :: 💻 ⚡️ Speed Demos (10 min): Builders sharing real-world AI solutions including their breakthrough code, diagrams and prompts ! 🎤 🦄 Pioneer Talks (20 min): Inspiring talk or demo from emerging Gen-AI leaders in Europe or Silicon Valley 🤝🍕🍻 Fun Vibes: Lots of time to connect with other builders over some yummy pizza & drinks. :: AGENDA :: 17:30 🤝 Drinks & Networking 18:00 🍕 Pizza (be early!) 18:30 🎤 🦄 Pioneer Talk (20m) --- Break --- 19:30 💻 ⚡️ Demos (4 x 10m) 20.10 🍻 Drinks & Networking 21.00 End :: FAQ :: • What's AI Builders? We're a self-organizing nonprofit community of 3000+ AI Nerds in Europe. Yes.. we're building our own AI CEO. • Why do I need to pay? 1). So we know how many people will come (max capacity of the space and reduce food waste) 2). Sponsor money doesn't cover all of our costs yet. • Can I get a free ticket? Can I volunteer as co-host? Message Cristian (+31636420602) if we still need co-hosts or request a free ticket. Co-hosts arrive 1.5h early and help set-up the event or welcome people. • *I'm not technical. Can I come? Yes, but to enjoy the meetup, we recommend to learn about these LLM Concepts: Multimodal, Vector Embeddings, RAG, Chaining, Structured Output, Function Calling, API calls, Knowledge Graphs, Reinforcement Learning, Fine-tuning, Agents. Additionally: Computer Vision, Diffusion Models, DevOps, MLOps. • Why go to AI Meetups?
🎟️ Get tickets: https://lu.ma/ai-builders 🎟️ ☝️This is a paid meetup (€20 - €10), Luma ticket is required! |
AI Builders Amsterdam :: Pizza, Demos & Networking (paid event)
|
|
Vibe Coding the Enterprise: From Flow State to Focused Delivery
2026-02-24 · 18:30
Provisional date -The official event date will be announced soon! 📌Abstract Vibe coding captures something real: momentum. The ability to move from an idea to a working prototype at surprising speed, guided by intent, context, and AI assistance. But in an enterprise environment, speed alone isn’t enough — software must also be secure, compliant, observable, and built to last. We'll see how teams can build applications in a single, end-to-end development flow — from business requirements and early prototypes to engineering, testing, delivery, and production — without breaking context or introducing late-stage rework. At the center of this flow is the software catalog: a living system of record that connects services, APIs, data, ownership, dependencies, and standards. More than documentation, the catalog becomes the shared interface for developers, platform teams, and AI agents to understand the system and act consistently, within enterprise rules and guardrails. The goal isn’t to limit creativity, but to make flow sustainable — turning AI-driven momentum into software that’s ready for production. 🔍 Key topics • From vibe-coding to ai-assisted delivery in production • Software catalog as the system of record for enterprise context • AI agents operating in context continuity within guardrails ⏲️ Agenda 18.30* Welcoming 18.45 Talk 19:30 Q&A 19:45 Closing remarks & Networking and see you at the next Meetup! *You will receive detailed information on how to access the building as soon as available. 🎙️Speaker Giulio Roggero CTO @ Mia-Platform 25 years of experience in software engineering, serial entrepreneur with more than 10 business initiatives launched, today is co-founder and CTO at Mia-Platform, the Internal Developer Platform named by Gartner Cool Vendor for Software Engineering Technologies and in the Cloud Application Platforms Magic Quadrant. Principal focus: cloud native, platform engineering, data fabric and omnichannel experience. He likes to paint Blood Bowl miniatures, construct Lego, build and drive RC Cars and learn piano. |
Vibe Coding the Enterprise: From Flow State to Focused Delivery
|
|
Virtual Open Coffee Club
2026-02-22 · 18:00
Hosted by Santa Monica New Tech / Los Angeles NEW: Click here for your membership and a downloadable resource pack for entrepreneurs! This is a virtual meeting on Zoom. Watch for the meeting ID and passcode after you RSVP. VIDEO CONFERENCE DETAILS: London Time: 18:00-19:00 Introductions on Zoom. 19:00-20:00 Informal Networking using Zoom breakout rooms or other channels. You can download the Zoom client ahead of time at https://zoom.us/download#client_4meeting. I recommend you familiarize yourself with the Zoom app if you've never used it before. Here are some helpful tutorials: https://bit.ly/3alN4NV. A few things to prepare: • You should check in about 10 minutes early to make sure you can connect and your audio and video are working right. • To handle introductions, you will be called on when it's your turn to introduce yourself. A microphone will be required for us to hear you. If you choose computer audio please make sure yours is set up and working with Zoom — there is a test function within the app. • Having your own camera and having it on is required. • Admissions close 10 minutes past the hour to avoid disrupting the meeting in progress. • By default, you will be muted at the start. You can unmute yourself when it’s your turn to speak. Please mute yourself again afterward to avoid disruptions to the meeting. • Once we are done with the introductions we'll have some themed breakout rooms (tech, sales/marketing, legal, finance, other) for networking. • IMPORTANT: If you'd like to connect with other participants after the session by using the messaging feature in Meetup make sure to change your Privacy Settings to allow being contacted by other members. TRADITIONAL COFFEE SHOP MEETING DETAILS: 10AM-11AM Introductions. 11AM-12PM Informal Networking. Watch a short video here: https://youtu.be/f42n5a92zy0 The OpenCoffee Club is a concept that originated in London with the goal to encourage entrepreneurs, developers, and investors to organize real-world informal meetups to chat, network, and grow. There are now hundreds (thousands?) of such groups around the world. Come join the movement! Learn more about the OCC concept here: https://bit.ly/2MuyURs |
Virtual Open Coffee Club
|
|
Virtual Open Coffee Club
2026-02-15 · 18:00
Hosted by Santa Monica New Tech / Los Angeles NEW: Click here for your membership and a downloadable resource pack for entrepreneurs! This is a virtual meeting on Zoom. Watch for the meeting ID and passcode after you RSVP. VIDEO CONFERENCE DETAILS: London Time: 18:00-19:00 Introductions on Zoom. 19:00-20:00 Informal Networking using Zoom breakout rooms or other channels. You can download the Zoom client ahead of time at https://zoom.us/download#client_4meeting. I recommend you familiarize yourself with the Zoom app if you've never used it before. Here are some helpful tutorials: https://bit.ly/3alN4NV. A few things to prepare: • You should check in about 10 minutes early to make sure you can connect and your audio and video are working right. • To handle introductions, you will be called on when it's your turn to introduce yourself. A microphone will be required for us to hear you. If you choose computer audio please make sure yours is set up and working with Zoom — there is a test function within the app. • Having your own camera and having it on is required. • Admissions close 10 minutes past the hour to avoid disrupting the meeting in progress. • By default, you will be muted at the start. You can unmute yourself when it’s your turn to speak. Please mute yourself again afterward to avoid disruptions to the meeting. • Once we are done with the introductions we'll have some themed breakout rooms (tech, sales/marketing, legal, finance, other) for networking. • IMPORTANT: If you'd like to connect with other participants after the session by using the messaging feature in Meetup make sure to change your Privacy Settings to allow being contacted by other members. TRADITIONAL COFFEE SHOP MEETING DETAILS: 10AM-11AM Introductions. 11AM-12PM Informal Networking. Watch a short video here: https://youtu.be/f42n5a92zy0 The OpenCoffee Club is a concept that originated in London with the goal to encourage entrepreneurs, developers, and investors to organize real-world informal meetups to chat, network, and grow. There are now hundreds (thousands?) of such groups around the world. Come join the movement! Learn more about the OCC concept here: https://bit.ly/2MuyURs |
Virtual Open Coffee Club
|
|
Berliner IT-Stammtisch in Zusammenarbeit mit DBITS e.V. und 4Freelance
2026-02-13 · 18:00
Hallo, es ist wieder so weit, der zweite Freitag im Juli ist schon nächste Woche! Deshalb habe ich im Neffes draußen im Innenhof für uns reserviert. Wenn das Wetter schlecht werden sollte, haben wir die Ecke vor dem - ausgeschalteten - Fernseher für uns :-). Ich freue mich schon, euch wieder persönlich zu sehen! Und hier der übliche Ankündigungstext: Unser zweimonatliches Treffen steht wieder an! An jedem zweiten Freitag eines ungeraden Monats treffen wir Selbständige, Unternehmer und Freiberufler aus der IT uns wieder um 19:00 Uhr im Neffes. Bei einem gemütlichen Zusammensein werden wir uns über fast alle Themen unterhalten. Der Schwerpunkt liegt dabei auf den beruflichen Themen. Z. B. wie gehen andere mit den Kunden und deren Mitarbeitern um, welches sind die neuesten Insider Infos von den verschiedene Unternehmen, wie sieht es bei den anderen mit den Einsatzorten und Arbeitsbedingungen aus und vieles mehr. Es wird wie immer ein interessanter Abend mit einem umfangreichen Erfahrungsaustausch werden, der einen auch beruflich weiterbringt. Mit dem DBITS e.V. wollen wir die Anerkennung von IT-Freiberuflern in der Gesellschaft und Politik vorantreiben, dort gibt es leider noch einen sehr großen Aufhol- und Aufklärungsbedarf und damit viel zu besprechen und zu erfahren. Des Weiteren bietet uns der Verein ein breites Netzwerk an Leuten mit den gleichen Zielen und Problemen. Aber wie gesagt, das Ziel ist ein schöner, informativer Abend, es gibt kein Mitgliederzwang. Nimm teil, wir freuen uns auf dein kommen! Viele Grüße und ich freue mich schon auf ein Treffen am nächsten Freitag! |
Berliner IT-Stammtisch in Zusammenarbeit mit DBITS e.V. und 4Freelance
|
|
Feb 11 - Visual AI for Video Use Cases
2026-02-11 · 17:00
Join our virtual Meetup to hear talks from experts on cutting-edge topics at the intersection of Visual AI and video use cases. Time and Location Feb 11, 2026 9 - 11 AM Pacific Online. Register for the Zoom! VIDEOP2R: Video Understanding from Perception to Reasoning Reinforcement fine-tuning (RFT), a two-stage framework consisting of supervised fine-tuning (SFT) and reinforcement learning (RL) has shown promising results on improving reasoning ability of large language models (LLMs). Yet extending RFT to large video language models (LVLMs) remains challenging. We propose VideoP2R, a novel process-aware video RFT framework that enhances video reasoning by modeling perception and reasoning as distinct processes. In the SFT stage, we develop a three-step pipeline to generate VideoP2R-CoT-162K, a high-quality, process-aware chain-of-thought (CoT) dataset for perception and reasoning. In the RL stage, we introduce a novel process-aware group relative policy optimization (PA-GRPO) algorithm that supplies separate rewards for perception and reasoning. Extensive experiments show that VideoP2R achieves state-of-the-art (SotA) performance on six out of seven video reasoning and understanding benchmarks. Ablation studies further confirm the effectiveness of our process-aware modeling and PA-GRPO and demonstrate that model's perception output is information-sufficient for downstream reasoning. About the Speaker Yifan Jiang is a third-year Ph.D. student in the Information Science Institute at the University of Southern California (USC-ISI), advised by Dr. Jay Pujara, focusing on natural language processing, commonsense reasoning and multimodality large language models. Layer-Aware Video Composition via Split-then-Merge Split-then-Merge (StM) is a novel generative framework that overcomes data scarcity in video composition by splitting unlabeled videos into separate foreground and background layers for self-supervised learning. By utilizing a transformation-aware training pipeline with multi-layer fusion, the model learns to realistically compose dynamic subjects into diverse scenes without relying on expensive annotated datasets. This presentation will cover the problem of video composition and the details of StM, an approach looking at this problem from a generative AI perspective. We will conclude by demonstrating how StM is working, and outperforming state-of-the-art methods in both quantitative benchmarks and qualitative evaluations. About the Speaker Ozgur Kara is a 4th year Computer Science PhD student at the University of Illinois Urbana-Champaign (UIUC), advised by Founder Professor James M. Rehg. His research builds the next generation of video AI by tackling three core challenges: efficiency, controllability, and safety. Video Reasoning for Worker Safety Ensuring worker safety in industrial environments requires more than object detection or motion tracking; it demands a genuine understanding of human actions, context, and risk. This talk demonstrates how NVIDIA Cosmos Reason, a multimodal video-reasoning model, interprets workplace scenarios with sophisticated temporal and semantic awareness, identifying nuanced safe and unsafe behaviors that conventional vision systems frequently overlook. By integrating Cosmos Reason with FiftyOne, users achieve both automated safety assessments and transparent, interpretable explanations revealing why specific actions are deemed hazardous. Using a curated worker-safety dataset of authentic factory-floor footage, we show how video reasoning enhances audits, training, and compliance workflows while minimizing dependence on extensive labeled datasets. The resulting system demonstrates the potential of explainable multimodal AI to enable safer, more informed decision-making across manufacturing, logistics, construction, healthcare, and other sectors where understanding human behavior is essential. About the Speaker Paula Ramos has a PhD in Computer Vision and Machine Learning, with more than 20 years of experience in the technological field. She has been developing novel integrated engineering technologies, mainly in Computer Vision, robotics, and Machine Learning applied to agriculture, since the early 2000s in Colombia. Video Intelligence Is Going Agentic Video content has become ubiquitous in our digital world, yet the tools for working with video have remained largely unchanged for decades. This talk explores how the convergence of foundation models and agent architectures is fundamentally transforming video interaction and creation. We'll examine how video-native foundation models, multimodal interfaces, and agent transparency are reshaping enterprise media workflows through a deep dive into Jockey, a pioneering video agent system. About the Speaker James Le currently leads the developer experience function at TwelveLabs - a startup building foundation models for video understanding. He previously operated in the MLOps space and ran a blog/podcast on the Data & AI infrastructure ecosystem. |
Feb 11 - Visual AI for Video Use Cases
|
|
Feb 11 - Visual AI for Video Use Cases
2026-02-11 · 17:00
Join our virtual Meetup to hear talks from experts on cutting-edge topics at the intersection of Visual AI and video use cases. Time and Location Feb 11, 2026 9 - 11 AM Pacific Online. Register for the Zoom! VIDEOP2R: Video Understanding from Perception to Reasoning Reinforcement fine-tuning (RFT), a two-stage framework consisting of supervised fine-tuning (SFT) and reinforcement learning (RL) has shown promising results on improving reasoning ability of large language models (LLMs). Yet extending RFT to large video language models (LVLMs) remains challenging. We propose VideoP2R, a novel process-aware video RFT framework that enhances video reasoning by modeling perception and reasoning as distinct processes. In the SFT stage, we develop a three-step pipeline to generate VideoP2R-CoT-162K, a high-quality, process-aware chain-of-thought (CoT) dataset for perception and reasoning. In the RL stage, we introduce a novel process-aware group relative policy optimization (PA-GRPO) algorithm that supplies separate rewards for perception and reasoning. Extensive experiments show that VideoP2R achieves state-of-the-art (SotA) performance on six out of seven video reasoning and understanding benchmarks. Ablation studies further confirm the effectiveness of our process-aware modeling and PA-GRPO and demonstrate that model's perception output is information-sufficient for downstream reasoning. About the Speaker Yifan Jiang is a third-year Ph.D. student in the Information Science Institute at the University of Southern California (USC-ISI), advised by Dr. Jay Pujara, focusing on natural language processing, commonsense reasoning and multimodality large language models. Layer-Aware Video Composition via Split-then-Merge Split-then-Merge (StM) is a novel generative framework that overcomes data scarcity in video composition by splitting unlabeled videos into separate foreground and background layers for self-supervised learning. By utilizing a transformation-aware training pipeline with multi-layer fusion, the model learns to realistically compose dynamic subjects into diverse scenes without relying on expensive annotated datasets. This presentation will cover the problem of video composition and the details of StM, an approach looking at this problem from a generative AI perspective. We will conclude by demonstrating how StM is working, and outperforming state-of-the-art methods in both quantitative benchmarks and qualitative evaluations. About the Speaker Ozgur Kara is a 4th year Computer Science PhD student at the University of Illinois Urbana-Champaign (UIUC), advised by Founder Professor James M. Rehg. His research builds the next generation of video AI by tackling three core challenges: efficiency, controllability, and safety. Video Reasoning for Worker Safety Ensuring worker safety in industrial environments requires more than object detection or motion tracking; it demands a genuine understanding of human actions, context, and risk. This talk demonstrates how NVIDIA Cosmos Reason, a multimodal video-reasoning model, interprets workplace scenarios with sophisticated temporal and semantic awareness, identifying nuanced safe and unsafe behaviors that conventional vision systems frequently overlook. By integrating Cosmos Reason with FiftyOne, users achieve both automated safety assessments and transparent, interpretable explanations revealing why specific actions are deemed hazardous. Using a curated worker-safety dataset of authentic factory-floor footage, we show how video reasoning enhances audits, training, and compliance workflows while minimizing dependence on extensive labeled datasets. The resulting system demonstrates the potential of explainable multimodal AI to enable safer, more informed decision-making across manufacturing, logistics, construction, healthcare, and other sectors where understanding human behavior is essential. About the Speaker Paula Ramos has a PhD in Computer Vision and Machine Learning, with more than 20 years of experience in the technological field. She has been developing novel integrated engineering technologies, mainly in Computer Vision, robotics, and Machine Learning applied to agriculture, since the early 2000s in Colombia. Video Intelligence Is Going Agentic Video content has become ubiquitous in our digital world, yet the tools for working with video have remained largely unchanged for decades. This talk explores how the convergence of foundation models and agent architectures is fundamentally transforming video interaction and creation. We'll examine how video-native foundation models, multimodal interfaces, and agent transparency are reshaping enterprise media workflows through a deep dive into Jockey, a pioneering video agent system. About the Speaker James Le currently leads the developer experience function at TwelveLabs - a startup building foundation models for video understanding. He previously operated in the MLOps space and ran a blog/podcast on the Data & AI infrastructure ecosystem. |
Feb 11 - Visual AI for Video Use Cases
|
|
Feb 11 - Visual AI for Video Use Cases
2026-02-11 · 17:00
Join our virtual Meetup to hear talks from experts on cutting-edge topics at the intersection of Visual AI and video use cases. Time and Location Feb 11, 2026 9 - 11 AM Pacific Online. Register for the Zoom! VIDEOP2R: Video Understanding from Perception to Reasoning Reinforcement fine-tuning (RFT), a two-stage framework consisting of supervised fine-tuning (SFT) and reinforcement learning (RL) has shown promising results on improving reasoning ability of large language models (LLMs). Yet extending RFT to large video language models (LVLMs) remains challenging. We propose VideoP2R, a novel process-aware video RFT framework that enhances video reasoning by modeling perception and reasoning as distinct processes. In the SFT stage, we develop a three-step pipeline to generate VideoP2R-CoT-162K, a high-quality, process-aware chain-of-thought (CoT) dataset for perception and reasoning. In the RL stage, we introduce a novel process-aware group relative policy optimization (PA-GRPO) algorithm that supplies separate rewards for perception and reasoning. Extensive experiments show that VideoP2R achieves state-of-the-art (SotA) performance on six out of seven video reasoning and understanding benchmarks. Ablation studies further confirm the effectiveness of our process-aware modeling and PA-GRPO and demonstrate that model's perception output is information-sufficient for downstream reasoning. About the Speaker Yifan Jiang is a third-year Ph.D. student in the Information Science Institute at the University of Southern California (USC-ISI), advised by Dr. Jay Pujara, focusing on natural language processing, commonsense reasoning and multimodality large language models. Layer-Aware Video Composition via Split-then-Merge Split-then-Merge (StM) is a novel generative framework that overcomes data scarcity in video composition by splitting unlabeled videos into separate foreground and background layers for self-supervised learning. By utilizing a transformation-aware training pipeline with multi-layer fusion, the model learns to realistically compose dynamic subjects into diverse scenes without relying on expensive annotated datasets. This presentation will cover the problem of video composition and the details of StM, an approach looking at this problem from a generative AI perspective. We will conclude by demonstrating how StM is working, and outperforming state-of-the-art methods in both quantitative benchmarks and qualitative evaluations. About the Speaker Ozgur Kara is a 4th year Computer Science PhD student at the University of Illinois Urbana-Champaign (UIUC), advised by Founder Professor James M. Rehg. His research builds the next generation of video AI by tackling three core challenges: efficiency, controllability, and safety. Video Reasoning for Worker Safety Ensuring worker safety in industrial environments requires more than object detection or motion tracking; it demands a genuine understanding of human actions, context, and risk. This talk demonstrates how NVIDIA Cosmos Reason, a multimodal video-reasoning model, interprets workplace scenarios with sophisticated temporal and semantic awareness, identifying nuanced safe and unsafe behaviors that conventional vision systems frequently overlook. By integrating Cosmos Reason with FiftyOne, users achieve both automated safety assessments and transparent, interpretable explanations revealing why specific actions are deemed hazardous. Using a curated worker-safety dataset of authentic factory-floor footage, we show how video reasoning enhances audits, training, and compliance workflows while minimizing dependence on extensive labeled datasets. The resulting system demonstrates the potential of explainable multimodal AI to enable safer, more informed decision-making across manufacturing, logistics, construction, healthcare, and other sectors where understanding human behavior is essential. About the Speaker Paula Ramos has a PhD in Computer Vision and Machine Learning, with more than 20 years of experience in the technological field. She has been developing novel integrated engineering technologies, mainly in Computer Vision, robotics, and Machine Learning applied to agriculture, since the early 2000s in Colombia. Video Intelligence Is Going Agentic Video content has become ubiquitous in our digital world, yet the tools for working with video have remained largely unchanged for decades. This talk explores how the convergence of foundation models and agent architectures is fundamentally transforming video interaction and creation. We'll examine how video-native foundation models, multimodal interfaces, and agent transparency are reshaping enterprise media workflows through a deep dive into Jockey, a pioneering video agent system. About the Speaker James Le currently leads the developer experience function at TwelveLabs - a startup building foundation models for video understanding. He previously operated in the MLOps space and ran a blog/podcast on the Data & AI infrastructure ecosystem. |
Feb 11 - Visual AI for Video Use Cases
|
|
Feb 11 - Visual AI for Video Use Cases
2026-02-11 · 17:00
Join our virtual Meetup to hear talks from experts on cutting-edge topics at the intersection of Visual AI and video use cases. Time and Location Feb 11, 2026 9 - 11 AM Pacific Online. Register for the Zoom! VIDEOP2R: Video Understanding from Perception to Reasoning Reinforcement fine-tuning (RFT), a two-stage framework consisting of supervised fine-tuning (SFT) and reinforcement learning (RL) has shown promising results on improving reasoning ability of large language models (LLMs). Yet extending RFT to large video language models (LVLMs) remains challenging. We propose VideoP2R, a novel process-aware video RFT framework that enhances video reasoning by modeling perception and reasoning as distinct processes. In the SFT stage, we develop a three-step pipeline to generate VideoP2R-CoT-162K, a high-quality, process-aware chain-of-thought (CoT) dataset for perception and reasoning. In the RL stage, we introduce a novel process-aware group relative policy optimization (PA-GRPO) algorithm that supplies separate rewards for perception and reasoning. Extensive experiments show that VideoP2R achieves state-of-the-art (SotA) performance on six out of seven video reasoning and understanding benchmarks. Ablation studies further confirm the effectiveness of our process-aware modeling and PA-GRPO and demonstrate that model's perception output is information-sufficient for downstream reasoning. About the Speaker Yifan Jiang is a third-year Ph.D. student in the Information Science Institute at the University of Southern California (USC-ISI), advised by Dr. Jay Pujara, focusing on natural language processing, commonsense reasoning and multimodality large language models. Layer-Aware Video Composition via Split-then-Merge Split-then-Merge (StM) is a novel generative framework that overcomes data scarcity in video composition by splitting unlabeled videos into separate foreground and background layers for self-supervised learning. By utilizing a transformation-aware training pipeline with multi-layer fusion, the model learns to realistically compose dynamic subjects into diverse scenes without relying on expensive annotated datasets. This presentation will cover the problem of video composition and the details of StM, an approach looking at this problem from a generative AI perspective. We will conclude by demonstrating how StM is working, and outperforming state-of-the-art methods in both quantitative benchmarks and qualitative evaluations. About the Speaker Ozgur Kara is a 4th year Computer Science PhD student at the University of Illinois Urbana-Champaign (UIUC), advised by Founder Professor James M. Rehg. His research builds the next generation of video AI by tackling three core challenges: efficiency, controllability, and safety. Video Reasoning for Worker Safety Ensuring worker safety in industrial environments requires more than object detection or motion tracking; it demands a genuine understanding of human actions, context, and risk. This talk demonstrates how NVIDIA Cosmos Reason, a multimodal video-reasoning model, interprets workplace scenarios with sophisticated temporal and semantic awareness, identifying nuanced safe and unsafe behaviors that conventional vision systems frequently overlook. By integrating Cosmos Reason with FiftyOne, users achieve both automated safety assessments and transparent, interpretable explanations revealing why specific actions are deemed hazardous. Using a curated worker-safety dataset of authentic factory-floor footage, we show how video reasoning enhances audits, training, and compliance workflows while minimizing dependence on extensive labeled datasets. The resulting system demonstrates the potential of explainable multimodal AI to enable safer, more informed decision-making across manufacturing, logistics, construction, healthcare, and other sectors where understanding human behavior is essential. About the Speaker Paula Ramos has a PhD in Computer Vision and Machine Learning, with more than 20 years of experience in the technological field. She has been developing novel integrated engineering technologies, mainly in Computer Vision, robotics, and Machine Learning applied to agriculture, since the early 2000s in Colombia. Video Intelligence Is Going Agentic Video content has become ubiquitous in our digital world, yet the tools for working with video have remained largely unchanged for decades. This talk explores how the convergence of foundation models and agent architectures is fundamentally transforming video interaction and creation. We'll examine how video-native foundation models, multimodal interfaces, and agent transparency are reshaping enterprise media workflows through a deep dive into Jockey, a pioneering video agent system. About the Speaker James Le currently leads the developer experience function at TwelveLabs - a startup building foundation models for video understanding. He previously operated in the MLOps space and ran a blog/podcast on the Data & AI infrastructure ecosystem. |
Feb 11 - Visual AI for Video Use Cases
|
|
Feb 11 - Visual AI for Video Use Cases
2026-02-11 · 17:00
Join our virtual Meetup to hear talks from experts on cutting-edge topics at the intersection of Visual AI and video use cases. Time and Location Feb 11, 2026 9 - 11 AM Pacific Online. Register for the Zoom! VIDEOP2R: Video Understanding from Perception to Reasoning Reinforcement fine-tuning (RFT), a two-stage framework consisting of supervised fine-tuning (SFT) and reinforcement learning (RL) has shown promising results on improving reasoning ability of large language models (LLMs). Yet extending RFT to large video language models (LVLMs) remains challenging. We propose VideoP2R, a novel process-aware video RFT framework that enhances video reasoning by modeling perception and reasoning as distinct processes. In the SFT stage, we develop a three-step pipeline to generate VideoP2R-CoT-162K, a high-quality, process-aware chain-of-thought (CoT) dataset for perception and reasoning. In the RL stage, we introduce a novel process-aware group relative policy optimization (PA-GRPO) algorithm that supplies separate rewards for perception and reasoning. Extensive experiments show that VideoP2R achieves state-of-the-art (SotA) performance on six out of seven video reasoning and understanding benchmarks. Ablation studies further confirm the effectiveness of our process-aware modeling and PA-GRPO and demonstrate that model's perception output is information-sufficient for downstream reasoning. About the Speaker Yifan Jiang is a third-year Ph.D. student in the Information Science Institute at the University of Southern California (USC-ISI), advised by Dr. Jay Pujara, focusing on natural language processing, commonsense reasoning and multimodality large language models. Layer-Aware Video Composition via Split-then-Merge Split-then-Merge (StM) is a novel generative framework that overcomes data scarcity in video composition by splitting unlabeled videos into separate foreground and background layers for self-supervised learning. By utilizing a transformation-aware training pipeline with multi-layer fusion, the model learns to realistically compose dynamic subjects into diverse scenes without relying on expensive annotated datasets. This presentation will cover the problem of video composition and the details of StM, an approach looking at this problem from a generative AI perspective. We will conclude by demonstrating how StM is working, and outperforming state-of-the-art methods in both quantitative benchmarks and qualitative evaluations. About the Speaker Ozgur Kara is a 4th year Computer Science PhD student at the University of Illinois Urbana-Champaign (UIUC), advised by Founder Professor James M. Rehg. His research builds the next generation of video AI by tackling three core challenges: efficiency, controllability, and safety. Video Reasoning for Worker Safety Ensuring worker safety in industrial environments requires more than object detection or motion tracking; it demands a genuine understanding of human actions, context, and risk. This talk demonstrates how NVIDIA Cosmos Reason, a multimodal video-reasoning model, interprets workplace scenarios with sophisticated temporal and semantic awareness, identifying nuanced safe and unsafe behaviors that conventional vision systems frequently overlook. By integrating Cosmos Reason with FiftyOne, users achieve both automated safety assessments and transparent, interpretable explanations revealing why specific actions are deemed hazardous. Using a curated worker-safety dataset of authentic factory-floor footage, we show how video reasoning enhances audits, training, and compliance workflows while minimizing dependence on extensive labeled datasets. The resulting system demonstrates the potential of explainable multimodal AI to enable safer, more informed decision-making across manufacturing, logistics, construction, healthcare, and other sectors where understanding human behavior is essential. About the Speaker Paula Ramos has a PhD in Computer Vision and Machine Learning, with more than 20 years of experience in the technological field. She has been developing novel integrated engineering technologies, mainly in Computer Vision, robotics, and Machine Learning applied to agriculture, since the early 2000s in Colombia. Video Intelligence Is Going Agentic Video content has become ubiquitous in our digital world, yet the tools for working with video have remained largely unchanged for decades. This talk explores how the convergence of foundation models and agent architectures is fundamentally transforming video interaction and creation. We'll examine how video-native foundation models, multimodal interfaces, and agent transparency are reshaping enterprise media workflows through a deep dive into Jockey, a pioneering video agent system. About the Speaker James Le currently leads the developer experience function at TwelveLabs - a startup building foundation models for video understanding. He previously operated in the MLOps space and ran a blog/podcast on the Data & AI infrastructure ecosystem. |
Feb 11 - Visual AI for Video Use Cases
|
|
Feb 11 - Visual AI for Video Use Cases
2026-02-11 · 17:00
Join our virtual Meetup to hear talks from experts on cutting-edge topics at the intersection of Visual AI and video use cases. Time and Location Feb 11, 2026 9 - 11 AM Pacific Online. Register for the Zoom! VIDEOP2R: Video Understanding from Perception to Reasoning Reinforcement fine-tuning (RFT), a two-stage framework consisting of supervised fine-tuning (SFT) and reinforcement learning (RL) has shown promising results on improving reasoning ability of large language models (LLMs). Yet extending RFT to large video language models (LVLMs) remains challenging. We propose VideoP2R, a novel process-aware video RFT framework that enhances video reasoning by modeling perception and reasoning as distinct processes. In the SFT stage, we develop a three-step pipeline to generate VideoP2R-CoT-162K, a high-quality, process-aware chain-of-thought (CoT) dataset for perception and reasoning. In the RL stage, we introduce a novel process-aware group relative policy optimization (PA-GRPO) algorithm that supplies separate rewards for perception and reasoning. Extensive experiments show that VideoP2R achieves state-of-the-art (SotA) performance on six out of seven video reasoning and understanding benchmarks. Ablation studies further confirm the effectiveness of our process-aware modeling and PA-GRPO and demonstrate that model's perception output is information-sufficient for downstream reasoning. About the Speaker Yifan Jiang is a third-year Ph.D. student in the Information Science Institute at the University of Southern California (USC-ISI), advised by Dr. Jay Pujara, focusing on natural language processing, commonsense reasoning and multimodality large language models. Layer-Aware Video Composition via Split-then-Merge Split-then-Merge (StM) is a novel generative framework that overcomes data scarcity in video composition by splitting unlabeled videos into separate foreground and background layers for self-supervised learning. By utilizing a transformation-aware training pipeline with multi-layer fusion, the model learns to realistically compose dynamic subjects into diverse scenes without relying on expensive annotated datasets. This presentation will cover the problem of video composition and the details of StM, an approach looking at this problem from a generative AI perspective. We will conclude by demonstrating how StM is working, and outperforming state-of-the-art methods in both quantitative benchmarks and qualitative evaluations. About the Speaker Ozgur Kara is a 4th year Computer Science PhD student at the University of Illinois Urbana-Champaign (UIUC), advised by Founder Professor James M. Rehg. His research builds the next generation of video AI by tackling three core challenges: efficiency, controllability, and safety. Video Reasoning for Worker Safety Ensuring worker safety in industrial environments requires more than object detection or motion tracking; it demands a genuine understanding of human actions, context, and risk. This talk demonstrates how NVIDIA Cosmos Reason, a multimodal video-reasoning model, interprets workplace scenarios with sophisticated temporal and semantic awareness, identifying nuanced safe and unsafe behaviors that conventional vision systems frequently overlook. By integrating Cosmos Reason with FiftyOne, users achieve both automated safety assessments and transparent, interpretable explanations revealing why specific actions are deemed hazardous. Using a curated worker-safety dataset of authentic factory-floor footage, we show how video reasoning enhances audits, training, and compliance workflows while minimizing dependence on extensive labeled datasets. The resulting system demonstrates the potential of explainable multimodal AI to enable safer, more informed decision-making across manufacturing, logistics, construction, healthcare, and other sectors where understanding human behavior is essential. About the Speaker Paula Ramos has a PhD in Computer Vision and Machine Learning, with more than 20 years of experience in the technological field. She has been developing novel integrated engineering technologies, mainly in Computer Vision, robotics, and Machine Learning applied to agriculture, since the early 2000s in Colombia. Video Intelligence Is Going Agentic Video content has become ubiquitous in our digital world, yet the tools for working with video have remained largely unchanged for decades. This talk explores how the convergence of foundation models and agent architectures is fundamentally transforming video interaction and creation. We'll examine how video-native foundation models, multimodal interfaces, and agent transparency are reshaping enterprise media workflows through a deep dive into Jockey, a pioneering video agent system. About the Speaker James Le currently leads the developer experience function at TwelveLabs - a startup building foundation models for video understanding. He previously operated in the MLOps space and ran a blog/podcast on the Data & AI infrastructure ecosystem. |
Feb 11 - Visual AI for Video Use Cases
|