talk-data.com talk-data.com

Topic

llms

102

tagged

Activity Trend

19 peak/qtr
2020-Q1 2026-Q1

Activities

102 activities · Newest first

Vanilla Retrieval-Augmented Generation (RAG) is becoming more and more adopted - but What is next? Join us for an introduction to Generative Feedback Loops (GFL). GFLs can level-up your RAG architecture by generating more insights on top of your data. This empowers you to implement smarter chatbots, agents, and other AI-driven solutions. Based on hands-on examples, we will explore the following questions: a) What are GFLs are? b) How do GFLs work? c) What challenges and use-cases can we tackle with GFLs; and d) How can I define my AI workflows to implement GFLs. Learn about the next level of AI applications and join us for a Live Demo.

Overview of how Small Language Models (SLMs) can solve business problems and boost device intelligence, with a comparison to Large Language Models (LLMs). Focus on fine-tuning and customization, edge deployments for real-time processing, and a cloud-to-edge deployment roadmap using Azure to maintain IP ownership and integration. Includes perspectives on leveraging SLMs to revolutionize devices and mentions Microsoft’s Student Ambassador Program energizing students to advance intelligent tech.

During the session, we’ll discuss the challenges that prompt engineering has presented, both when it first gained popularity and as it continued to evolve. We’ll share how these challenges informed the development of our prompt engineering tooling and workflows. We’ll cover: Standardizing communication with LLMs; Using templating to customize prompts; Building prompt-centric production workflows; Working with structured LLM output; Ensuring the quality of LLM output; Creating tooling that supports our prompt engineering workflows.

One of the latest innovations in Generative AI (GenAI) technology are Large Language Models (LLMs). You can exploit LLMs to solve many tasks that were incredibly challenging not that long ago, and some say you can solve anything with GenAI. But can you? Recently, we decided to reshape an old application for fraud detection by introducing AI in it. This talk is a summary of both our successful and failed attempts to solve everything with AI.

Organizations develop feedback loops to continuously enhance quality. One such loop is the learning from user interactions with your data, retraining models, deploying new models and learning again. The learning curve to create a loop like this is steep, it requires ML experience and tools. However, most teams can easily provide labeled examples. In-Context Learning (ICL) is a method to add classification examples as input to foundation models (like LLMs).\nThis talk defines an Adaptive ICL strategy using Retrieval for Examples, where the output is used for content retrieval, example set expansion for future model training and real-time user behaviour tracking. Adaptive ICL is hence an easy way for teams to get immediate results with AI, while laying the foundation for more advanced ML loops in the future.

LLMs have unlocked new opportunities in NLP with their possible applications. Features that used to take months to be planned and developed now require a day to be prototyped. But how can we make sure that a successful prototype will turn into a high-quality feature useful for millions of customers? In this talk, we will explore real examples of the challenges that arise when ensuring the quality of LLM outputs and how we address them at Grammarly.

In this talk, we will delve into the emerging challenges that generative artificial intelligence GenAI and Large Language Models (LLMs) bring to the world of software quality and testing. We will explore how the integration, use, or design of solutions with GenAI models pose new challenges, including data quality and bias detection, dealing with non-determinism in our test automation and privacy and security concerns. We will review some proposals that are being implemented to adjust our testing tasks to the development of such systems, including approaches to automated testing and observability.

An intro to RAGHack, a global hackathon to develop apps using LLMs and RAG. A large language model (LLM) like GPT-4 can be used for summarization, translation, entity extraction, and question-answering. Retrieval Augmented Generation (RAG) is an approach that sends context to the LLM so that it can provide grounded answers. RAG apps can be developed on Azure using a wide range of programming languages and retrievers (such as AI Search, Cosmos DB, PostgreSQL, and Azure SQL). Get an overview of RAG in this session before diving deep in our follow-up streams.

LLMs like GPT can give useful answers to many questions, but there are also well-known issues with their output: The responses may be outdated, inaccurate, or outright hallucinations, and it’s hard to know when you can trust them. And they don’t know anything about you or your organization private data (we hope). RAG can help reduce the problems with “hallucinated” answers, and make the responses more up-to-date, accurate, and personalized - by injecting related knowledge, including non-public data. In this talk, we’ll go through what RAG means, demo some ways you can implement it - and warn of some traps you still have to watch out for.