Behind the growing interest in Generate AI and LLM-based enterprise applications lies an expanded set of requirements for data integrations and ML orchestration. Enterprises want to use proprietary data to power LLM-based applications that create new business value, but they face challenges in moving beyond experimentation. The pipelines that power these models need to run reliably at scale, bringing together data from many sources and reacting continuously to changing conditions. This talk focuses on the design patterns for using Apache Airflow to support LLM applications created using private enterprise data. We’ll go through a real-world example of what this looks like, as well as a proposal to improve Airflow and to add additional Airflow Providers to make it easier to interact with LLMs such as the ones from OpenAI (such as GPT4) and the ones on HuggingFace, while working with both structured and unstructured data. In short, this shows how these Airflow patterns enable reliable, traceable, and scalable LLM applications within the enterprise.
talk-data.com
J
Speaker
Julian LaNeve
1
talks
Chief Technology Officer at Astronomer
Astronomer
CTO at Astronomer.
Bio from: NYC Airflow Rooftop Happy Hour ft. PMC Member Jarek Potiuk!
Frequent Collaborators
Filtering by:
Airflow Summit 2023
×
Filter by Event / Source
Talks & appearances
Showing 1 of 6 activities