Building Real-Time AI/ML Agents with Distributed Data using Apache Cassandra and Astra DB
Are your LLMs stuck on outdated data? Learn how the RAG pattern and a vector database can build smarter, context-aware AI agents.
#1about 3 minutes
Introducing the DataStax real-time data cloud
The platform combines Apache Cassandra, Apache Pulsar, and Kaskada to provide a flexible database, streaming, and machine learning solution for developers.
#2about 3 minutes
Interacting with Astra DB using GraphQL and REST APIs
A live demonstration shows how to create a schema, ingest data, and query tables in Astra DB using both GraphQL and REST API endpoints.
#3about 1 minute
Understanding real-time AI and its applications
Real-time AI leverages the most recent data to power predictive analytics and automated actions, as seen in use cases from Uber and Netflix.
#4about 2 minutes
What is Retrieval Augmented Generation (RAG)?
RAG is a pattern that allows large language models to access and use your proprietary, up-to-date data to provide contextually relevant responses.
#5about 3 minutes
Key steps for building a generative AI agent
The process involves defining the agent's purpose, choosing an LLM, selecting context data, picking an embedding model, and performing prompt engineering.
#6about 3 minutes
Exploring the architecture of a RAG system
A RAG system uses a vector database to perform a similarity search on data embeddings, finding relevant context to enrich the prompt sent to the LLM.
#7about 3 minutes
Generating vector embeddings from text content
A Jupyter Notebook demonstrates splitting source text into chunks and using an embedding model to create vector representations for storage and search.
#8about 4 minutes
The end-to-end data flow of a RAG query
A user's question is converted into an embedding, used for a similarity search in the vector store, and the results are combined with other context to build a final prompt.
#9about 3 minutes
Executing a RAG prompt to get an LLM response
The demo shows how the context-enriched prompt is sent to an LLM to generate a relevant answer, including how to add memory for conversational history.
#10about 3 minutes
Getting started with the Astra DB vector database
Resources are provided for getting started with Astra DB, including quick starts, a free tier for developers, and information on multi-cloud region support.
Related jobs
Jobs that call for the skills explored in this talk.
Matching moments
02:42 MIN
Powering real-time AI with retrieval augmented generation
Scrape, Train, Predict: The Lifecycle of Data for AI Applications
09:46 MIN
Code walkthrough for building a RAG-based chatbot
Creating Industry ready solutions with LLM Models
03:08 MIN
Leveraging Astra DB for high-relevance vector search
Accelerating GenAI Development: Harnessing Astra DB Vector Store and Langflow for LLM-Powered Apps
Panel Discussion: Responsible AI in Practice - Real-World Examples and ChallengesIntroductionIn the ever-evolving landscape of artificial intelligence, the concept of "responsible AI" has emerged as a cornerstone for ethical and practical AI implementation. During the WWC24 Panel discussion, three eminent experts—Mina, Bjorn Brin...
Stephan Gillich - Bringing AI EverywhereIn the ever-evolving world of technology, AI continues to be the frontier for innovation and transformation. Stephan Gillich, from the AI Center of Excellence at Intel, dove into the subject in a recent session titled "Bringing AI Everywhere," sheddi...
Eli McGarvie
13 AI Tools You Have to TryFirst, it was NFTs, then it was Web3, and now it’s generative AI… it’s probably time to stop collecting pictures of monkeys and kitties. Chatbots and generative AI are the next big thing. This time we’ve jumped on a trend that has real-world applicat...
From learning to earning
Jobs that call for the skills explored in this talk.