Skip to content

Demonstration modelling how a chained LLM agent workflow could be used to expedite the generation of personal contextual data for personalised LLM experiences and personal RAG development

Notifications You must be signed in to change notification settings

danielrosehill/Agentic-Context-Development-Interview-Demo

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Context Extraction Demo

Made with Streamlit Python Version License: MIT Hugging Face Space

This project demonstrates an innovative approach to enhancing personalized Large Language Model (LLM) experiences through agentic workflow-based context extraction. The system showcases how AI agents can proactively generate and collect contextual data to improve the quality and relevance of LLM interactions.

Purpose

The primary goal of this project is to illustrate how an agent-driven workflow can:

  • Proactively identify and extract relevant contextual information
  • Generate meaningful data that enhances LLM understanding
  • Create more personalized and context-aware AI interactions
  • Demonstrate practical implementation of agentic workflows in LLM systems

Screenshots

Step 1: AI agent asks you questions according to your preferred area of focus

alt text

Step 2: Sit back and talk about yourself for a while

The bot asked me if I could wake up with one magic power in the morning what would it be? This is the SFW version:

alt text

When you're done, click on 'End Interview' and your interview with the bot is done:

alt text

The chat transcript is then parsed, mined for contextual data, and reformatted for ingress to a RAG pipeline / vector DB

(Behind the scenes)

You get your context data out the other end!

alt text

Download and load into an agent for personalised LLM!

The reformatted contextual data snippets from the interviews are provided as downloadable markdown files. Markdown was chosen for its compact nature, its versatility, and its ubiquitous presence in the world of large language models.

alt text

These marked on files can then be aggregated, uploaded, or added to a RAG pipeline and added to an agent for personalized large language model (LLM) experiences.

An iterative workflow is envisioned whereby the user engages in a few interviews at a time, feeding these into vector database storage and progressively increasing the pool of personal context data available to the tools being worked with.

About

This project was developed through collaboration between Daniel Rosehill and Claude (Anthropic). It serves as a practical demonstration of how AI systems can be designed to actively participate in context generation and enhancement, leading to more effective and personalized LLM experiences.

Implementation

The system implements an agentic workflow that enables:

  • Automated context extraction from user interactions
  • Proactive generation of contextual metadata
  • Integration of extracted context into LLM inference processes
  • Enhanced personalization through accumulated contextual understanding

Attribution

Development: Claude (Anthropic) Project Direction and Implementation: Daniel Rosehill

About

Demonstration modelling how a chained LLM agent workflow could be used to expedite the generation of personal contextual data for personalised LLM experiences and personal RAG development

Topics

Resources

Stars

Watchers

Forks

Languages