Skip to content

Latest commit

 

History

History
7 lines (5 loc) · 1 KB

README.md

File metadata and controls

7 lines (5 loc) · 1 KB

About this repository

This is a 3-part series on the topic 'talk to your slide deck' using multimodal foundation models hosted on Amazon Bedrock and Amazon SageMaker.

  • Part 1 of the series implements a solution design that first converts slides into embeddings and stores these embeddings in a vector store (Amazon OpenSearch Serverless). When a user queries this data, LLaVA 1.5 is invoked and inference returned to user.
  • Part 2 of the series follows a different solution design. This approach will generate and store LLaVA 1.5 inferences in a vector store (Amazon OpenSearch Serverless) and use those stored inferences to respond to user queries.
  • Part 3 of the series will compare the two approaches.