Skip to content

The Document Question and Answering RAG Application is designed to provide intelligent and accurate responses to questions based on the content of uploaded documents. By combining state-of-the-art machine learning models and cloud technologies, this application offers a seamless and scalable

Notifications You must be signed in to change notification settings

manoharpalanisamy/Build-and-deploy-RAG-in-AWS-lambda-ECR-docker-langchain-Huggingface-

Repository files navigation

Document Question and Answering RAG APPLICATION

🚀 Welcome to the Document Question and Answering (Q&A) RAG Application! This project leverages cutting-edge technologies to create an efficient Q&A system. From model inference to deployment, I've got you covered.

Overview

ℹ️ This project aims to streamline the process of extracting information from documents by answering user questions using the Retrieve and Generate (RAG) approach. It harnesses the power of AWS Bedrock, AWS AppRunner, AWS ECR, Langchain, FAISS, Docker, GitHub Actions, LangSmith, and Streamlit.

📈 Continuous Integration, Delivery, and Deployment

The project follows best practices for CI/CD/CD, leveraging GitHub Actions for automation:

  • Continuous Integration: Ensures that every change is tested and validated(GitHub).
  • Continuous Delivery: Automates the release process, making deployments frequent and reliable(AWS ECR).
  • Continuous Deployment: Automatically deploys every validated change to production, ensuring fast and reliable updates(AWS AppRunner).

⚙️ AWS Integration: Leveraging AWS services like Bedrock, AppRunner, and ECR, I've created a scalable and reliable infrastructure for hosting our application.

🔑 Environment Variables: I've securely managed our project's deployment environment variables using GitHub Actions, ensuring a smooth deployment process.

Technologies Used

🛠️ AWS Bedrock: Provides the foundational infrastructure for the application, ensuring scalability and reliability.

📦 AWS AppRunner: Simplifying container-based application deployment and management.

🐳 Docker: Containerizing our application for easy deployment and scalability.

🤖 Langchain: Harnessing the power of language models for advanced natural language processing tasks.

🔍 FAISS: Employing efficient similarity search for document retrieval.

🛡️ GitHub Actions: Automating our CI/CD pipeline for efficient development workflows.

🔗 LangSmith: Integrating powerful language models for inference.

🌐 Streamlit: Creating interactive and user-friendly web applications with ease.

Model Inference and Embeddings

🧠 Model Inference: I utilize meta.llama2-13b-chat-v1 for model inference, ensuring accurate responses to user queries.

🔠 Vector Embeddings: Leveraging amazon.titan-embed-text-v1 for generating vector embeddings, enabling efficient document representation.

📧 Contact

For any questions or feedback, feel free to reach out!

Contributing

🤝 Contributions are welcome! Feel free to fork the repository and submit pull requests with your enhancements.

License

📄 This project is licensed under the MIT License, allowing for open collaboration and use.

Happy coding! 💻✨

About

The Document Question and Answering RAG Application is designed to provide intelligent and accurate responses to questions based on the content of uploaded documents. By combining state-of-the-art machine learning models and cloud technologies, this application offers a seamless and scalable

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published