-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Anirud/add rag context to chat thread #199
Merged
anirudTT
merged 24 commits into
anirud/multi-modal-model-support
from
anirud/add-rag-context-to-chat-thread
Feb 25, 2025
Merged
Anirud/add rag context to chat thread #199
anirudTT
merged 24 commits into
anirud/multi-modal-model-support
from
anirud/add-rag-context-to-chat-thread
Feb 25, 2025
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
* Update CONTRIBUTING.md and Git-management.png to v1.1.0 --------- Co-authored-by: bgoelTT <bgoel@tenstorrent.com>
* copy in agent control dir * add TAVILY API key to .env.default * copy run agent container + helper func * copied in updated docker views * copied in model utils to stream agent response * copy in model views * added agent view * copy in all frontend components * add search api key to docker compose yml * copy in updated model urls * added requirements for dockerfile * rename hf_model_id * remove commetned code intepretor tool * added fix so agent works with other llama models * fix requirements in dockerfile * add thread id to match stateful chat * add readme * add agent workflow diagram * Update README.md * Delete app/api/agent_control/Agent.png * Add files via upload * Delete app/api/agent_control/Agent.png * Add files via upload * Update README.md * Delete app/api/agent_control/Agent.png * Add files via upload * Delete app/api/agent_control/Agent.png * Add files via upload * Update README.md
…upport - Add RAG datasource integration with metadata display - Create reusable FileDisplay component for file management - Implement FileViewerDialog for improved file preview experience - Support both image and non-image file types with download option - Clean up file handling logic and separate from image-specific code - Add visual indicator for RAG-enabled messages
- Use the RAG datasource from the message if available
- Remove redundant and commented-out interfaces - Group related interfaces together (chat, inference, file, voice) - Add proper JSDoc comments for better documentation - Consolidate duplicate type definitions - Add explicit typing for RAG-related interfaces
- show images in better aspect ratio
- File extensions mapping for code files and other file types
* update readme to reflect new flow * fix readme issues * add Supported models tab: pointing to tt-inference-server readme * docs: Update main readme - add better quick start guide - add better notes for running in development mode * docs: re add Mock model steps * docs: fix links * docs: fix vllm * Update HowToRun_vLLM_Models.md Co-authored-by: Benjamin Goel <bgoel@tenstorrent.com> * Update HowToRun_vLLM_Models.md --------- Co-authored-by: Benjamin Goel <bgoel@tenstorrent.com>
ac2d756
to
1b6ade1
Compare
This was
linked to
issues
Feb 25, 2025
anirudTT
added a commit
that referenced
this pull request
Feb 25, 2025
* copy run agent container + helper func * copied in updated docker views * copied in model utils to stream agent response * copy in model views * added agent view * copy in all frontend components * add search api key to docker compose yml * copy in updated model urls * added requirements for dockerfile * rename hf_model_id * remove commetned code intepretor tool * added fix so agent works with other llama models * fix requirements in dockerfile * add thread id to match stateful chat * add readme * add agent workflow diagram * Update README.md * Delete app/api/agent_control/Agent.png * Add files via upload * Delete app/api/agent_control/Agent.png * Add files via upload * Update README.md * Delete app/api/agent_control/Agent.png * Add files via upload * Delete app/api/agent_control/Agent.png * Add files via upload * Update README.md * fix link href (#180) * refactor(chat history component): improve file handling and add RAG support - Add RAG datasource integration with metadata display - Create reusable FileDisplay component for file management - Implement FileViewerDialog for improved file preview experience - Support both image and non-image file types with download option - Clean up file handling logic and separate from image-specific code - Add visual indicator for RAG-enabled messages * Show RAG pill based on the message's stored RAG context * feat(add support in chat component): - Use the RAG datasource from the message if available * move image display to its own component * include rag source name when selected * refactor(types): clean up and organize type definitions - Remove redundant and commented-out interfaces - Group related interfaces together (chat, inference, file, voice) - Add proper JSDoc comments for better documentation - Consolidate duplicate type definitions - Add explicit typing for RAG-related interfaces * add pdfjs-dist to test * feat: improve file display - show images in better aspect ratio * display file display for images ,code files and other file types * add icons for file display in chat thread * extend types * extend to add - File extensions mapping for code files and other file types * extend to allow for files to be passed as text * fix alignment * limit upload to a single image file * set focused state in input area * feat: add ability to process multiple code and or text file types and send to model * re add resizing input are * fix copy button logic * Anirud/update vllm setup steps (#189) * update readme to reflect new flow * fix readme issues * add Supported models tab: pointing to tt-inference-server readme * docs: Update main readme - add better quick start guide - add better notes for running in development mode * docs: re add Mock model steps * docs: fix links * docs: fix vllm * Update HowToRun_vLLM_Models.md * Update HowToRun_vLLM_Models.md
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Changelog