Replies: 2 comments 2 replies
-
|
Anyone? :) |
Beta Was this translation helpful? Give feedback.
0 replies
-
|
Hi, |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hello,
I’m working with RAGFlow on an Azure VM and want to create a dataset for enterprise use.
I have a large local folder (~90GB) containing many PDF and DOCX files.
My questions are:
Do I need to upload all files into the system in order to use them as a dataset, or is there a way for RAGFlow to reference a folder/directory directly?
Is there a method to “see” or register existing files in the dataset without uploading them again (for example, by mounting a folder, indexing, or using external storage like S3)?
What is the recommended workflow for handling very large datasets (tens of GBs, thousands of files) in RAGFlow for analytics and Q&A use cases?
I’m mainly looking for an efficient approach that avoids manual upload of thousands of files one by one.
Thanks in advance!
Beta Was this translation helpful? Give feedback.
All reactions