-
Notifications
You must be signed in to change notification settings - Fork 43
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
connect timeout error for Ollama server when trying to insert embeddings for a large number of documents #191
Comments
Do you have an idea of how many documents it can embed safely before this error appears? This timeout seems to be based on the capacity of the machine itself. I tried in two different machines and got two different sizes before timeout. I am thinking it may make sense to provide an optional parameter to specify the chunk size which will break the insert into different groups inserted sequentially. |
I will test incrementally and let you know by the end of today. As of now, I have been able to insert 2000 documents in one go. Note: When I use the term document, I mean it in the sense of an article with anywhere between a couple of hundred to several thousand words (these are excerpts of articles from scientific papers).
I am using an M1 MacBook Pro with 16 GB RAM. The SQLite database (with almost a million articles) was created on the same machine without any problem, and took a couple of hours to create. The original data were individual XML files that were parsed and broken up into tables, with the main table containing a lot of metadata as well as the fulltext of the articles. This fulltext is what I am using to create embeddings and store as the vector database, also in SQLite format. I will report back today with the limit at which my process breaks. Many thanks. |
an update:
|
From what I understand, input text has to be tokenized and chunked so that embeddings of the right size can be generated. This level of chunking of my input text, the individual articles, is already happening internally managed by Using rough calculations, if 5000 articles took ~15 mins, a million articles will take about 300 mins, or 5 hours, to generate embeddings. |
This issue is stale because it has been open for 14 days with no activity. |
🐛 Describe the bug
Using the following code (that works for smaller number of documents -- tested with
LIMIT 200
)I got the following error
Ollama is running and is available on port 11434
The text was updated successfully, but these errors were encountered: