You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
ONNXRuntime takes up too much memory (more like accumulating cause I believe it's not freeing up unused memory), when trying to embed large collections of data.
Am I missing something or is this a problem with the runtime itself.
Iam trying to embed about 10000 documents (Average Size - 3000 characters) using the JinaAI Colbert Model (Late Interaction Model)
GPU - Tesla T4 16 GB VRam
The text was updated successfully, but these errors were encountered:
What's the batch size you're using?
Are you keeping the embeddings in memory or are you uploading them somewhere else / write to disk?
Colbert embeddings are quite huge since colbert produce 128-dim embedding per token
As for now, I was able to reproduce the issue and it indeed seems like a problem with onnxruntime not freeing up the space
However, we might need more time to investigate it, thank you for pointing it out
ONNXRuntime takes up too much memory (more like accumulating cause I believe it's not freeing up unused memory), when trying to embed large collections of data.
Am I missing something or is this a problem with the runtime itself.
Iam trying to embed about 10000 documents (Average Size - 3000 characters) using the JinaAI Colbert Model (Late Interaction Model)
GPU - Tesla T4 16 GB VRam
The text was updated successfully, but these errors were encountered: