Replies: 1 comment
-
Yes we’re always happy to make things faster. Afaik the time spent in this function is mostly necessary though. It’s primarily spending time on encoding the passages with BERT, not quite something we can optimize away. you could confirm with profiling that the vast majority of time is spent inside BERT, not in our code. That said, recent pytorch and transformers optimizations (over the past year or two) are not reflected into this component. You may for instance be able to compile the BERT model into a faster version that’s more static. I’d look into ONNX BERT. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I am trying to index large collections more quickly. It seems the doc function in colbert.py
ColBERT/colbert/modeling/colbert.py
Line 94 in 83658dc
Thanks,
Marc Mason
Beta Was this translation helpful? Give feedback.
All reactions