Open
Description
Description
Currently, we are truncating long articles. This was a temporary measure to reduce the context length required but has some significant drawbacks.
Now that ColBERT is demonstrably fast enough for online tasks, we should consider chunking all provided articles and re-ranking the chunks instead of the full articles. This can further reduce our input token budget while resolving the issue of some articles being too long.
There are some concerns regarding the chunking methods available. Will the LLM be able to compensate for missing breadcrumb-like title context?
Metadata
Assignees
Labels
No labels
Activity