- 
          
- 
                Notifications
    You must be signed in to change notification settings 
- Fork 10.9k
Closed
Labels
performancePerformance-related issuesPerformance-related issuesstaleOver 90 days of inactivityOver 90 days of inactivity
Description
Proposal to improve performance
Currently, hash_request_tokens executes in engine core to compute hashes of blocks based on the request token IDs (and lora IDs, MM tokens, etc). And the current design make it to become the hard blocker of inferences.
As shown in the following charts, for small models opt128m with QPS 200 (input=700, output=1) scenarios, noticeable amount of time is used compute the hash.

Ideally, in order to compute the hashes, all dependent metadata should be ready when the data received on input_socket processing threads who is running in parallel with engine core thread. With this move, we would be able to move the hashes computation out from critical path, as shown in the following chart.

Report of performance regression
N/A
Misc discussion on performance
N/A
Your current environment (if you think it is necessary)
The output of `python collect_env.py`
N/A
Before submitting a new issue...
- Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
yeqcharlotte
Metadata
Metadata
Assignees
Labels
performancePerformance-related issuesPerformance-related issuesstaleOver 90 days of inactivityOver 90 days of inactivity