v0.1.5
What's Changed #27
- add vllm>=0.4.1 by @liyucheng09 in #19, #44
- Feature(MInference): update HF demo information, thanks @ak's sponsoring by @iofu728 in #22
- Feature(MInference): add unittest by @iofu728 in #31, #32
- Feature(MInference): add triton-based decoding in case flash_attn is not available by @liyucheng09 in #35
- Feature(MInference): add e2e benchmark using vllm by @iofu728 in #49
- Feature(MInference): support llama 3.1 by @iofu728 in #54
- Hotfix(MInference): fix the import warnings, fix the apply_rotary_pos… by @iofu728 in #30
New Contributors
- @liyucheng09 made their first contribution in #19
Full Changelog: v0.1.4...v0.1.5