-
Notifications
You must be signed in to change notification settings - Fork 1.5k
Issues: mlc-ai/mlc-llm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug] Llama example inference using Vulkan gives build error
bug
Confirmed bugs
#2977
opened Oct 14, 2024 by
asfarkTii
[Question] Fail to pip install mlc-llm
question
Question about the usage
#2974
opened Oct 13, 2024 by
xiboliyaxiangjiaojun
Error with LLVM Configuration on Windows for GPU Inference in mlc-llm
bug
Confirmed bugs
#2969
opened Oct 9, 2024 by
sreejith-ios
[Question] Why phi3.5v inference image is much longer than HF
question
Question about the usage
#2967
opened Oct 8, 2024 by
Liuuuu54
InternalError when running llava model
question
Question about the usage
#2966
opened Oct 7, 2024 by
plufz
[Bug] ValueError: Unknown model type: rwkv
bug
Confirmed bugs
#2963
opened Oct 7, 2024 by
haohenggang
[Question] Android app related sentence generation
question
Question about the usage
#2958
opened Oct 2, 2024 by
j0h0k0i0m
[Question] How do you abort in the event of a memory pressure warning on iOS?
question
Question about the usage
#2930
opened Sep 22, 2024 by
bryan1anderson
[Question] How should I build for iOS Simulator? I need to be able to compile simulators (even if I don't run the model in them)
question
Question about the usage
#2929
opened Sep 20, 2024 by
bryan1anderson
[Question] If I can run mlc_llm on an arm64 cpu without any gpu devices?
question
Question about the usage
#2927
opened Sep 20, 2024 by
AIarong
[Bug] Own 2B model is crashing with errors on Snapdragon 8 gen 2 Android phones
bug
Confirmed bugs
#2926
opened Sep 20, 2024 by
AspenFPS
[Doc] Flashinfer has supported the sm75 CUDA device. Can MLC_LLM compile and install on 2080Ti with set(USE_FLASHINFER ON)?
documentation
Improvements or additions to documentation
#2925
opened Sep 20, 2024 by
ZanePoe
[Bug] Segmentation fault while building runtime and model libraries for Android
bug
Confirmed bugs
#2922
opened Sep 19, 2024 by
iamlixiao
[Question] How can I access my own models in the MLC-LLM App ???
question
Question about the usage
#2911
opened Sep 17, 2024 by
Ruban-07
[Feature Request] Support for returning log probabilities of both the prompt and the response tokens in the MLC-LLM API, similar to the functionality provided by OpenAI API.
feature request
New feature or request
#2908
opened Sep 17, 2024 by
Notborntodie
[Bug] Speculative decoding small draft doesn't work on macOS
bug
Confirmed bugs
#2907
opened Sep 16, 2024 by
vlbosch
[Feature Request] Move MLCChat/mlc-package-config.json to Android/data folder
feature request
New feature or request
#2905
opened Sep 15, 2024 by
Jhyrachy
[Bug] Crash on iPads and iPhones with less than 4 gb of ram
bug
Confirmed bugs
#2904
opened Sep 14, 2024 by
PabloButron
[Bug] do not running Llama-3-8B-Instruct-q4f16_1-MLC on windows
bug
Confirmed bugs
#2899
opened Sep 12, 2024 by
BlindDeveloper
[Question] Proper way to run nn.Modules for testing
question
Question about the usage
#2896
opened Sep 11, 2024 by
mitchelldehaven
Previous Next
ProTip!
Follow long discussions with comments:>50.