-
Notifications
You must be signed in to change notification settings - Fork 1.5k
Issues: mlc-ai/mlc-llm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Question] Please explain how the VRAM mechanism of MLC-LLM works on AMD ROCm 6.1?
question
Question about the usage
#2955
opened Sep 30, 2024 by
shahizat
[Question] How can I implement pytorch torch.stack, torch.outer equivalent function when define a new model?
question
Question about the usage
#2941
opened Sep 25, 2024 by
zhihuij
[Bug] mlc_llm serve error on Mac M1 (git clone failed with error 128)
bug
Confirmed bugs
#2938
opened Sep 24, 2024 by
pchalasani
[Question] after pip install mlc-llm, No module named 'mlc_llm'
question
Question about the usage
#2932
opened Sep 22, 2024 by
kidhan1234
[Question] How do you abort in the event of a memory pressure warning on iOS?
question
Question about the usage
#2930
opened Sep 22, 2024 by
bryan1anderson
[Question] How should I build for iOS Simulator? I need to be able to compile simulators (even if I don't run the model in them)
question
Question about the usage
#2929
opened Sep 20, 2024 by
bryan1anderson
[Question] If I can run mlc_llm on an arm64 cpu without any gpu devices?
question
Question about the usage
#2927
opened Sep 20, 2024 by
AIarong
[Bug] Own 2B model is crashing with errors on Snapdragon 8 gen 2 Android phones
bug
Confirmed bugs
#2926
opened Sep 20, 2024 by
AspenFPS
[Doc] Flashinfer has supported the sm75 CUDA device. Can MLC_LLM compile and install on 2080Ti with set(USE_FLASHINFER ON)?
documentation
Improvements or additions to documentation
#2925
opened Sep 20, 2024 by
ZanePoe
[Bug] Segmentation fault while building runtime and model libraries for Android
bug
Confirmed bugs
#2922
opened Sep 19, 2024 by
iamlixiao
[Question] How can I access my own models in the MLC-LLM App ???
question
Question about the usage
#2911
opened Sep 17, 2024 by
Ruban-07
[Feature Request] Support for returning log probabilities of both the prompt and the response tokens in the MLC-LLM API, similar to the functionality provided by OpenAI API.
feature request
New feature or request
#2908
opened Sep 17, 2024 by
Notborntodie
[Bug] Speculative decoding small draft doesn't work on macOS
bug
Confirmed bugs
#2907
opened Sep 16, 2024 by
vlbosch
[Feature Request] Move MLCChat/mlc-package-config.json to Android/data folder
feature request
New feature or request
#2905
opened Sep 15, 2024 by
Jhyrachy
[Bug] Crash on iPads and iPhones with less than 4 gb of ram
bug
Confirmed bugs
#2904
opened Sep 14, 2024 by
PabloButron
[Bug] do not running Llama-3-8B-Instruct-q4f16_1-MLC on windows
bug
Confirmed bugs
#2899
opened Sep 12, 2024 by
BlindDeveloper
[Bug] llama2 radix_tree->MatchPrefix fail cause lots of redunction prefilling.
bug
Confirmed bugs
#2897
opened Sep 12, 2024 by
sunzj
[Question] Proper way to run nn.Modules for testing
question
Question about the usage
#2896
opened Sep 11, 2024 by
mitchelldehaven
[Bug] Any Model with the Suffix _1 Crashes Android
bug
Confirmed bugs
#2894
opened Sep 11, 2024 by
Melgark
[Bug report] running on wsl2 also on windows?
question
Question about the usage
#2893
opened Sep 11, 2024 by
BlindDeveloper
Previous Next
ProTip!
What’s not been updated in a month: updated:<2024-08-30.