-
Notifications
You must be signed in to change notification settings - Fork 1.4k
Issues: mlc-ai/mlc-llm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug] error: no template named 'is_base_of_v' in namespace 'std'; did you mean 'is_base_of'?
bug
Confirmed bugs
#2620
opened Jul 3, 2024 by
haohenggang
In Confirmed bugs
tvm_runtime.h
, there are extra spaces in STRINGIFY_MACRO(EXPAND(n1) EXPAND(n2))
.
bug
#2618
opened Jul 3, 2024 by
XuanqiCui
[Feature Request] Support for Qualcomm Snapdragon XElite PCs (arm64 Windows and WSL2-Linux) as target
feature request
New feature or request
#2617
opened Jul 2, 2024 by
Sing-Li
[Bug] Could not find a version that satisfies the requirement during installation of prebuilt packages
bug
Confirmed bugs
#2605
opened Jun 25, 2024 by
SamKG
[Bug] Check failed: (it != type_key2index_.end()) is false: Cannot find type ObjectPath. Did you forget to register the node by TVM_REGISTER_NODE_TYPE
bug
Confirmed bugs
#2602
opened Jun 24, 2024 by
raj-khare
[Bug] fine-tuned model deployed with webllm not working
bug
Confirmed bugs
#2601
opened Jun 24, 2024 by
JLKaretis
[Bug] "Build Android APP from Source" step2: "mlc_llm package" returns error ""path/to/mlc-llm/android/android/mlc4j" does not exist"
bug
Confirmed bugs
#2600
opened Jun 22, 2024 by
mrbrantofgithub
[Question] It takes too much time for the first token to be returned after a reqeust issued.
question
Question about the usage
#2595
opened Jun 19, 2024 by
dkjung
[Question] Can you programmatically clear the kv cache?
question
Question about the usage
#2593
opened Jun 19, 2024 by
0xLienid
[Question] How to use function calling in MLCChat Android app?
question
Question about the usage
#2589
opened Jun 17, 2024 by
wqwz111
[Question] How to use cpp in project
question
Question about the usage
#2588
opened Jun 17, 2024 by
Moxoo
[Question] batchsize of prefill step
question
Question about the usage
#2583
opened Jun 14, 2024 by
Jack-liu1998
[Bug] FP8 quantization accuracy loss with TinyLlama-1.1B-Chat-v1.0
bug
Confirmed bugs
#2579
opened Jun 14, 2024 by
razetime
[Question] multiple gpu seting: Check failed num_running_rsentries <= engine_config_->max_num_sequence (81 vs. 80) :
question
Question about the usage
#2575
opened Jun 13, 2024 by
aaronlyt
[Question] How can I build Android demo app with llava model?
question
Question about the usage
#2572
opened Jun 12, 2024 by
emphasis10
[Bug] Does MLC LLM support function call ?
bug
Confirmed bugs
#2570
opened Jun 12, 2024 by
Stephen888888
[Bug] The performance accuracy of large models is severely lost after quantization on Qwen2-1.5B-Instruct ,please fix it
bug
Confirmed bugs
#2568
opened Jun 12, 2024 by
Stephen888888
[Question] Proper way to use multiple GPUs
question
Question about the usage
#2562
opened Jun 10, 2024 by
0xLienid
Exiting all the time. Android, Redmi Note 13 pro plus [Bug]
bug
Confirmed bugs
#2558
opened Jun 9, 2024 by
condr-at
[Question] Unable to download and compile custom model from Hugging Face using Question about the usage
mlc_llm package
command
question
#2525
opened Jun 6, 2024 by
AbhayGopal
[Bug] chatglm4 mlc_llm shows error "TVMError: Check failed: append_length > 0 (0 vs. 0) : Append with length 0 is not allowed." during mlc_llm chat CLI
bug
Confirmed bugs
#2517
opened Jun 6, 2024 by
lihaofd
[Question] Running mlc_llm into a multi-phase container build
question
Question about the usage
#2512
opened Jun 5, 2024 by
oglok
Previous Next
ProTip!
What’s not been updated in a month: updated:<2024-06-03.