This repository has been archived by the owner on Dec 6, 2024. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 49
Issues: QwenLM/qwen.cpp
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
crash if compliing in debug mode, everything is ok if in release mode
#75
opened Jan 31, 2024 by
feixyz10
请问用qwen.cpp量化后的模型如何使用optimum-benchmark进行性能基准测试,现在参照readme中所述只得到一个build文件夹,不清楚如何进行下一步的测试
#72
opened Jan 18, 2024 by
suyu-zhang
[BUG] Qwen-1.8-Chat,用llama.cpp量化为f16,然后推理回答错乱,请问1.8在llama.cpp还不支持吗?
#69
opened Dec 26, 2023 by
Lyzin
2 tasks done
💡 [Question] - 您好,请教个问题,qwen-cpp BaseStreamer 如何通过std::string 构造一个 BaseStreamer?C++代码少一个构造方式
question
Further information is requested
#62
opened Dec 18, 2023 by
micronetboy
您好,请教个问题,qwen-cpp BaseStreamer 如何通过std::string 构造一个 BaseStreamer?C++代码少一个构造方式
#61
opened Dec 18, 2023 by
micronetboy
💡 [Question] - <title>qwen-cpp 只使用 cpu 和 启用 cpu BLAS 加速, 在都不使用GPU的情况下,速度有多大差别?我测试没有差别
question
Further information is requested
#63
opened Dec 15, 2023 by
micronetboy
💡 [Question] - QwenCPP Python Binding 如何 支持 BLAS CPU 加速
question
Further information is requested
#64
opened Dec 15, 2023 by
micronetboy
💡 [REQUEST] - CPU 的 qwen-cpp 如何封装为一个 http 服务?
question
Further information is requested
#65
opened Dec 14, 2023 by
micronetboy
CUDA error 2 at /home/qwen.cpp/third_party/ggml/src/ggml-cuda.cu:7196: out of memory
#55
opened Dec 8, 2023 by
youngallien
Previous Next
ProTip!
Add no:assignee to see everything that’s not assigned.