-
Notifications
You must be signed in to change notification settings - Fork 385
Issues: SJTU-IPADS/PowerInfer
Meta: Implementing hybrid inference across key desktop platforms
#92
opened Dec 27, 2023 by
hodlen
Open
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Source for v2 (mobile inference engine)
question
Further information is requested
#194
opened Jun 12, 2024 by
peeteeman
Need quite a long time to load the model
question
Further information is requested
#188
opened May 21, 2024 by
meicale
Will this work with Falcon 2?
question
Further information is requested
#186
opened May 14, 2024 by
aaronrmm
关于在A100显卡上测得的效果异常的疑问
question
Further information is requested
#184
opened May 4, 2024 by
bulaikexiansheng
在A100-80G上无法找到cuda的情况
question
Further information is requested
#182
opened Apr 24, 2024 by
bulaikexiansheng
Where is the definition or addition location of GGML_USE_HYBRID_THREADING?
question
Further information is requested
#172
opened Mar 25, 2024 by
wfloveiu
two questions that i want to solve
question
Further information is requested
#167
opened Mar 18, 2024 by
yeptttt
Will we have instruct fine-tuned model support in the future?
question
Further information is requested
#164
opened Mar 13, 2024 by
ZeonfaiHo
3 tasks done
[Question]: High PPL on wikitext2 of ReLU-LLAMA-7B for language modeling tasks
question
Further information is requested
#162
opened Mar 11, 2024 by
llCurious
3 tasks done
24GB的显存只能占用12GB,CUDA占用也不到10%。但是CPU占用100%内存占用35GB
question
Further information is requested
#159
opened Mar 3, 2024 by
NerounCstate
[ROCm] Is AMD ROCm support available in near future?
enhancement
New feature or request
#158
opened Feb 29, 2024 by
Orion-zhen
3 tasks done
The CUDA compiler identification is unknown And PowerInfer was compiled without cuBLAS
question
Further information is requested
#157
opened Feb 29, 2024 by
LHQUer
关于LLaMA-70B-PowerInfer-GGUF的chat版本
question
Further information is requested
#143
opened Feb 6, 2024 by
NerounCstate
possible to do one that can fit into 7GB vram?
question
Further information is requested
#141
opened Feb 4, 2024 by
sprappcom
Will using only CPU be faster than llama.cpp?
question
Further information is requested
#140
opened Feb 2, 2024 by
liutt1312
How to understand the codes of llama.cpp?
question
Further information is requested
#130
opened Jan 24, 2024 by
BHbean
not enough space in the buffer with long prompts
bug-unconfirmed
Unconfirmed bugs
#129
opened Jan 23, 2024 by
RachelShalom
3 tasks done
how to show the outputs result on a web-service? or how can i get the result of inferrence for other application?
question
Further information is requested
#126
opened Jan 22, 2024 by
xujiangyu
2 of 3 tasks
Previous Next
ProTip!
Add no:assignee to see everything that’s not assigned.