Issues: ggerganov/llama.cpp
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
./perplexity should allow multiple files, and macro-averaging
enhancement
New feature or request
#7066
opened May 3, 2024 by
turian
4 tasks done
Llama3 GGUF conversion with merged LORA Adapter seems to lose training data randomly
bug-unconfirmed
#7062
opened May 3, 2024 by
Sneakr
llava 1.5 invalid output after first inference (llamacpp server)
bug-unconfirmed
#7060
opened May 3, 2024 by
CaptainOfHacks
Flash Attention not working with NVIDIA Quadro P3200 Pascal Architecture GPU
bug-unconfirmed
#7055
opened May 3, 2024 by
countzero
Non-deterministic output of the llama.cpp server when using multiple slots
bug-unconfirmed
#7052
opened May 3, 2024 by
reuank
CodeQwen returns extra white space for code completion (w/server + fim)
bug-unconfirmed
#7050
opened May 2, 2024 by
wsxiaoys
Different tokenization than AutoTokenizer when word is adjacent to non-special added token
bug-unconfirmed
#7049
opened May 2, 2024 by
JohanAR
Significantly different results (and WRONG) inference when GPU is enabled.
bug-unconfirmed
#7048
opened May 2, 2024 by
phishmaster
Command-R-Plus unable to convert or use after BPE pretokenizer update
bug-unconfirmed
duplicate
This issue or pull request already exists
#7040
opened May 2, 2024 by
sealad886
Server UI: Code snippets are being mangled by <em> italic emphasize replacement.
#7023
opened May 1, 2024 by
jboero
Supporting phi-2 tokenizer
enhancement
New feature or request
#7022
opened May 1, 2024 by
BramVanroy
4 tasks done
server: self context extent broken
bug
Something isn't working
server/webui
#7005
opened Apr 30, 2024 by
phymbert
Pythonic way for quantization
enhancement
New feature or request
#7003
opened Apr 30, 2024 by
MohamedAliRashad
LLamaCpp embedding returns an empty array for long text(While HuggingFaceEmbeddings works fine)
bug-unconfirmed
#6996
opened Apr 30, 2024 by
mokeyish
Segmentation fault on finetune with -ngl > 0, Debian 12 stable
bug-unconfirmed
#6994
opened Apr 30, 2024 by
Basiliotornado
About dialogue training mode
enhancement
New feature or request
#6993
opened Apr 30, 2024 by
andyyumiao
Intel(R) Arc(TM) A770M Setting as default instead of Iris Xe Graphics
bug-unconfirmed
#6991
opened Apr 29, 2024 by
shailesh837
main Segfault using cmake & -march=armv8.4a flag
bug-unconfirmed
#6990
opened Apr 29, 2024 by
Jeximo
Previous Next
ProTip!
Type g p on any issue or pull request to go back to the pull request listing page.