Issues: huggingface/candle
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Latest commit on cudarc seems to have broken running the examples codes
#2175
opened May 8, 2024 by
hololite
Incorrect EOS token(s) in meta-llama/Meta-Llama-3-8B-Instruct example
#2164
opened May 6, 2024 by
socathie
Error: Metal error Error while loading function: "Function 'cast_bf16_f16' does not exist" with llama3
#2163
opened May 6, 2024 by
yIllusionSky
Quantized Phi-3 example fails "cannot find llama.attention.head_count in metadata"
#2154
opened May 2, 2024 by
MoonKraken
broadcast_as
error when processing multiple tokens at once in quantized example
#2153
opened May 2, 2024 by
EricLBuehler
Low time effiency when run cnn on mnist-traning only with CPU
#2144
opened Apr 29, 2024 by
Viewer-HX
DriverError(CUDA_ERROR_ILLEGAL_ADDRESS, "an illegal memory access was encountered")
#2131
opened Apr 27, 2024 by
VakeDomen
flash attention does not yield speed gains on llama example
#2069
opened Apr 15, 2024 by
jorgeantonio21
Previous Next
ProTip!
no:milestone will show everything without a milestone.