-
Notifications
You must be signed in to change notification settings - Fork 3.1k
Issues: karpathy/llm.c
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Speedup
attention_forward_kernel2
by implementing Flash Attention 2 kernel
#60
opened Apr 11, 2024 by
leloykun
Loading…
A CS 231n-style port of this project, implementing LLMs solely with NumPy
#784
opened Nov 18, 2024 by
davidtag
Loading…
[layernorm.cuh] Minor fix with 32 replaced by WARP_SIZE
#544
opened Jun 4, 2024 by
lancerts
Loading…
[attension.cuh] Move assert outside of attn kernel to launcher
#545
opened Jun 4, 2024 by
lancerts
Loading…
Rewrite the encoder_forward float4 kernel with pack128
#302
opened Apr 30, 2024 by
lancerts
Loading…
Use the command 'brew --prefix libomp' to retrieve the location where libomp would be installed on macOS.
#87
opened Apr 12, 2024 by
linmajia
Loading…
[dev/cuda] Added warpsize as a constant expr for dev/cuda files
#438
opened May 20, 2024 by
lancerts
Loading…
[dev/cuda] Include a matmul_backward_bias kernel based on PMPP CoarsenedSumReduction kernel in 10.15
#419
opened May 16, 2024 by
lancerts
Loading…
Add
repkv_backward_kernel2
and repkv_kernel2
(llama3 branch)
#771
opened Sep 28, 2024 by
insop
Loading…
Check for the existence of CUDNN_FRONTEND_PATH before looking in default directories. If it exists, skip additional checks.
#628
opened Jun 21, 2024 by
koparasy
Loading…
only save missing bits to reconstruct fp32 master weights
#432
opened May 19, 2024 by
ngc92
Loading…
feat(attention_forward.cu): Gentle introduction to CuTe(cutlass)
#233
opened Apr 23, 2024 by
FeSens
Loading…
Previous Next
ProTip!
Type g p on any issue or pull request to go back to the pull request listing page.