/llama3.cuda-ko

llama3.cuda is a pure C/CUDA implementation for Llama 3 model.

Primary LanguageCudaMIT LicenseMIT

llama3.cuda

llama3.cuda

llama3.cuda is a pure C/CUDA implementation for Llama 3 model.

Following up on my previous implementation of the Llama 3 model in pure NumPy, this time I have implemented the Llama 3 model in pure C/CUDA (This repository!).

The Llama model implementation and UTF-8 tokenizer implementation were based on llama2.c previous implemented by Andrej Karpathy, while the CUDA code adopted the kernel implemented by rogerallen. It also heavily referenced the early CUDA kernel implemented by ankan-ban.

The key features are:

  • No dependency
    It's simple, readable, and dependency-free to ensure easy compilation anywhere. Both Makefile and CMake are supported.
  • No C++
    It's a pure C implementation that does not use C++, and most values are treated as pointers.
  • One single file
    Even including a lot of boilerplate code, such as UTF-8 byte sequence processing, It kept the entire code to under 900 lines in a single file.
  • Same result
    To achieve exactly the same results as the NumPy implementation, I debugged the logit values manually to reduce the floating-point arithmetic error rate, and reduced error rate to less than 0.5%.
  • High performance
    While the NumPy implementation on the M2 MacBook Air processed 33 tokens/s, the CUDA version processed 2,823 tokens/s on a NVIDIA 4080 SUPER, which is approximately 85 times faster. This experiment really demonstrated why we should use GPU.

Usage

$ make
$ ./runcuda "I have a dream"
"""
I have a dream. He dreams of a big, beautiful garden full of flowers and trees. He dreams of playing with his friends and eating yummy snacks.
One day, he was walking in the garden when he saw
Token count: 50, elapsed: 0.017000s, 2823 tokens/s
"""

Next steps

I did some patches, including tokenizer, to ensure the same results as the NumPy version, and interestingly, I noticed that Andrej Karpathy commented in the tokenizer code "I don't have the energy to read more of the sentencepiece code to figure out what it's doing". I spent a lot of time fixing this code, but unfortunately, I didn't get good results and had to do it as a messy monkey patch. I'll try again in the future with further refinements.

In the future, I will try to verify it on other platforms besides CUDA by using AMD's ROCm implementation and Intel's oneAPI implementation. Also, there is still an issue that Multi-Head Attention is handled by a single kernel, which has a similar effect to Flash Attention, but it is somewhat inefficient because it performs GEMV operations instead of GEMM operations. In the future, I plan to improve this and implement Flash Attention correctly.

Citing llama3.cuda

If you use or discuss llama3.cuda in your academic research, please cite the project to help spread awareness:

@misc{llama3.cuda,
  title = {llama3.cuda: pure C/CUDA implementation for Llama 3 model},
  author = {Sang Park}, 
  howpublished = {\url{https://github.com/likejazz/llama3.cuda}},
  note = {llama3.cuda, MIT License}
  year = {2024},
}

References

I've adopted most of the code from the authors below:

For more information on implement Llama 3 model, see the following article I wrote:

And title image was generated by DALL-E

License

MIT