A repository sharing the literatures and resources about Large Language Models (LLMs) and beyond.
Hope you find this repository useful! 😊
- 2024.10.24
- Welcome to watch our new online free LLMs intro course on bilibili!
- We also open-source and keep updating the course assignments for you to take a deep dive into LLMs.
- If you like this course or this repository, you can subscribe to the teacher's bilibili account and maybe ⭐ this GitHub repo 😜.
- 2024.03.07
- We offer a comprehensive notebook tutorial on efficient GPU kernel coding using Triton, building upon the official tutorials and extending them with additional hands-on examples, such as the Flash Attention 2 forward/backward kernel.
- In addition, we also provide a step-by-step math derivation of Flash Attention 2, enabling a deeper understanding of its underlying mechanics.
- Tutorials
- Development Stages
- Applications
- Study
- Basics
- Assets
Note:
-
Each markdown file contains collected papers roughly sorted by
published year
in descending order; in other words, newer papers are generally placed at the top. However, this arrangement is not guaranteed to be completely accurate, as thepublished year
may not always be clear. -
The taxonomy is complex and not strictly orthogonal, so don't be surprised if the same paper appears multiple times under different tracks.