You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In [*Build a Large Language Model (from Scratch)*](http://mng.bz/orYv), you'll discover how LLMs work from the inside out. In this book, I'll guide you step by step through creating your own LLM, explaining each stage with clear text, diagrams, and examples.
12
+
In [*Build a Large Language Model (From Scratch)*](http://mng.bz/orYv), you'll discover how LLMs work from the inside out. In this book, I'll guide you step by step through creating your own LLM, explaining each stage with clear text, diagrams, and examples.
13
13
14
14
The method described in this book for training and developing your own small-but-functional model for educational purposes mirrors the approach used in creating large-scale foundational models such as those behind ChatGPT.
15
15
@@ -31,21 +31,20 @@ Alternatively, you can view this and other files on GitHub at [https://github.co
31
31
<br>
32
32
<br>
33
33
34
-
| Chapter Title | Main Code (for quick access) | All Code + Supplementary |
| Ch 1: Understanding Large Language Models | No code | No code |
37
-
| Ch 2: Working with Text Data | - [ch02.ipynb](ch02/01_main-chapter-code/ch02.ipynb)<br/>- [dataloader.ipynb](ch02/01_main-chapter-code/dataloader.ipynb) (summary)<br/>- [exercise-solutions.ipynb](ch02/01_main-chapter-code/exercise-solutions.ipynb)|[./ch02](./ch02)|
| Ch 4: Implementing a GPT Model from Scratch | - [ch04.ipynb](ch04/01_main-chapter-code/ch04.ipynb)<br/>- [gpt.py](ch04/01_main-chapter-code/gpt.py) (summary)<br/>- [exercise-solutions.ipynb](ch04/01_main-chapter-code/exercise-solutions.ipynb)|[./ch04](./ch04)|
40
-
| Ch 5: Pretraining on Unlabeled Data | Q1 2024 | ... |
41
-
| Ch 6: Finetuning for Text Classification | Q2 2024 | ... |
42
-
| Ch 7: Finetuning with Human Feedback | Q2 2024 | ... |
43
-
| Ch 8: Using Large Language Models in Practice | Q2/3 2024 | ... |
| Ch 1: Understanding Large Language Models | No code | - |
37
+
| Ch 2: Working with Text Data | - [ch02.ipynb](ch02/01_main-chapter-code/ch02.ipynb)<br/>- [dataloader.ipynb](ch02/01_main-chapter-code/dataloader.ipynb) (summary)<br/>- [exercise-solutions.ipynb](ch02/01_main-chapter-code/exercise-solutions.ipynb)|[./ch02](./ch02)|
| Ch 4: Implementing a GPT Model from Scratch | - [ch04.ipynb](ch04/01_main-chapter-code/ch04.ipynb)<br/>- [gpt.py](ch04/01_main-chapter-code/gpt.py) (summary)<br/>- [exercise-solutions.ipynb](ch04/01_main-chapter-code/exercise-solutions.ipynb)|[./ch04](./ch04)|
40
+
| Ch 5: Pretraining on Unlabeled Data | Q1 2024 | ... |
41
+
| Ch 6: Finetuning for Text Classification | Q2 2024 | ... |
42
+
| Ch 7: Finetuning with Human Feedback | Q2 2024 | ... |
43
+
| Ch 8: Using Large Language Models in Practice | Q2/3 2024 | ... |
| Appendix B: References and Further Reading | No code | - |
46
+
| Appendix C: Exercises | No code | - |
47
+
| Appendix D: Adding Bells and Whistles to the Training Loop | - [appendix-D.ipynb](appendix-D/01_main-chapter-code/appendix-D.ipynb)|[./appendix-D](./appendix-D)|
0 commit comments