Skip to content

ChatGPT-like LLM in PyTorch from scratch, step by step

License

Notifications You must be signed in to change notification settings

quantumKaz/llm-scratch

Repository files navigation

Build a Large Language Model (From Scratch)

This repository contains code for developing, pretraining, and finetuning a GPT-like LLM.

Bonus Material

Several folders contain optional materials as a bonus for interested readers:

Questions, Feedback, and Contributing to This Repository

I welcome all sorts of feedback, best shared via GitHub Discussions. Likewise, if you have any questions or just want to bounce ideas off others, please don't hesitate to post these in the forum as well.

Please note that since this repository contains the code corresponding to a print book, I currently cannot accept contributions that would extend the contents of the main chapter code, as it would introduce deviations from the physical book. Keeping it consistent helps ensure a smooth experience for everyone.

Citation

If you find this book or code useful for your research, please consider citing it.

Chicago-style citation:

Raschka, Sebastian. Build A Large Language Model (From Scratch). Manning, 2024. ISBN: 978-1633437166.

BibTeX entry:

@book{build-llms-from-scratch-book,
  author       = {Sebastian Raschka},
  title        = {Build A Large Language Model (From Scratch)},
  publisher    = {Manning},
  year         = {2024},
  isbn         = {978-1633437166},
  url          = {https://www.manning.com/books/build-a-large-language-model-from-scratch},
  github       = {https://github.com/rasbt/LLMs-from-scratch}
}

About

ChatGPT-like LLM in PyTorch from scratch, step by step

Topics

Resources

License

Stars

Watchers

Forks

Packages

No packages published