MuzzammilShah/people-names
Viewer • Updated • 32k • 10
Introduced to the concept of a bigram character-level language model, this repository explores its training, sampling, and evaluation processes. The model evaluation was conducted using the Negative Log Likelihood (NLL) loss to assess its quality.
The model was trained in two distinct ways, both yielding identical results:
This demonstrated that both methods converge to the same result, showcasing their equivalence in achieving the desired outcome.
For a better reading experience and detailed notes, visit my Road to GPT Documentation Site.
Notes and implementations inspired by the Makemore - Part 1 video by Andrej Karpathy.
For more of my projects, visit my Portfolio Site.