About this Digital Document
As a means to "demystify" generative models, I will develop a course module that begins with basic (first generation) language models to create simple autocomplete processes, and then move on to more complex implementations, such as RNNs, LSTMs, and transformers. This module will use R and Python, as well as open-source tools like Ollama.
While GUIs for generative models and integration into existing apps is proliferating, many
implementations obfuscate the underlying mechanics of generative technologies. As a result, students generally do not understand how these models work, and are not able to anticipate when and how they are likely to fail. By getting hands-on experience with each advancement in language modeling and demonstrating how such advancements solved problems revealed by previous models, students will get a better understanding of how contemporary generative models operate. Students will also train these various models on a corpus of English science fiction, and thus build bots that write science fiction.