Please use this identifier to cite or link to this item: https://hdl.handle.net/20.500.12202/8957
Title: From Naive Bayes to transformers: Evolution of natural language processing models for improved text understanding
Authors: Feltenberger, Dave
Cantor, Nissim
Keywords: ChatGPT
natural language processing (NLP)
AI
boilerplate code
Naive Bayes classifier
transformer architecture
attention mechanism
Issue Date: May-2023
Publisher: Yeshiva University
Citation: Cantor, N. (2023, May). From Naive Bayes to transformers: Evolution of natural language processing models for improved text understanding [Undergraduate honors thesis, Yeshiva University].
Series/Report no.: Jay and Jeanie Schottenstein Honors Program;May 2023
Abstract: When ChatGPT was released to the public in late 2022, people quickly came to realize that many things were about to change. While teachers started to get headaches trying to figure out whether their students’ essays were written by AI, software engineers rejoiced, as they boosted their productivity levels by leveraging the new AI to write boilerplate code much more quickly than a human is capable of. How does this technology work, and what makes it so different than what came before it? In this paper, we will explore the history of natural language processing (NLP), starting with the basic, statistics based approach of the Naive Bayes classifier. Then we’ll explore two different types of neural networks, and end with the transformer model that backs GPT and other large language models, doing a deep-dive into the attention mechanism that is the basis of transformer architecture.
Description: Honors thesis / YU only
URI: https://hdl.handle.net/20.500.12202/8957
Appears in Collections:Jay and Jeanie Schottenstein Honors Student Theses

Files in This Item:
File Description SizeFormat 
Nissim Cantor Honors Thesis.pdf
  Restricted Access
677.97 kBAdobe PDFView/Open


This item is licensed under a Creative Commons License Creative Commons