Large Language Model Techniques (1)

Thursday, March 21, 2024 - 15:00

TECHNICAL WORKSHOP SERIES - Large Language Model Techniques (1)
 

Presenter: Zahra Taherikhonakdar
Date: Thursday, March 21, 2024
Time: 3:00 PM
Location: 4th Floor (Workshop space) at 300 Ouellette Avenue (School of Computer Science Advanced Computing Hub)

Abstract:
A language model is a probabilistic representation of a natural language, capable of estimating the likelihood of word
sequences based on the textual data it was trained on, encompassing one or more languages. The initial significant
statistical language model emerged in 1980, followed by IBM's 'Shannon-style' experiments in the same decade.
These experiments involved observing and analyzing human subjects' performance in predicting or correcting text,
leading to the identification of potential avenues for enhancing language modeling.
These models find application in various tasks, such as speech recognition (aiding in avoiding predictions of
improbable or nonsensical sequences), machine translation, natural language generation (producing text that
resembles human language more closely), optical character recognition, handwriting recognition, grammar
induction, and information retrieval.
 
Workshop Outline:
In this workshop you will be introduced to the techniques that are used in LM:
  • Introduction to N grams
  • Estimating N gram Probabilities
  • Evaluation and Perplexity
  • Generalization
     
Prerequisites:
Computer Science knowledge
 
Biography:
Zahra is a PhD student at the University of Windsor. Her research is in the area of Information Retrieval. Particularly
about how to improve query refinement as a technique to make search engines to retrieve the most related documents
based on users' initial query.