Skip to content

ebell495/CSE576_Project2

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

26 Commits
 
 
 
 

Repository files navigation

CSE576_Project2

Colab Notebooks

Pretraining Colab Notebook: https://colab.research.google.com/drive/1XYdqwQoAqZPUa524reehwcOe-zcy93Zy?usp=sharing

Finetuning Colab Notebook: https://colab.research.google.com/drive/1dmb7WYaPWnsYVHtO37DsIQVKCNjAM2Q3?usp=sharing

(Use to test the finetuned models)

Evaluation Colab Notebook: https://colab.research.google.com/drive/1IxyxmCI_fCDB1wRaOisNZOicUkUONdgU?usp=sharing

Saved Models

Pretrained Models: https://drive.google.com/drive/folders/17r6JXdVbzpj9qE8ZPSu_XZQEl86N6_g_?usp=sharing

Finetuned Models: https://drive.google.com/drive/folders/1ECsp-OmDDUrXAKPyo5Eqx4F2T6XDffTO?usp=sharing

Results: https://drive.google.com/drive/folders/12fXNKFBKsxZmmrtORfhBlR2AHKMftNoH?usp=sharing

Masking Methodologies

  • Many-Mask Numeration
    • This will randomly mask one or more of the words in the numeration.
    • Three thousand four hundred fifty-five -> <mask> thousand <mask> <mask> fifty-five
  • Many-Mask Number
    • This will random mask one or more digits in the number
    • 56483 -> 5<mask><mask>83 or 5<mask>4<mask>3
  • Mixed-Mask
    • This will randomly mask one or more digits in the number then mask the complement in the numeration
    • Three thousand four hundred fifty-five is the number 3455 -> <mask> <mask> four hundred fifty-<mask> is the number 3<mask>5<mask>

Finetuning

Bulding off of this paper: https://aclanthology.org/2021.emnlp-main.563.pdf we will try finetuning using an e based representations. This takes the number and labels the powers of tens. For example 8721 becomes 8 e 3 7 e 2 2 e 1 1 e 0. The authors of the paper above find that this representations performs well on extrapolation tasks, but have not tested it with this numeration task.

Results

The finetuning domain is from 0-9999. The out of domain values are from 0-999999. We have pretrained 2 sets of models, one set up to the in domain range, and the other up to the out of domain range. We are interested in seeing if the out of domain results--predicting the number from numeration--performs better based on the pretraining techniques and ranges.

In Domain

In Domain 0-9999 0-999999
T5-Base (Control) 89.656% --
Many-Mask Numeration 90.557% 82.249%
Many-Mask Number 99.566% 99.433%
Mixed-Mask 99.666% 99.833%

Out of Domain

These are the results of testing a pre-trained and fine tuned model on an out of domain range.

Out-Of-Domain 0-9999 0-999999
T5-Base (Control) 12.575% --
Many-Mask Numeration 2.311% 2.888%
Many-Mask Number 2.230% 4.029%
Mixed-Mask 2.409% 4.117%

In Domain E Representation

In Domain 0-9999 0-999999
T5-Base (Control) 88.055 --
Many-Mask Numeration 82.449% 61.862%
Many-Mask Number 99.933% 98.365%
Mixed-Mask 99.766% 99.933%

Out of Domain E Representation

These are the results of testing a pre-trained and fine tuned model on an out of domain range.

Out-Of-Domain 0-9999 0-999999
T5-Base (Control) 0% --
Many-Mask Numeration 0% 0%
Many-Mask Number 0% 0%
Mixed-Mask 0% 0%

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published