Skip to content

simon-benigeri/lstm-language-model

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

73 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Training an LSTM with Dropout on wikitext-2 and NYT covid 19 text.

We implement an LSTM with Dropout Regularization and train it on 2 corpora: wikitext-2 and NYT covid 19.

The model architecture is inspired by the paper "Recurrent Neural Network Regularization" by Zaremba et al. (2014), one of the earliest successful implementations of Dropout Regularization on recurrent neural networks.

Paper: https://arxiv.org/abs/1409.2329
Original code, in Lua and Torch: https://github.com/wojzaremba/lstm

To create the environment, go to the repo directory run the commands:
conda env create -f environment.yml
conda activate lstm_lm

The src directory contains four scripts:

  • lstm.py contains the model described as in the paper.
  • main.py runs the pipeline: initialize dataset -> create model -> train model.
  • datasets.py is used to load data to train the model. We used wikitext-2 and some articles scraped from NYT on covid 19.

The data directory contains the 2 corpora:

  • wiki.train.txt, wiki.valid.txt, wiki.test.txt
  • nyt_covid.train.txt, nyt_covid.valid.txt, nyt_covid.test.txt

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •  

Languages