Model 1:
- Using Snapshot for fast ensembling
- GloVe embeddings were then applied with both GRU and LSTM combined with Max Pooling and Attention techniques
- DICE loss combined with Binary Cross Entropy - Loss function
- Trained all these tags as chunks to obtain the top tags within the dataset
- Manual fine tuning of the punctuations was done to filter out what the training had done just to boost the scores
Model 2:
- Extract top n tags based on boxplot to set a base for the model
- Extract TF-IDF Feature vectors which serves as the input to the LR function
- FastText and Word2Vec models are trained on article's text - Input to Bidirectional LSTM
- Calculate using F1 score
Model 3:
- Apply KNN algorithm