Understanding Representation Learning Paradigms with Applications to Low Resource Text Classification

dc.contributor.authorGarg, Siddhant
dc.date.accessioned2020-06-01T21:14:53Z
dc.date.available2020-06-01T21:14:53Z
dc.date.issued2020-05-21
dc.description.abstractA crucial component of modern machine learning systems is learning input representations which can be used for prediction tasks. The expensive cost of labelling and easy availability of unlabelled data has led to the popularity of representation learning techniques on unlabelled data. In thesis we present two ideas in the domain of representation learning. Firstly, we show that self-supervised representation learning approaches like variational auto-encoders and masked self-supervision can be viewed as imposing a regularization on the representation via a learnable function. We present a discriminative theoretical framework for analysing the underlying assumptions and sample complexities of representation learning via such functional regularizations. Our results show that functional regularization on unlabelled data can prune the hypothesis space and reduce the sample complexity of labelled data. We then consider the domain of NLP where fine-tuning pre-trained sentence embedding models like BERT has become the default transfer learning approach. We propose an alternative transfer learning approach called SimpleTran for low resource text classification characterized by small sized datasets. We train a simple sentence embedding model on the target dataset, combine its output embedding with that of the pre-trained model via concatenation or dimension reduction, and finally train a classifier on the combined embedding either by fixing the embedding model weights or training the classifier and the embedding models end-to-end. With end-to-end training, SimpleTran outperforms fine-tuning on small and medium sized datasets with negligible computational overhead. We provide theoretical analysis for our method, identifying conditions under which it has advantages.en_US
dc.identifier.urihttp://digital.library.wisc.edu/1793/80196
dc.relation.ispartofseriesTR1862
dc.subjectrepresentation learningen_US
dc.subjectself supervised learningen_US
dc.subjectself-supervised learningen_US
dc.subjecttransfer learningen_US
dc.subjectsentence embeddingsen_US
dc.titleUnderstanding Representation Learning Paradigms with Applications to Low Resource Text Classificationen_US
dc.typeTechnical Reporten_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
TR1862 Siddhant Garg.pdf
Size:
2.06 MB
Format:
Adobe Portable Document Format
Description:

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
1.92 KB
Format:
Item-specific license agreed upon to submission
Description: