Check a prototype of 2nd version of this platform being developed by cKnowledge.org in collaboration with MLCommons.

Learning Emotion from 100 Observations: Unexpected Robustness of Deep Learning under Strong Data Limitations

lib:12b0853d94434590 (v1.0.0)

Authors: Sven Buechel,João Sedoc,H. Andrew Schwartz,Lyle Ungar
ArXiv: 1810.10949
Document:  PDF  DOI 
Abstract URL: https://arxiv.org/abs/1810.10949v2


One of the major downsides of Deep Learning is its supposed need for vast amounts of training data. As such, these techniques appear ill-suited for NLP areas where annotated data is limited, such as less-resourced languages or emotion analysis, with its many nuanced and hard-to-acquire annotation formats. We conduct a questionnaire study indicating that indeed the vast majority of researchers in emotion analysis deems neural models inferior to traditional machine learning when training data is limited. In stark contrast to those survey results, we provide empirical evidence from three typologically diverse languages that commonly used neural architectures can be trained on surprisingly few observations, outperforming $n$-gram based ridge regression on only 100 data points. Our analysis suggests that high-quality, pre-trained word embeddings are a main factor for achieving those results.

Relevant initiatives  

Related knowledge about this paper Reproduced results (crowd-benchmarking and competitions) Artifact and reproducibility checklists Common formats for research projects and shared artifacts Reproducibility initiatives

Comments  

Please log in to add your comments!
If you notice any inapropriate content that should not be here, please report us as soon as possible and we will try to remove it within 48 hours!