Data Science Asked by Qwertiops on September 29, 2020
I’m trying to use the Continuous Bag Of Words method for word embedding on a corpus of 7503 tweets.
In particular, I’m trying to use CBOW on this Kaggle competition, which involves classifying tweets depending on whether they refer to disasters.
I followed the instructions in this article, and then trained a linear classifier on the average vector of the words in a tweet. The classifier performs very poorly, even on the training set (it only classifies less than 60% of tweets correctly).
There are two reasons I think this might be happening:
I can’t find any accessible literature online about best practices for either of these. In particular, I have no idea how much data is usually needed for training (since the neural network is so shallow, I suspect that the usual rules don’t apply), or how much training is necessary.
Any help would be very much appreciated!
Get help from others!
Recent Questions
Recent Answers
© 2024 TransWikia.com. All rights reserved. Sites we Love: PCI Database, UKBizDB, Menu Kuliner, Sharing RPP