Data Science Asked by VM_AI on May 20, 2021
Language model with attention layer is not learning after 20 epochs. Both training and validation loss increase together, while the accuracy flattens at around 7%
The way input data is pipelined is by applying window of length 10 for each sentence so that the model will be able to make inferences on all words of vocab size rather than masking words randomly in each sentence. You can find the code here https://drive.google.com/file/d/1La83LKaZNHsGfCtxKtAWgAqAzLwHo9U5/view?usp=sharing
Orange = train
Blue = validation
Any suggestions on how to get this working would be helpful. Thanks.
Get help from others!
Recent Questions
Recent Answers
© 2024 TransWikia.com. All rights reserved. Sites we Love: PCI Database, UKBizDB, Menu Kuliner, Sharing RPP