Skip to main content

Blog

Learn About Our Meetup

5000+ Members

MEETUPS

LEARN, CONNECT, SHARE

Join our meetup, learn, connect, share, and get to know your Toronto AI community. 

JOB POSTINGS

INDEED POSTINGS

Browse through the latest deep learning, ai, machine learning postings from Indeed for the GTA.

CONTACT

CONNECT WITH US

Are you looking to sponsor space, be a speaker, or volunteer, feel free to give us a shout.

[D] Attention layer yields inconsistent results

Hi reddit,

I am currently working on a problem that involves Recurrent Neural Networks. More precisely, I am dealing with sequences of inputs and try to make a prediction at each time step. As such, I decided to try to include an attention layer that is to look on the left context only.

The problem with this is that the results vary depending on the validation data, and by this, I mean a difference of 10-15% accuracy ! I suspect that something is not going so well. I even wonder if the attention layer does not look at both the left and right contexts, which would partly explain why sometimes performances are so good (but I set history_only=True ).

Has it ever happened to you, and what would you do in such situations ? Thank you 🙂

submitted by /u/lazywiing
[link] [comments]