Skip to main content

Blog

Learn About Our Meetup

5000+ Members

MEETUPS

LEARN, CONNECT, SHARE

Join our meetup, learn, connect, share, and get to know your Toronto AI community. 

JOB POSTINGS

INDEED POSTINGS

Browse through the latest deep learning, ai, machine learning postings from Indeed for the GTA.

CONTACT

CONNECT WITH US

Are you looking to sponsor space, be a speaker, or volunteer, feel free to give us a shout.

[D] Better or worse to include relational data in training/output?

I am solving for a list of 2D features [x,y].

Do you think it’s better to just solve for just the outputs or should I help guide it by reinforcing with relational data. E.g. the vector between 2 points:

param1 [x1,y1],
param2 [x2,y2],
Vector1 [(x1-x2), (y1-y2)],

=> [x1, y1, x2, y2, xVector1, yVector1]
vs
[x1, y1, x2, y2]

My intuition says it’s better to include the ‘hints’ provided by the vectors. But the cautionary voice in my head says “don’t try to impose on the networks, if it’s valuable, it will eventually discover it itself. ” Am I being a naive beginner in trying to help it with additional clues or am I being paranoid in thinking the training will just magically find the optimal solution on its own?

I’ve trained a model on just points and it’s doing okish with my current dataset. And I’m increasing the size of my dataset right now. But I’m also seeing some “obvious” mistakes like points 1 pixel apart from one another. If it was solving for both I feel like the error rate would more accurately reflect whatever is causing the training to not catch those edge cases.

submitted by /u/im_thatoneguy
[link] [comments]