Blog

Learn About Our Meetup

5000+ Members

MEETUPS

LEARN, CONNECT, SHARE

Join our meetup, learn, connect, share, and get to know your Toronto AI community. 

JOB POSTINGS

INDEED POSTINGS

Browse through the latest deep learning, ai, machine learning postings from Indeed for the GTA.

CONTACT

CONNECT WITH US

Are you looking to sponsor space, be a speaker, or volunteer, feel free to give us a shout.

[D] Neural Differential Equations

I had a question about these. I know that you calculate the whole network in one go and then you just evaluate it at some points along the the depth. However, I was wondering how the parameters work.

1) How are the weights and biases updated? I know they are “shared” through the whole network (and hence less parameters than the usual network) however, how do the individual evaluations work then? For the network. Like say the network is defined from t = 0 to t = 5, and I evaluate at t = 1 and t = 2; are the weights the same here and the only thing that changes is t? And if so, what’s the point even? Why not evaluate just at the end point (i.e. the maximum depth you want) ?

2) Going off of that, what is the point of those in between evaluations if the parameters are shared anyway? Wouldn’t they be updated the same way every time? Or is it that.. multiple evaluations means that the derivatives and the updates are “better”?

I’m just really confused about this whole shared parameters thing. Please help!

submitted by /u/bthi
[link] [comments]

Next Meetup

 

Days
:
Hours
:
Minutes
:
Seconds

 

Plug yourself into AI and don't miss a beat

 


Toronto AI is a social and collaborative hub to unite AI innovators of Toronto and surrounding areas. We explore AI technologies in digital art and music, healthcare, marketing, fintech, vr, robotics and more. Toronto AI was founded by Dave MacDonald and Patrick O'Mara.