Skip to main content

Blog

Learn About Our Meetup

5000+ Members

MEETUPS

LEARN, CONNECT, SHARE

Join our meetup, learn, connect, share, and get to know your Toronto AI community. 

JOB POSTINGS

INDEED POSTINGS

Browse through the latest deep learning, ai, machine learning postings from Indeed for the GTA.

CONTACT

CONNECT WITH US

Are you looking to sponsor space, be a speaker, or volunteer, feel free to give us a shout.

Category: Reddit MachineLearning

[D] Getting into research teams in large tech companies

Hi all,

I’m currently interviewing for positions as either research scientist or SWE with a couple of big N companies. I am finishing a PhD in an ML related subject, and what I would really like is to find a job where I can do interesting applied research and maybe publish the occasional paper, but in industry rather than academia. However I have no previous experience with the tech industry and so I am flying a bit blind, applying to companies that have large ML teams and hoping to get lucky.

My question is how do you make it into those interesting teams, is it the same process as for generalist new grad roles? As far as the recruiter at Google and FB told me, if I pass the interview and accept the offer I will be matched with a team and can put down some preferences then, but how much leverage do I really have? Do the majority of PhDs just end up on product related teams? Does it depend on the office, what projects are going on at the moment, and if so how easy is it to transfer to research oriented teams later on?

Thanks!

submitted by /u/d73urhi
[link] [comments]

[D] Machine Learning – WAYR (What Are You Reading) – Week 77

This is a place to share machine learning research papers, journals, and articles that you’re reading this week. If it relates to what you’re researching, by all means elaborate and give us your insight, otherwise it could just be an interesting paper you’ve read.

Please try to provide some insight from your understanding and please don’t post things which are present in wiki.

Preferably you should link the arxiv page (not the PDF, you can easily access the PDF from the summary page but not the other way around) or any other pertinent links.

Previous weeks :

1-10 11-20 21-30 31-40 41-50 51-60 61-70 71-80
Week 1 Week 11 Week 21 Week 31 Week 41 Week 51 Week 61 Week 71
Week 2 Week 12 Week 22 Week 32 Week 42 Week 52 Week 62 Week 72
Week 3 Week 13 Week 23 Week 33 Week 43 Week 53 Week 63 Week 73
Week 4 Week 14 Week 24 Week 34 Week 44 Week 54 Week 64 Week 74
Week 5 Week 15 Week 25 Week 35 Week 45 Week 55 Week 65 Week 75
Week 6 Week 16 Week 26 Week 36 Week 46 Week 56 Week 66 Week 76
Week 7 Week 17 Week 27 Week 37 Week 47 Week 57 Week 67
Week 8 Week 18 Week 28 Week 38 Week 48 Week 58 Week 68
Week 9 Week 19 Week 29 Week 39 Week 49 Week 59 Week 69
Week 10 Week 20 Week 30 Week 40 Week 50 Week 60 Week 70

Most upvoted papers two weeks ago:

/u/cafedude: https://arxiv.org/pdf/1911.13299.pdf

/u/nivter: On Mutual Information Maximization for Representation Learning: https://arxiv.org/abs/1907.13625

The authors ran experiments to show that MI maximization between two representations is not directly tied to learning good representations. They did so by maximizing MI while also adversarially training the model to perform badly on linear classification models. One key takeaway for me was that encoders that learn good representations tend to ignore unwanted information as a result of which they are hard to invert (high condition number of Jacobian of output wrt inputs)

Besides that, there are no rules, have fun.

submitted by /u/ML_WAYR_bot
[link] [comments]

[D] Neural network : Applying activity regularization in layer space

Hi all,

I’m trying to build an auto-encoder with an activity regularization that is a function of the output of the layer for a single observation. As i understand it, activity regularization is usually done over the output of a cell for each batch, to promote balanced and sparse activation of each cell. Is that correct?

In this particular case, i want to promote a sparse activation of the layer in such a way that 1/ cells activation influence each others 2/ the mean activation per cell will not be necessarily balanced . I’m aware that if i’m using the L1 norm, the axis over which the sum is done doesn’t matter, so i plan to use a tweaked L(1/2) norm.

Does this make sense, and is there a simple way to do this in Keras?

Thanks

PS : so far i’ve done something in that spirit by alternatively training the encoder on its own modified output (like putting the maximum activation to 1 and the rest to 0, pretty brutal) and the autoencoder on the training data. kind of works, but it’s slow and could be a lot better

submitted by /u/koctogon
[link] [comments]

[P] Use Gaussian Proecess to model maneuvering error

[P] Use Gaussian Proecess to model maneuvering error

Motivation

I try to build some novel application for GP and find a interesting application on maneuvering error modeling. In Napoleonic Wars age, soldiers are lined to fire on their foe. They’re trained to maintain a straight batthleline but that line will not be that perfect straight at all. We can build a generative model to simulate it using GP to help realistic animation in war game.

Why GP?

A i.i.d benchmark can be considered, it will look like this:

i.i.d movement

We can assign time and space correlation to them naturally using GP. A squared exponential can be specifed to denote the fact that more closer two points (X,Y,T) in time and space, the more positive correlation their “error” term hold.

We may expect at same time unit which are close to each other will have consistent deviating tendency, leading a smoother placement instead of a zigzag shape. In other hand, closing to each other at different timestamp having same tendency may imply a “unseen” terrain obstacle, which slow units passing it temporarily.

Result

Take 30 frames:

Result of Gaussian Process

Links

Original post with detailed equations.

https://yiyuezhuo.github.io/blog/tech/2019/12/07/emperor-gaussian-process.html

submitted by /u/yiyuezhuo
[link] [comments]

[Project] Balancing the strengths of a generator and discriminator?

In my DCGAN I have 3 hidden layers for both G and D. Previously I would generate images of 64×64 but decided to double that for 128 x 128. I went ahead and also doubled the inputs and outputs of both D and G.

HOWEVER

On the first epoch the Discriminator became exremely weak and is not able to distinguish any of the generated images from fake. Could of any of you neural net boys guide me to a place where I can find out how to balance them out?

submitted by /u/CasualTrip
[link] [comments]

[D] WeBank & Mila Strategic Partnership

[D] WeBank & Mila Strategic Partnership

What would happen if Bengio or anyone from Mila were to say they support Hong Kong, Taiwan is an independent country or free Uyghurs? TenCent has also proven they’re willing to censor content for the Chinese Communist Party (CCP) and it’s law in China that all companies must have a seat on their board for a CCP member(s).

https://preview.redd.it/2pze3ip61r441.jpg?width=4032&format=pjpg&auto=webp&s=0bdaf241094e59e5edc4a3617075f6bd164c73f0

https://preview.redd.it/7h6h3pp61r441.jpg?width=4032&format=pjpg&auto=webp&s=8d3f1ddaf3d3f3b70c8ee823d925715ffcb51ccb

submitted by /u/unscentedfrowning
[link] [comments]