Skip to main content

Blog

Learn About Our Meetup

5000+ Members

MEETUPS

LEARN, CONNECT, SHARE

Join our meetup, learn, connect, share, and get to know your Toronto AI community. 

JOB POSTINGS

INDEED POSTINGS

Browse through the latest deep learning, ai, machine learning postings from Indeed for the GTA.

CONTACT

CONNECT WITH US

Are you looking to sponsor space, be a speaker, or volunteer, feel free to give us a shout.

Category: Reddit MachineLearning

[D] NAS: Has anyone tried yet to search for new basic operations like convolution, pooling?

From Neural Architecture Search: A Survey, first published in 2018:

“Moreover, common search spaces are also based on predefined building blocks, such as different kinds of convolutions and pooling, but do not allow identifying novel building blocks on this level; going beyond this limitation might substantially increase the power of NAS.”

I wonder now if somebody tried that since this was written or if anybody has some thoughts about the feasability of this idea.

submitted by /u/creiser
[link] [comments]

[D] Which problems would you say are solvable today if we just had more data and compute? Using just the theoretical knowledge we have today.

In other words what could we expect to achieve sooner or later if our theoretical work stopped today and we just focused on collecting more data and adding more compute? Are there any types of problems that definitely not wouldn’t be solvable?

If you have any paper discussion this subject, feel free to share them.

submitted by /u/mrconter1
[link] [comments]

[D] Pretraining vs learning from scratch

Hi, I saw this video: https://www.youtube.com/watch?v=AhEVk7TLVjQ explaining the difference and I’m quite confused.

What is the actual basis for deciding whether to use pre-trained models or learn from scratch? My understanding is, if you don’t have resources you have to use pre-trained models. If you have enough resources you can learn from scratch with your own data.

submitted by /u/good_profile
[link] [comments]

[P] ML Buddies for Projects

Hey all,

Not sure if this is the best place to post this, but I am looking to enhance my portfolio and ML skills more importantly through Kaggle competitions and side projects.

About me:

cornell grad with coursework experience in ML, did an internship dealing with NLP and CV applications. now working as a data engineer and doing a project dealing with using nlp learning from unstructured data.

currently following along with cs231 and 230 to improve my deep learning skills. Almost done with the andrew ng

looking for like minded individuals to partake in side projects or kaggle contests with!

submitted by /u/SWEbyday
[link] [comments]

[D] 2 Titan RTX’s or 4 2080Tis?

In terms of performance, which would be better? The price is about the same

With the 2080Tis, I have the overhead of inter gpu communication, and have 44 GB of memory, 17408 cuda cores total, 2176 tensor cores total, 1545 Mhz speed

With the 2 Titans, I dont have as much overhead i suppose, have 48GB, only 9216 cuda cores total, 1152 tensor cores total, 1770 Mhz speed.

Which do I go with. Once again, the price is about the same

submitted by /u/soulslicer0
[link] [comments]

[P] Loss for restricted boltzmann machines

Hello guys, I am trying to implement this paper. They used restricted boltzmann machines to get evolution of a system. Due to nature of the problem “activation” function is not sigmoid but tanh which is later transformed into probability. Different activation function means that gradient is different, although in the paper they wrote “standard” gradient equations. When I tried to train model with standard gradient equations I couldn’t get any meaningful results.I am aware that RBM minimizes KL divergence, but I couldn’t figure out how to implement that (how to get probability function). After trying to find what should I put for loss function when training with eg. Adam I stumbled upon some random github code which for loss function had:

L = energy(data_vec) – energy(genrated_vec)

(yes without absolute value). Well, I plugged that in and it worked like a charm. When I tried to put absolute value over loss it did not worked. I am having trouble understanding why is this correct loss for RBM and I would appreciate help 🙂

submitted by /u/matej1408
[link] [comments]

[D]What is the latest on this Fujitsu Laboratories announcement from Oct?

I found this :

https://www.fujitsu.com/global/about/resources/news/press-releases/2019/1015-01.html

Looks like it was posted about a month ago, I haven’t really found much more info on it, was trying to see if it was anything groundbreaking, the details provided are slim to none. Just gives a basic overview of how neural networks seem to work, and some fancy terms for normalization/transformations/landmark detection. Thanks.

They make some bold claims:

This technology has achieved a high detection accuracy rate of 81% even with limited training data.

This technology is also more accurate than other existing technologies according to certain facial expression recognition technology benchmarks

submitted by /u/FreckledMil
[link] [comments]

[D] Amd or Intel CPU for deep learning server?

Building a Deep Learning home server with 4x 2080 ti blower style GPUs and I am wondering what CPU to get for this machine. The delima is that I am using python Pytorch and Numpy which has a lot of support with Intels MLK packages that sabotage AMD performance. However, I am utilizing the GPUs and CUDA technology for the matrix multiplication and can get away with Numpy Openblas for my environments on the AMD machine. I’m trying to keep this project under $8000 and want the best performance. Also, I am primary using DDPG and DQN networks. What would you do?

Type Item Price
CPU AMD Threadripper 3960X 3.8 GHz 24-Core Processor
CPU Cooler be quiet! Dark Rock Pro TR4 59.5 CFM CPU Cooler
Motherboard ASRock TRX40 Creator ATX sTRX4 Motherboard
Memory G.Skill Ripjaws V Series 16 GB (2 x 8 GB) DDR4-3200 Memory $59.99 @ Newegg
Storage HP EX950 1 TB M.2-2280 NVME Solid State Drive $129.99 @ Newegg
Video Card PNY GeForce RTX 2080 Ti 11 GB Blower Video Card
Case Corsair Air 540 ATX Mid Tower Case $129.98 @ Newegg
Power Supply EVGA SuperNOVA P2 1600 W 80+ Platinum Certified Fully Modular ATX Power Supply
Prices include shipping, taxes, rebates, and discounts
Total (before mail-in rebates) $339.96
Mail-in rebates -$20.00
Total $319.96
Generated by PCPartPicker 2019-11-29 12:22 EST-0500

submitted by /u/thinking_computer
[link] [comments]