Skip to main content

Blog

Learn About Our Meetup

5000+ Members

MEETUPS

LEARN, CONNECT, SHARE

Join our meetup, learn, connect, share, and get to know your Toronto AI community. 

JOB POSTINGS

INDEED POSTINGS

Browse through the latest deep learning, ai, machine learning postings from Indeed for the GTA.

CONTACT

CONNECT WITH US

Are you looking to sponsor space, be a speaker, or volunteer, feel free to give us a shout.

Category: Reddit MachineLearning

[Discussion] Understanding Subscale WaveRNN & usage of Masked Dilated CNN as conditioning network

Related Paper: Efficient Neural Audio Synthesis

I have been reading the sections relating to Subscale WaveRNN in where the DeepMind team was able to generate B samples in a single step. They have discussed about conditioning a particular sample using past samples and up to F samples from the previous sub-tensors future context. In their case, they used a masked dilated CNN (this can be found on the last paragraph of 4.1 Subscale Dependency Scheme). Here’s the excerpt specifically to this:

The Subscale WaveRNN that generates a given sub-tensor is conditioned on the future context of previous sub-tensors using a masked dilated CNN with relus and the mask applied over past connections instead of future ones.

My first question is: how could a masked dilated CNN help with this?

Next, Nal Kalchbrenner has tweeted this quick demo of the Subscale WaveRNN. This one confuses me a lot when I’m referring back to the original paper.

My final question is: does anyone have taken a look at subscaling more closely?

Any insights would be appreciated.

(Note: This is my first post and I am hoping that I followed the format correctly.)

submitted by /u/bigbawsboy
[link] [comments]

[P] cleanlab: accelerating ML and deep learning research with noisy labels

[P] cleanlab: accelerating ML and deep learning research with noisy labels

Hey folks. Today I’ve officially released the cleanlab Python package, after working out the kinks for three years or so. It’s the first standard framework for accelerating ML and deep learning research and software for datasets with label errors. cleanlab has some neat features:

  1. If you have model outputs already (predicted probabilities for your dataset), you can find label errors in one line of code. If you don’t have model outputs, its two lines of code.
  2. If you’re a researcher dealing with datasets with label errors, cleanlab will compute the uncertainty estimation statistics for you (noisy channel, latent prior of true labels, joint distribution of noisy and true labels, etc.)
  3. Training a model (learning with noisy labels) is 3 lines of code.
  4. cleanlab is full of examples — how to find label errors in ImageNet, MNIST, learning with noisy labels, etc.

Full cleanlab announcement and documentation here: [LINK]

GitHub: https://github.com/cgnorthcutt/cleanlab/ PyPI: https://pypi.org/project/cleanlab/

As an example, here is how you can find label errors in a dataset with PyTorch, TensorFlow, scikit-learn, MXNet, FastText, or other framework in 1 line of code.

# Compute psx (n x m matrix of predicted probabilities)# in your favorite framework on your own first, with any classifier.# Be sure to compute psx in an out-of-sample way (e.g. cross-validation)# Label errors are ordered by likelihood of being an error.# First index in the output list is the most likely error.

from cleanlab.pruning import get_noise_indices

ordered_label_errors = get_noise_indices(s=numpy_array_of_noisy_labels,psx=numpy_array_of_predicted_probabilities,sorted_index_method='normalized_margin', # Orders label errors)

cleanlab logo and my cheesy attempt at a slogan.

P.S. If you happen to work at Google, cleanlab is incorporated in the internal code base (as of July 2019).P.P.S. I don’t work there, so you’re on your own if Google’s version strays from the open-source version.

submitted by /u/cgnorthcutt
[link] [comments]

[N] China forced the organizers of the International Conference on Computer Vision (ICCV) in South Korea to change Taiwan’s status from a “nation” to a “region” in a set of slides.

Link: http://www.taipeitimes.com/News/front/archives/2019/11/02/2003725093

The Ministry of Foreign Affairs yesterday protested after China forced the organizers of the International Conference on Computer Vision (ICCV) in South Korea to change Taiwan’s status from a “nation” to a “region” in a set of slides.

At the opening of the conference, which took place at the COEX Convention and Exhibition Center in Seoul from Tuesday to yesterday, the organizers released a set of introductory slides containing graphics showing the numbers of publications or attendees per nation, including Taiwan.

However, the titles on the slides were later changed to “per country/region,” because of a complaint filed by a Chinese participant.

“Taiwan is wrongly listed as a country. I think this may be because the person making this chart is not familiar with the history of Taiwan,” the Chinese participant wrote in a letter titled “A mistake at the opening ceremony of ICCV 2019,” which was published on Chinese social media under the name Cen Feng (岑峰), who is a cofounder of leiphone.com.

The ministry yesterday said that China’s behavior was contemptible and it would not change the fact that Taiwan does not belong to China.

Beijing using political pressure to intervene in an academic event shows its dictatorial nature and that to China, politics outweigh everything else, ministry spokeswoman Joanne Ou (歐江安) said in a statement.

The ministry has instructed its New York office to express its concern to the headquarters of the Institute of Electrical and Electronics Engineers, which cosponsored the conference, asking it not to cave in to Chinese pressure and improperly list Taiwan as part of China’s territory, she said.

Beijing has to forcefully tout its “one China” principle in the global community because it is already generally accepted that Taiwan is not part of China, she added.

As China attempts to force other nations to accept its “one China” principle and sabotage academic freedom, Taiwan hopes that nations that share its freedoms and democratic values can work together to curb Beijing’s aggression, she added.

submitted by /u/Only_Assist
[link] [comments]

[D] What are some interesting questions we can answer using AI?

Hi everyone,

I have created a simple concept: “Train an AI to answer everyday questions and upload the simulations online”

I already created the first project/simulation a couple of months ago, but I’m now looking for the next questions we could answer using AI. I have a few ideas like:

“A.I. Learns: Is it better to walk or run in the rain?”

“A.I. Learns: How to make a profit from cryptocurrency”

etc…..

Topics could be serious or lighter and funny sometimes 🙂

So, what are some interesting questions we can answer using AI?

I’ll pick up interesting answers for my next project.

Thank you.

submitted by /u/mrsailor23
[link] [comments]

[P] WindHive: ML based coding assistant to boost coding productivity

Hi r/MachineLearning, we were tired of constantly having to search Google, StackOverflow, and GitHub for code examples and API documentation when writing code. We built WindHive.ai, a smart coding assistant that provides you contextually relevant code snippets and docs directly from your editor!

WindHive does this by using machine-learned code representations/embeddings. We have trained neural networks on hundreds of publicly available code repositories to create embeddings for the task of deciding which snippets of code would be most useful to show you at any given time. Using these embeddings, we are able to index and search through tens of thousands of code snippets and show you exactly what you need.

We believe WindHive can help you increase your programming productivity and avoid reinventing the wheel! Please to to windhive.ai to find out more and share your feedback with us!

submitted by /u/crystal_alpine
[link] [comments]

[D] An Open Source Stack for Managing and Deploying ML Models – DVC & Cortex – Tutorial

https://towardsdatascience.com/an-open-source-stack-for-managing-and-deploying-models-c5d3b98160bc

In this tutorial, we’re going to use DVC to create a model capable of analyzing StackOverflow posts, and recognizing which ones are about Python. We are then going to deploy our model as a web API, ready to form the backend of a piece of production software.

DVC stores your model weights and training data in a centralized location, allowing collaborators to get started easily, while also tracking changes and ensuring an accurate version history.

As a final step in this tutorial, we’re going to integrate DVC with another open source tool—Cortex—that allows us to deploy DVC-generated models as web APIs, ready for production.

submitted by /u/thumbsdrivesmecrazy
[link] [comments]