Skip to main content

Blog

Learn About Our Meetup

5000+ Members

MEETUPS

LEARN, CONNECT, SHARE

Join our meetup, learn, connect, share, and get to know your Toronto AI community. 

JOB POSTINGS

INDEED POSTINGS

Browse through the latest deep learning, ai, machine learning postings from Indeed for the GTA.

CONTACT

CONNECT WITH US

Are you looking to sponsor space, be a speaker, or volunteer, feel free to give us a shout.

Category: Reddit MachineLearning

[R] MetaInit: Initializing learning by learning to initialize

Abstract

Deep learning models frequently trade handcrafted features for deep features learned with much less human intervention using gradient descent. While this paradigm has been enormously successful, deep networks are often difficult to train and performance can depend crucially on the initial choice of parameters. In this work, we introduce an algorithm called MetaInit as a step towards automating the search for good initializations using meta-learning. Our approach is based on a hypothesis that good initializations make gradient descent easier by starting in regions that look locally linear with minimal second order effects. We formalize this notion via a quantity that we call the gradient quotient, which can be computed with any architecture or dataset. MetaInit minimizes this quantity efficiently by using gradient descent to tune the norms of the initial weight matrices. We conduct experiments on plain and residual networks and show that the algorithm can automatically recover from a class of bad initializations. MetaInit allows us to train networks and achieve performance competitive with the state-of-the-art without batch normalization or residual connections. In particular, we find that this approach outperforms normalization for networks without skip connections on CIFAR-10 and can scale to Resnet-50 models on Imagenet.

https://papers.nips.cc/paper/9427-metainit-initializing-learning-by-learning-to-initialize

submitted by /u/hardmaru
[link] [comments]

[D] Is Google Colab the future for reproducible research?

Colab and similar products can help with the reproducibility of code and more importantly the code underlying academic results. These free resources not only make code transparency easier, from here forward, it makes unpublished Python code highly suspect. There are no limitations to sharing code and data anymore and no limitation in accessing this code, the data and the necessary processing power to analyse the results.

from medium

submitted by /u/OppositeMidnight
[link] [comments]

[D] Current state of the Topic Segmentation problem

Recently, I did a little research in the literature for “Topic segmentation” since “Text segmentation” seems to be more related to identifying text in images. From the results, it appears that the most recent survey is from 2011 [1], while the most recent papers in big conferences are from 2008 to 2013 [2, 3, 4].

Is this the current state of the problem, or there are more recent and relevant works?

It’s also possible that I’m using the wrong terms. So, for clarification, I’m most interest in segmenting a collection of documents in a small and well-known number of sections / topics.

[1] Purver, Matthew. “Topic segmentation.” In Spoken language understanding: systems for extracting semantic information from speech (2011)

[2] Eisenstein, Jacob, and Regina Barzilay. “Bayesian unsupervised topic segmentation.” In Proceedings of the Conference on Empirical Methods in Natural Language Processing (2008)

[3] Riedl, Martin, and Chris Biemann. “TopicTiling: a text segmentation algorithm based on LDA.” In Proceedings of ACL 2012 Student Research Workshop (2012)

[4] Du, Lan, Wray Buntine, and Mark Johnson. “Topic segmentation with a structured topic model.” In Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics (2013)

submitted by /u/Daango_
[link] [comments]

[R] We released our Oktoberfest Food Dataset

[R] We released our Oktoberfest Food Dataset

Data sample example

Abstract:
” We release a realistic, diverse, and challenging dataset for object detection on images. The data was recorded at a beer tent in Germany and consists of 15 different categories of food and drink items. We created more than 2,500 object annotations by hand for 1,110 images captured by a video camera above the checkout. We further make available the remaining 600GB of (unlabeled) data containing days of footage. Additionally, we provide our trained models as a benchmark. Possible applications include automated checkout systems which could significantly speed up the process. “

Arxiv link

git

If you have any feedback or comments on our work, we are more than happy to hear that.

submitted by /u/comp_vision_
[link] [comments]

[R] C-Flow: Conditional Generative Flow Models for Images and 3D Point Clouds

[R] C-Flow: Conditional Generative Flow Models for Images and 3D Point Clouds

https://preview.redd.it/x1qxpi3p76541.jpg?width=736&format=pjpg&auto=webp&s=7761c84809112c8bda57e24cbbc9810c2870a84d

We introduce a novel conditioning scheme that brings normalizing flows to an entirely new scenario for multi-modal data modeling. We demonstrate our conditioning method to be very adaptable, being applicable to image manipulation, style transfer and multi-modal mapping in a diversity of domains, including RGB images, 3D point clouds, segmentation maps, and edge masks.

Paper PDF: https://arxiv.org/pdf/1912.07009.pdf

Project Website: https://www.albertpumarola.com/research/C-Flow/index.html

submitted by /u/APumarola
[link] [comments]

[P] For NLP researchers, Easy-to-use Text Preprocessing Package, PreNLP

Do very simple text-preprocessing (a.k.a dirty work) with PreNLP Package !

I’m working in NLP part, and implementing a package to do iterative but necessary works for NLP. Here are some exmaples to preprocess text for following NLP tasks.

  • Frequently used normalization functions for text pre-processing are provided in prenlp. General use cases are as follows:

>>> from prenlp.data import Normalizer >>> normalizer = Normalizer() >>> normalizer.normalize('Visit this link for more details: https://github.com/') Visit this link for more details: [URL] >>> normalizer.normalize('Use HTML with the desired attributes: <img src="cat.jpg" height="100" />') Use HTML with the desired attributes: [TAG] >>> normalizer.normalize('Hello 🤩, I love you 💓 !') Hello [EMOJI], I love you [EMOJI] ! >>> normalizer.normalize('Contact me at lyeoni.g@gmail.com') Contact me at [EMAIL] >>> normalizer.normalize('Call +82 10-1234-5678') Call [TEL] 
  • Quick tour to Text classification : The following example code trains fastText classification model on IMDB. The code below has only 16 lines of code (except blank lines and comments).

import fasttext import prenlp from prenlp.data import Normalizer from prenlp.tokenizer import NLTKMosesTokenizer # Data Preparation imdb_train, imdb_test = prenlp.data.IMDB() # Preprocessing tokenizer = NLTKMosesTokenizer() normalizer = Normalizer(url_repl=' ', tag_repl=' ', emoji_repl=' ', email_repl=' ', tel_repl=' ') for dataset in [imdb_train, imdb_test]: for i, (text, label) in enumerate(dataset): dataset[i][0] = ' '.join(tokenizer(normalizer.normalize(text.strip()))) # both prenlp.data.fasttext_transform(imdb_train, 'imdb.train') prenlp.data.fasttext_transform(imdb_test, 'imdb.test') # Train model = fasttext.train_supervised(input='imdb.train', epoch=20) # Evaluate print(model.test('imdb.train')) print(model.test('imdb.test')) # Inference print(model.predict(imdb_test[0][0])) 

For more details, follows: https://github.com/lyeoni/prenlp

p.s. And I want to know what you want to implement on this issue. I’ll implement that on this package.

submitted by /u/lyeoni
[link] [comments]