Skip to main content

Blog

Learn About Our Meetup

5000+ Members

MEETUPS

LEARN, CONNECT, SHARE

Join our meetup, learn, connect, share, and get to know your Toronto AI community. 

JOB POSTINGS

INDEED POSTINGS

Browse through the latest deep learning, ai, machine learning postings from Indeed for the GTA.

CONTACT

CONNECT WITH US

Are you looking to sponsor space, be a speaker, or volunteer, feel free to give us a shout.

Category: Reddit MachineLearning

[Project] Bbox labeler – labeling datasets on mobile

Hi,

I’m working on a project called bbox labeler which will be a react native mobile app for creating bounding boxes for datasets. The functionality will be similar to labelmg with which many of you are familiar.

A demo of creating, moving, and resizing bboxes can be viewed here: https://youtu.be/3S3IgoY3XqA

I expect to have it released in the next couple of weeks.

submitted by /u/ransudz
[link] [comments]

[P] Implementing BERT-model for NER

Hi all,

I will try to be as concise as possible, but here is some background. The subject of my master thesis is ‘dutch named entity recognition using BERT’. This means that I will have to do entity extraction on dutch clinical notes, using google’s BERT model. The problem I have is that I’ve only taken two university programming courses (in python) and because the field of NLP is literally booming, I have a difficult time sketching out a strategic plan towards implementing this model successfully. The following is a list of things I consider doing, and I have no idea which of these are relevant here, or which important things I am potentially missing out on that would be necessary…

  1. Studying the book ‘Hands-On Machine Learning with Scikit-Learn and TensorFlow’ by Aurélien Géron
  2. Following 3 to 4 introductory courses on NLP, TensorFlow, Machine Learning on Datacamp (online learning platform)
  3. Following the Stanford CS224N: NLP with Deep Learning course
  4. Familiarizing myself with Github, trying to implement and play around with the open-source models.
  5. Reading blog posts on NLP
  6. Reading papers on NLP
  7. …?

Feel free to add to this list, or to provide comments on some of the listed elements!

FYI: I have a bachelor in Math (so I don’t expect any difficulties regarding the theoretics of ML)

My current professor doesn’t seem to show great interest in guiding me, so I have to refer to you guys! I would really greatly appreciate your input as I am a little bit lost at the moment to be honest.

Thanks!

submitted by /u/SquareConfidence7
[link] [comments]

[P] Updates to my machine learning 20 questions-style game…

I posted this a few months back and have been working on the engine which seems to be more accurate now. The game uses decision trees and learning to optimise the tolerance for missing data, in the aim of guessing the object in the fewest questions possible. Happy to explain more to anyone who’s interested and all feedback welcome! Thanks!

Try it here: https://incredicat.com

submitted by /u/twm7
[link] [comments]

[Project] Topological analysis of narratives

Hi!

I am a soon to be computational narratology PhD student. I decided to take it upon myself to try to increase public awareness of narratology (Its a cool field, more people should know about it!)

As such, I decided to start a blog series about formally analyzing plot holes and showing how these plot holes become apparent in the topological features of an embedded narrative. This directly correlates with my PhD thesis (creating a DNN to automatically detect plot holes in narratives, and suggest ways to fix them) so I thought I’d be a prime candidate for writing a blog about it!

https://www.louiscastricato.com/post/topology-and-you-what-the-future-of-nlp-has-to-do-with-algebraic-topology

The first entry is more NLP oriented, but I plan to dig into solely computational narratology in the coming entries. I’ll be posting every few months, and each post will be a 5 – 10min light read (They are written for non-experts). This is the first time I have written a blog post, and all things considered I think it came out pretty good!

submitted by /u/FerretDude
[link] [comments]

[Project] TensorFlow Implementation of Graphical Attention RNNs (Cirstea et. al, 2019)

I really enjoyed this paper on graphical attention RNNs. It is basically a clever way to combine a Graph Attention Mechanism (Veličković et al., 2017) with a Diffusion Convolutional RNN (Li et al., 2017). As the authors did not provide an implementation, I decided to create one myself. My implementation can be found here.

Hope it may be of use to somebody. Any feedback would be greatly appreciated.

submitted by /u/Gedanke
[link] [comments]

[D] Accuracy metric in LSTM not considers time offset for multivariate time-series classification?

[D] Accuracy metric in LSTM not considers time offset for multivariate time-series classification?

So this is a kind of complex question, so I hope I formulate it good enough.

I have a human activity detection task that binary classifies if a user does a specific action or not. For me, it is enough if the system detects the action within 3 seconds after it initially happened.

I am using smartphone sensor data with a frequency of 50Hz, which I then combine with a windowing approach with windows of 1sec length and 0.5sec overlap (i.e. I calculate statistics such as `mean` or `std` for each sensor data for a set time of 1 sec, store these in “windows” and overlap these “windows” by 0.5sec).

For LSTM to learn longterm data I use 5 such windows as timesteps (which would represent 3sec of data) and shift each timestep by one window. So the shape of the data fed to the model is:

[13000 instances, 5 timesteps, 21 features]

Now let’s consider the following case of a finished classification of such a model where all of the large squares in the image are labeled as an event, but only some are classified as such:

https://preview.redd.it/bcscy38nvz541.png?width=1976&format=png&auto=webp&s=034919109a9daee0d7dc6bc21ee1851c08dfdda1

As I understand it, LSTM using the `binary_crossentropy` loss function and `accuracy` as a metric in Keras will evaluate the results in a way that the above accuracy would be 2 out of 5 correctly classified instances. However, the accuracy, in this case, should be 100% because my goal is to detect the event within 3 sec, so as long one of these 5 timesteps are labeled as the event I should get 100% accuracy.

So my questions are:

  1. Do I understand the metric correctly and is this a problem for my current goal?
  2. If yes, how could I overcome this evaluation problem?

submitted by /u/rick854
[link] [comments]

[D] Christmas gifts: books on ML

Hi, I’m considering a few books as Christmas gifts, and I’d like your opinion about which one would be the best choice: ofc it’s ok to suggest something outside of this list. Also, since I understand the list is quite long, it’s ok if you don’t have an opinion about all the books in it. Just let me know which one you’d choose and why. Requirements:

Having said that much, let’s get on with the list:

submitted by /u/arkady_red
[link] [comments]

[D] I have an idea that I think is important and possibly commercially viable but I don’t have the time or domain expertise to make it happen

Privacy is becoming more and more important, especially with the availability of facial detection. For people who insist on using Facebook and Instagram, I’ve had the idea to apply adversarial noise to people’s faces to sanitize them before they’re uploaded online. The problem is how to generalize the noise so it’s effective against as many models as possible, and I imagine if that’s even possible that it would distort the image too much to be worth using. Another question I’ve had is are the models even available to test against? I know black box attacks exist, but how effective are they? And how hard is it to generate an effective attack if the weights are changed?

Like I said, I don’t have the expertise to fully follow through on this and I’d love if it actually came to fruition, or maybe the field isn’t quite there yet, but I wanted to put it out there just in case.

submitted by /u/Boozybrain
[link] [comments]