Skip to main content

Blog

Learn About Our Meetup

5000+ Members

MEETUPS

LEARN, CONNECT, SHARE

Join our meetup, learn, connect, share, and get to know your Toronto AI community. 

JOB POSTINGS

INDEED POSTINGS

Browse through the latest deep learning, ai, machine learning postings from Indeed for the GTA.

CONTACT

CONNECT WITH US

Are you looking to sponsor space, be a speaker, or volunteer, feel free to give us a shout.

Category: Reddit MachineLearning

[D] Best way to cluster text paragraphs?

My boss wants me to do a hack project where I cluster user feedback / complaints (e.g. people saying “wtf I can’t log in” or “this UI is ugly bla bla” etc.) We have >100k unlabeled data points. There may be jargon in there but it’s mostly legible English. Our goal is to cluster these things so that those talking about the same issue get grouped, and we can take care of them in chunks as nobody wants to read a thousand of these per day.

I’m not an NLP guy by any stretch, so I’ve been reading papers all day to try catching up, however I’m kind of in the middle of the ocean right now. There’s a lot of stuff out there and being inexperienced I thought I’d summon you folks for a discussion on what to try.

My idea now is to use some kind of Transformer model to embed each data point (paragraph) but stuck here as I’m learning that the vectors coming out of those encoders don’t cluster well by text meaning. Let me know any ideas.

P.S. simple models like counting keywords failed me because 1) the data points have a lot of shared vocab so irrelevant things get clustered together, and 2) there are many ways of talking about the same thing with different words.

Ciao

submitted by /u/ME_PhD
[link] [comments]

[D] My experience with Paperspace virtual machines

I was looking for a VM with a GPU to train my model. I was going to use Google Cloud but unfortunately they don’t do business with people from my country so I had to look elsewhere.

That’s when I remembered of paperspace which looked pretty nice. They even have a separate option for ml which allows you send calculations to the cloud and launch notebooks.

But the system wouldn’t accept my card. It simply said “Card is declined”. I reached support and they said that’s probably because their system cannot determine my ip because of VPN or firewall and that I need to turn that off to add card info. Pretty strange thing to ask for IP to simply add payment info but that worked.

I quickly understood that I’m not comfortable with this Gradient service and that I’d like to operate from PyCharm, using vm as a remote interpreter via ssh.

So I tried to rent a regular VM but all the options were locked saying that I need to send a request, describing reasons and ways in which I want to use it. Strange, but I send a request, saying that thing about using PyCharm. Waited a day, no response and sent one more request.

Later that day I get an email from their security staff saying that my account rated highly on their risk matrix and was flagged as suspicious and that I must send them:

  • a photo of my ID with name matching the card
  • contact information
  • company or personal website
  • link to github or social media accounts
  • detailed description of what I’m going to do with the service

And if I don’t do it in 24 hours they will ban me forever.

tl;dr accused me of being suspicious and potentially fraudulent and asked all kinds of personal info to unblock me

Well, imo they should balance their false positive rate and improve customer service greatly.

What are other good alternatives for VMs for machine learning? What do you use?

submitted by /u/Darell1
[link] [comments]

What does the global minima of a non-convex loss function look like?

For LeNet trained on MNIST with the lowest possible loss (global minima),

  • What would the test error rate look like? Is there a benchmark for best possible performance?
  • Can we achieve global minima on non-convex loss functions for a classification task with a minimum number of parameters? Or conversely, how does adding more parameters to a NN help with this?

submitted by /u/liqui_date_me
[link] [comments]

[D] Is there research that focuses primarily on speed of learning, or minimizing required dataset size instead of results?

I see a lot of brilliant techniques that produce incredible results, but are there any techniques that try to bring learning speeds an order of magnitude or two up even if it costs them half the accuracy, or techniques that try to learn from 400 images instead of 40k?

Or in other words, I would love if someone were to link me some research pursuing non-conventional goals.

submitted by /u/derpderp3200
[link] [comments]

[P] Short films scripted by GPT-2

I started a project where I’m going to shoot short stories generated by GPT-2. I insert the first line and it completes it. Of course, there is a lot of filling the gaps as sometimes all we get is dialogue and other times only a setting.

This is my first video and I’ve got a few others already shot and plenty scripts selected.

https://www.youtube.com/watch?v=QfI0Pu0jz3E

I think this is an interesting application of GPT-2 and illustrates visually that although there is coherence in the text, meaning is many times lost in the process.

It would be amazing if in a few years after posting consistently we can see the evolution of AI generated scripts and who knows we can actually have an interesting story which is not pure absurd comedy as they are mostly now.

Hope you enjoy the project and, if you do, please suggest starting lines.

submitted by /u/brunoplak
[link] [comments]

[P] Clustering Pollock

Hi all,

I applied kmeans clustering to some of the Pollock’s paintings. The idea was to track the artist’s usage of #colors through the years. Here’s the outcome!

I had really good fun in mixing computer science and art. I used Python with the standard data science stack (pandas, numpy, scikitlearn) plus opencv. echarts for the visualizations at the end of the article .

Let me know what you think!

https://medium.com/@andrea.ialenti/clustering-pollock-1ec24c9cf447

submitted by /u/travellingsalesman2
[link] [comments]

[D] Tensorflow User Experience

https://nostalgebraist.tumblr.com/post/189464877164/attention-conservation-notice-machine-learning

Originally saw this on Hackernews

https://news.ycombinator.com/item?id=21710863

Are things really this bad? Isn’t the TF 2.0 API cleaning supposed to make Keras the standard API for TPUs? Why doesn’t he use that?

Edit: also, is this an indictment of TF in general or just TPUs?

submitted by /u/justin285
[link] [comments]

[P] How can I make my rendered training data match real data better?

[P] How can I make my rendered training data match real data better?

I’m trying to detect a single type of boxes from a camera image. Instead of using hand labelled images for training, I want to create the data from a 3D model using blender and a python script.

So far I successfully created a dataset and trained RetinaNet on it. I do apply some augmentation (color shifts, saturation changes, noise, blurring, sharpening).

The results on a validation set (consisting of synthetic data too) are great, but the localization performance on real images is way worse.

What changes should I make to my rendering process to match real images better?

Since it’s a virtual environment, I have pretty much unlimited control over everything, but I have no clue what makes sense to try varying. Some of the detections are flawless, but others are way off and I can’t tell what’s the visual difference that throws the network off.

An example for a rendered image (training set)

Excellent results on validation set (halfway hidden boxes are supposed to be not detected)

Localization problems on real images

submitted by /u/Single_Blueberry
[link] [comments]