Skip to main content

Blog

Learn About Our Meetup

5000+ Members

MEETUPS

LEARN, CONNECT, SHARE

Join our meetup, learn, connect, share, and get to know your Toronto AI community. 

JOB POSTINGS

INDEED POSTINGS

Browse through the latest deep learning, ai, machine learning postings from Indeed for the GTA.

CONTACT

CONNECT WITH US

Are you looking to sponsor space, be a speaker, or volunteer, feel free to give us a shout.

Category: Reddit MachineLearning

[D] Best approach to fine-classification of image objects?

I’m basically looking to make a classifier that gives a binary (yes/no) response to an image’s subject matching my model. For example let’s say I wanted it to match copperhead snakes and only copperhead snakes. Is there a good cloud solution for this?

I tried AWS Rekognition but it wanted a minimum of 2 labels and I felt the results wouldn’t be good enough after doing a test run with ~120 images. I know that’s relatively small but don’t want to waste time/money with a cloud solution if it won’t get me there.

Will Google Vision have better results? Would I need to roll my own solution in Python? Or is it unrealistic for a novice solution to tell one type of snake from another?

submitted by /u/TRAINS_CHOOCHOO
[link] [comments]

[D] Concerns about “Face Beautification: Beyond Makeup Transfer”

I came across the paper “Face Beautification: Beyond Makeup Transfer” and was appalled at the poor ethical and scientific practice shown by the paper. I emailed the PC and the D&I chairs, but I wanted to share my critique with the community as well:

I came across the paper “Face Beautification: Beyond Makeup Transfer” that was published at NeurIPS this year. I was deeply concerned by the apparent complete lack of care for the social and ethical repercussions of the paper. The goal of the paper is to change photos of women to make them more attractive. While it may be possible to do this in a way that isn’t objectionable, the paper there is zero discussion of or acknowledgement of the social, political, and power-dynamical (is that a word?) aspects of what is judged as attractive. The paper also contains serious methodological issues and blatantly contradicts itself in a fashion that I would expect to disqualify the paper from publication in the first place.

The examples in the paper make it clear that the algorithm’s concept of “attractive” is “light skinned white people.” Of the 114 demo examples of computer-generated “attractive people” in the paper, 100% are white. Not only that, almost all of them have extremely light skin. Only a couple of the shown inputs appear to be non-white people (e.g., Table 2 appears to contain a South Asian woman), and the algorithm clearly makes them into white people both by lightening their skin and by changing other morphological features to make the person appear more white. None of the inputs appear to be black people. Even among white people it strongly prefers people with lighter skin; there are zero examples where the algorithm appears to darken the skin tone of the person to be beautified and in the majority of cases it is significantly lightened.

This isn’t just an issue of using white people as “attractive references,” as it even happens when the reference attractive image is a photo of a non-white person, as seen in the table at the top of the first page. Two east Asian people are used as reference images to beautify white people, but the resulting image has typically white features such as a less ovular face shape and doubled eyelids.

Not only does this appear as a persistent pattern in the images, the authors don’t even mention that it happens, let alone critically engage with this. Given how much NeurIPS appears to pride itself on social awareness in AI research, I am saddened and disheartened to see that this paper was viewed as having sufficient merit and ethical practice to warrant publication.

Another major issue is the very last paragraph of their paper. It says

>Personalized beautification is expected to attract increasingly more attention in the incoming years. This work we have only focused on the beautification of female Caucasian faces. A similar question can be studied for other populations even though the relationship between gender, race, cultural background and the perception of facial attractiveness has remained under-researched in the literature. How can AI help reshape the practice of personal makeup and plastic surgery is an emerging field for future research.

This paragraph is clearly false for several reasons. As I mentioned, they have reference photos of non-Caucasian people in the paper itself and appear to input at least a couple non-Caucasian people. Secondly, the authors use data sets that contain a large number of non-Caucasian people. Since they mention the number of training and testing data points used, it is easy to verify that either their “Experimental Setup” section is not wrong or this paragraph is. Given that the images given as examples in the paper itself appear to falsify this paragraph, it seems clear that this paragraph is not true. At no point in the entire paper other than this paragraph do they say anything about only being interested in Caucasian people, and they do not mention Sundering the data to Caucasians. They do mention subsetting the data to women.

While I generally believe in making the most charitable assumptions, it seems uncredible that this might be a mistake or that the authors might be unaware that this paragraph is false. Not only do the reference images in their own paper falsify it, one of their data sets is drawn from a paper titled “SCUT-FBP5500‡ : A Diverse Benchmark Dataset for Multi-Paradigm Facial Beauty Prediction.” The very first page of this paper prominently features a graphic showing non-Caucasian people and the word “diverse” appears in their abstract three times. For their other data set (CelebA), the project website shows demo examples of non-Caucasian people. It does not appear possible for the authors to have done their due diligence and not noticed this. Additionally, it took me only a couple seconds to find dozens of papers studying “the relationship between gender, race, cultural background and the perception of facial attractiveness” and even Googling that phrase brings up lots of papers. Given the location of this paragraph within the paper and the fact that the paragraph blatantly contradicts the description of the experiment in the paper itself I fear this paragraph was added later after concerns about the paper were raised in order to mislead the reader and justify their poor ethical practice.

I also believe that the validation methodologies considered by the paper are extremely insufficient, even setting aside social and ethical concerns.The authors say

>To evaluate the image quality from human’s perception, we develop a user study and ask users to vote the most attractive one among ours and the baseline. 100 face images from testing set are submitted to Amazon Mechanical Turk (AMT), and each survey requires 20 users. We collect 2000 data points in total to evaluate human preference. The final results demonstrate the superiority of out model, showing in Table 1.

This is a rather small sample size, especially as no analysis of variance or estimation of uncertainty is done. Despite the extensive literature on how socioeconomic and racial factors influence assessments of attraction, these attributes are never discussed in the Mechanical Turk population. Additionally, they never actually assess if people find the computer generated images more attractive than the reference images, which is purportedly the entire purpose of the paper. They only ask if the image their algorithm generates is more attractive than other computer-generated images. The only further validation is that they ask their algorithm to score the beauty of the new images and find that on average the beauty rating goes up. This isn’t evidence of anything meaningful at all, as they’re using the same algorithm to evaluate if the beauty increased as they used inside their GAN to make the image more beautiful in the first place.

This is all the validation that they do in the paper.

You can find the paper on arXiv here: https://arxiv.org/abs/1912.03630

Edit: This post is based on the email I sent but I want to be clear that it is not the exact text. It has been edited for grammar and clarity. The content has not been substantively changed.

submitted by /u/StellaAthena
[link] [comments]

[D] What is the best implementation of a trainable TTS network for creating custom TTS voices?

In this instance, TTS refers to Text-To-Speech.

As the title implies I am looking for the best way to train a network to produce high-quality text to speech results in a custom voice pulled from training data. Assuming access to large amounts of high-quality speech data from a single speaker, the English language, powerful machines, and extended training times what is the best implementation/codebase to use?

I have done quite a lot of research into this but have found my results to be quite confusing. Tacotron-2 seems to me to provide the highest quality results with an open-source implementation. However, implementations such as ESPnet(1) seem to be geared more towards testing different methods rather than developing your own custom voice. I am not new to Machine Learning but I am new to applying ML to audio or language-related problems thus I am very behind on my understanding of the state of such lines of research.

If I was looking to replicate something like the results from “Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions”(2) where they used 20+ hours of data from an English to produce very natural sounding speech(3) what would be my best option? I just figured I would ask the experts of Reddit before I took the plunge on setting up a codebase and dataset only to realize there were significantly better options available.

Thanks!

(1) https://github.com/espnet/espnet

(2)(paper link) https://arxiv.org/abs/1712.05884

(3)(audio sample link) https://google.github.io/tacotron/publications/tacotron2/index.html

submitted by /u/blackfish_88
[link] [comments]

[D] attribution models

Hello,

Suppose I have the following data:

—A——–A——B——C—B———-A———–X

–B-A—-B——A——C—C—–B—X

A, B, and C represent different contacts or touch points between the firm and the customer. X is the desired event for the customer (for example, a purchase). The —— represents the time in between events.

What kind of attribution models can I build to understand the relationship between A, B, C, and X?

I can create variables based on the recency and frequency of A, B, and C. What else can I do?

I read somewhere that neural network (maybe recurrent neural network) can handle non-tabular data a lot better. Can I treat those sequences of events and their timing as non-tabular data and utilize them?

Please educate me or provide pointers. Thanks!

submitted by /u/sonicking12
[link] [comments]

[D] Early Career Advice for a Machine Learning PhD Student

I worked at company A this past summer as a PhD intern (the position only required the candidate to be pursuing a Bachelor’s degree) where about half of my responsibility was directly related to machine learning. Overall, the experience went very well and they offered me the opportunity to continue working for them part time during the semester while I was away at school; it was implied that this could lead to further employment in the form of an internship or full time position in the future, but it was never formally stated or agreed upon.

Fast forward to now: Companies B, C, and D are contacting me with further internship opportunities. These companies would likely look better on a CV (or at least offer my CV a larger diversity of experience), and put me in roles that appear to align more closely with my career goals; I’m also not crazy about the industry of company A.

My Question: Is it unethical for me to accept a position with company B, C, or D this summer? I think it’s better for me as a researcher to get a diversity of experience by joining a new company, and the roles could teach me things I won’t learn by going back to company A a second time. That being said, I’m not ruling out company A as a possible landing location when I get a full time position after graduation; I don’t want to burn any bridges.

submitted by /u/green-top
[link] [comments]

[D] Why is everyone against PhD for money/prestige/job security reasons?

People who go to med school spend an equivalent of time and effort to become physicians/surgeons and many of them openly admit to doing it for the money/prestige/job security. I am definitely interested in machine learning and would be happy to do research, but why is it wrong to say I also want to it for those reasons?

Asking as someone considering a PhD in CS at top 20-40 after my B.S. in math given that it is funded…

submitted by /u/mfdoomlives
[link] [comments]

[D] Unpopular opinion: Analytics is NOT data science

There seems to be an attempt to rebrand the field of analytics and data analysis as AI or “data science”. I see too many social groups putting together “analytics and deep learning”, while obviously the two are barely related.

In my (probably unpopular) opinion, if there’s no degree of predictive modelling in your work, it’s not data science. If it doesn’t require you to implement some ML architecture from a paper, or at least fit/predict an existing one, it’s not data science. If there is no optimization problem, it could be an interesting data-related problem, it could be incredibly elegant and helpful to describe statistically, but what science is there exactly?

Why do I care? Because it does no good to the “AI hype”. There are researchers who spend years in accademia, and self-taught deep learning experts who mastered half a dozen subjects to even begin to understand how to model real world problems with data. Of course data analysis is necessary and helpful to many businesses, but it’s a different profession that requires different skills. And it’s not the same as data science.

What do you think?

submitted by /u/bob3421o
[link] [comments]

[D] Tensorboard equivalent for post-training evaluation (in other datasets)

Hi,

I am training some models to predict wave propagation. See here if you want to get an idea.

Now, I want to test my model post-training in other datasets. There is a caveat here that my problem is spatiotemporal and the evaluation metric is not a scalar but a curve along the time axis, something like this

Right now I do the post-processing in a notebook which is ok but I was wondering if there’s a better solution out there. I am considering using tensorboard actually and write an extra event file for each experiment. It also makes sense since I got all the hparams in it anyway.

I just wanted to get some feedback and in general ask if there are tools for that out there that you peeps use.

Cheers

submitted by /u/sfotiadis
[link] [comments]